One , Basic command of linear regression
regress y x1 x2 ( Red means that the command can be abbreviated to red )
with Nerlove Data, for example ( The data are attached below )
regress lntc lnq lnpf lnpk lnpl
The upper part of the table is an analysis of variance , Including the sum of squares of regression , Sum of squares of residuals , mean square ,F Inspection, etc . The goodness of fit is given on the right side of the upper part R2 And adjusted R2.root MSE
Represents the standard error of the equation .
The lower part of the table shows the point estimates and interval estimates of regression coefficients , Standard error sum t Test value .
If the regression does not require a constant term , Can be added after the command noconstant
regress lntc lnq lnpf lnpk lnpl,noc
Two , Regression with constraints
If only the samples satisfying certain conditions are regressed , Conditions can be added if condition
If in this case q>=6000 Identified as a large enterprise , The following commands can be set
regress lntc lnq lnpf lnpk lnpl if q>=6000
Or virtual variables , Define a new variable large , If it's a big business , Then the value is 1, Otherwise 0, The code is
g large=(q>=6000) regress lntc lnq lnpf lnpk lnpl if large
The output is equivalent to the above if Conditional results .
If we return to non large enterprises , It can be expressed as
regress lntc lnq lnpf lnpk lnpl if large==0
If the regression coefficient needs to satisfy some specified conditions , such as a1+a2+a3=1,a1=2a2 etc. , This can be done by setting constraints :
constraint def 1 lnpl+ lnpk+ lnpf=1 cnsreg lntc lnq lnpf lnpk lnpl,c(1)
constraint def 1 Define the first constraint ,cnsreg Represents a constrained regression ,c(1) Indicates that the constraint condition is satisfied 1
If several conditions need to be satisfied at the same time , Conditions can be further set 2, condition 3, To add constraints lnq=1 take as an example
cons def 2 lnq=1 cnsreg lntc lnq lnpf lnpk lnpl,c(1-2)
Three , forecast , Inspection and related calculation
If you want to calculate the fitting value of the dependent variable , And save to the new variable yhat in , Can be used predict:
Take the prediction of unconstrained regression as an example :
regress lntc lnq lnpf lnpk lnpl predict yhat
The prediction results are saved in the original data set
The residual is further calculated , And save to e1 in , Then we can ( among residual Can be abbreviated ):
predict e1,residual
residual e1 The results are stored in the original data
If you want to calculate a coefficient , It can be used directly display Expression mode , Such as computing lnq Square of
di _b[lnq]^2
If it is necessary to test the coefficient under certain conditions , Can be used test condition , To test lnq=1,lnpl+lnpk+lnpf=1 take as an example
te lnq=1 te (lnq=1)(lnpl+lnpk+lnpf=1)
Separate inspection lnq=1,F The test showed that the original hypothesis was rejected ; Joint inspection lnq=1,lnpl+lnpk+lnpf=1, Rejecting the original hypothesis of joint establishment .
For two coefficients at the same time equal to 0 Make assumptions , Commands can be used te variable 1 variable 2 , To test lnpl lnpk Union equals 0 take as an example :
test lnpl lnpk
F The test showed that both were not rejected 0 The original hypothesis of .
Technology