Professional Documents
Culture Documents
Econometric Lec3
Econometric Lec3
3
Multiple Regression
LECTURE
1
CHAPTER 7
Multiple Regression
The Problem of Estimation
2
NOTATION
Y = 1 + 2 X2 + 3 X3 +…+ k Xk + u
3
Explaination for the parial coefficients
Y = 1 + 2X2 + 3X3 + u (suppose this is a true model)
Y measures the change in the mean
= 2 : 2
X2 values of Y, per unit change in X2,
holding X3 constant.
or The ‘direct’ or ‘net’ effect of a unit change
in X2 on the mean value of Y
Y holding X2 constant, the direct effect
= 3
X3 of a unit change in X3 on the mean
value of Y.
^u = Y - ^ - ^ X -
^X
1 2 2 3 3
RSS
^ =2 ( Y - ^1- ^2X2 - ^3X3)(-1) = 0
1
RSS ^ -
^ X - ^ X )(-X ) = 0
=2 ( Y -
^2
1 2 2 3 3 2
^ X + ^ X X + ^ X 2 = X Y
1 3 2 2 3 3 3 3
^ 2= u^2 u^2
and =
u n-3 n- k
k=3
# of independent variables
( including the constant term)
9
Scalar forms of Var and SE of OLS estimators
2 2
var(ˆ2 ) se( ˆ3 )
2i 23 )
x 2
(1 r 2
3i 23 )
x 2
(1 r 2
2
2
var(ˆ2 ) se( ˆ2 )
2i 23 )
x 2
(1 r 2
2i 23 )
x 2
(1 r 2
r23 2
cov(ˆ2 , ˆ3 )
(1 r232 2i
x 2
3i
x 2
2
( x2i x3i ) 2
r
2 i 3i
23 2 2
x x
10
Properties of multiple OLS estimators
_ _ _
1. The regression line(surface)passes through the mean of Y1, X2, X3
_ _ _
i.e., ^ ^ ^
1 = Y - 2X2 - 3X3 Linear in parameters
_
^ ^_ ^ _ Regression through the mean
==> Y = 1 + 2X2 + 3X3
_
2. ^=Y+^
Y 2x2 + ^3x3 Unbiased: E(^)=
i i
or y =^x +
2 2
^x
3 3
5. ^^
uY=0 random sample
11
Properties of multiple OLS estimators
6. As X2 and X3 are closely related ==> var(^2) and var(^3)
become large and infinite. Therefore the true values of 2
and 3 are difficult to know.
All the normality assumptions in the two-variables case regression
are also applied to the multiple variable regression.
But one addition assumption is
No exact linear relationship among the independent variables.
(No perfect collinearity, i.e., Xk Xj )
14
Applying elasticities
percentage change in y y/y y x
= = =
percentage change in x x/x x y
Y = 1 + 2 X
Y
X = 2
Y X X
= = 2
X Y Y
15
Estimating elasticities
^ Y X X
= = b2
X Y Y
^Y = b1 + b2X t = 4 + 1.5X t
t
X = 8 = average number of years of experience
Y = $10 = average wage rate
^
= b2 X 8
= 1.5 = 1.2
Y 10 16
log-log models: Nonlinearities
ln(Y) = 1 + 2 ln(X)
ln(Y) ln(X)
= 2 X
X
1 Y 1 X
Y X = 2 X
X 17
X Y
Y X = 2
18
Summary
dY
Model Equation dY ( Y )
( ) dX
dX X
Y 1 2 X dY X
2( )
linear 2
dX Y
dY
Log-log ln Y 1 2 ln X d ln Y
Y 2 2
d ln X dX
X
dY Y
==> 2( )
dX X
19
Summary(Count.)
dY
Log-lin ln Y 1 2 X d ln Y Y 2X
2
dX dX
dY
==> 1Y
dX
Lin-log Y 1 2 ln X
dY dY 2 1
d ln X dX
2
Y
X
dY 1
==> 2
dX X
1 dY dY
2 ( -1 )
Reciprocal Y 1 2 1 1 2
X d ( 2 ) dX XY
X X
dY -1
==> 2
dX X2 20
Application of functional form regression
1. Cobb-Douglas Production function:
Y 1L
K e
u
Transforming:
ln Y ln 1 2 ln L 3 ln K u
==> ln Y ’1 2 ln L 3 ln K u
d ln Y : elasticity of output w.r.t. labor input
2
d lnL
d ln Y : elasticity of output w.r.t. capital input.
3
d lnK
costs
i.e. Y 1 2 X 3 X u (MC)
2
MC
costs
or
Y 1 2 X 3 X 4 X u (TC)
2 3
TC
y 22
Chapter 8
23
Hypothesis Testing in multiple regression:
1. Testing individual partial coefficient
2. Testing the overall significance of all coefficients
3. Testing restriction on variables (add or drop): Xk = 0 ?
4. Testing partial coefficient under some restrictions
Such as 2+ 3 = 1;
or 2 = 3 (or 2+ 3 = 0); etc.
5. Testing the functional form of regression model.
24
1. Individual partial coefficient test
1 holding X3 constant: Whether X2 has the effect on Y ?
H0 : 2 = 0 Y
= 2 = 0?
X2
H1 : 2 0
^ -0
0.726
t= 2 = = 14.906
Se (^ )
2 0.048
Compare with the critical value tc0.025, 12 = 2.179
^
Answer : Yes, 2 is statistically significant and is
significantly different from zero.
25
1. Individual partial coefficient test (cont.)
2 holding X2 constant: Whether X3 has the effect on Y?
H0 : 3 = 0 Y
= 3 = 0?
X3
H1 : 3 0
^
3 - 0 2.736-0
t= = = 3.226
Se (^) 0.848
3
^
Answer: Yes, 3 is statistically significant and is
significantly different from zero.
26
2. Testing overall significance of the multiple regression
27
^ ^u
Analysis of Variance: Since y = y +
==> y2 = y^2 + u^2
ANOVA TABLE TSS = ESS + RSS
(SS) (MSS)
Source of variation Sum of Square df Mean sum of Sq.
y^2
Due to regression(ESS) y ^ 2
k-1
k-1
Due to residuals(RSS) u^2 n-k u^2 ^2
n-k = u
H0 : 2 = … = k = 0
H1 : 2 … k 0 if F > Fcα,k-1,n-k ==> reject Ho 28
Three- 2x2 + ^3x3 + u^
y= ^
variable
y2 = ^2 x2 y + ^
3 x3 y + u^2
case
TSS = ESS + RSS
ANOVA TABLE
Source of variation SS df(k=3) MSS
ESS ^ x y + ^ x y
3-1 ESS/3-1
2 2 3 3
RSS ^2
u n-3 RSS/n-3
(n-k)
TSS y2 n-1
ESS / k-1 (^ x y + ^ x y) / 3-1
2 2 3 3
F-Statistic = = ^2 / n-3
RSS / n-k u 29
An important relationship between R2 and F
ESS / k-1 ESS (n-k)
F= =
RSS / n-k RSS (k-1)
ESS n-k
=
TSS-ESS k-1 For the three-variables case :
ESS/TSS n-k R2 / 2
=
ESS F=
1 - TSS k-1 (1-R2) / n-3
R2 n-k
=
1 - R2 k-1
R2 / (k-1) (k-1) F
F = R =
2
R 2
/ k-1
F =
* =
(1-R ) / n- k
2
0.9710 / 3
=
(1-0.9710) /16
= 179.13
Fc(0.05, 4-1, 20-4) = 3.24
k-1 n-k
Since F* > Fc ==> reject H0. 32
Construct the ANOVA Table (8.4) .(Information from EViews)
Source of
variation
SS Df MSS
2 2 2 2
Due to R (y ) k-1 R (y )/(k-1)
regression =(0.971088)(28.97771)2x19
(SSE) =15493.171 =3 =5164.3903
2 2 2 2 2
Due to (1- R )(y ) or ( u ) n-k (1- R )(y )/(n-k)
Residuals =(0.0289112)(28.97771) )2x19
(RSS) =461.2621 =16 =28.8288
2
Total (y ) n-1
(TSS) =(28.97771) 2x19
=15954.446 =19
H 0 : 1 = 2 = 3 = 0
ESS / k-1 R2
/ k-1 0.707665 / 2
F =
*
= =
RSS/(n- k) (1-R ) / n- k (1-0.707665)/ 61
2
F* = 73.832
Since F* > Fc
==> reject H0.
34
Construct the ANOVA Table (8.4) .(Information from EVIEWS)
Source of
variation
SS Df MSS
2 2 2 2
Due to R (y ) k-1 R (y )/(k-1)
regression =(0.707665)(75.97807)2x64
(SSE) =261447.33 =2 =130723.67
2 2 2 2 2
Due to (1- R )(y ) or ( u ) n-k (1- R )(y )/(n-k)
Residuals =(0.292335)(75397807)2x64
(RSS) =108003.37 =61 =1770.547
2
Total (y ) n-1
(TSS) =(75.97807)2x64
=369450.7 =63
H0 : 2 = 0, 3= 0,
H1 : 2 0 ; 3 0
36
37
3.Testing the addition variable in the regression model
Old model :
Y = 1 + 2 X2 + u1
New model :
Y = 1 + 2X2 + 3 X3 + u2
Old model
40
New model H0 : add X3 (R) is not suitable
3 = 0
Fc0.05, 1, 39
= 4.17
(add or drop)
(R new - R old) / df
2 2
# of new regressors
F* =
(1 - R2new) / df in the new
n-k ( )
model
(0.9872 - 0.9868) / 1
= = 1.218 Since
(1 - 0.9872) / 39 F* < Fc ==> not reject 41H0.
Add an relevant variable X3: (Studenmund, pp.166)
Old model
Y = 0 + 1 X1 + 2 X2 + u
Next:
New model: Y = 0 + 1 X1 + 2 X2 + 3 X3 + u’
H0 : add X3 (YD variable is not suitable, 3 = 0
42
Add an relevant variable X3(YD): (Studenmund, pp.166)
New model
Fc
= 4.08 Since F* > Fc ==> reject H0.
(0.05, 1, 40)
43
Add variables in general discussion:
46
Use the R2 instead of ESS or RSS in the F-test.
Restriction Test:
48
New Model or unrestricted model
50
Dropping variable: 2 H0: No effect of X5, i.e., 5 = 0
Since t*(β ) < tc ==> not reject H0
5
Y = 1 X22 X3 3 eu
Restricted least squares:
Constant returns to scales 2 = 1 - 3
2 + 3 = 1
3 = 1 - 2
ln Y = 1 + 2 ln X2 + 3 ln X3 + u unrestricted
model
=> ln Y = 1 + ( 1 - 3 ) ln X2 + 3 ln X3 + u
=> ln Y = 1 + ln X2 + 3 ( ln X3 – lnX2 ) + u
=> (ln Y - ln X2) = 1 + 3 ( ln X3 – lnX2) + u
Y X3 restricted
=> ln(X ) = ’1 +’3ln( X ) + u’
2 2 model
Y* = ’1 + ’3 X* + u’ 52
OR
=> ln Y = 1 + 2 ln X2 + (1- 2 ) ln X3 + u
=> ln Y = 1 + 2 ln X2 + lnX3 - 2 ln X3 + u
=> (ln Y - ln X3) = 1 + 2 ( lnX2 – lnX3) + u
Y X2
=> ln(X ) = ”1 + ”2 ln ( X ) + u” restricted
3 3
model
Y** = ”1+”3 X** + u”
53
Unrestricted equation: Restricted equation:
lnY = 1+ 2lnX2+3lnX3 + u ln(Y/X2) = ’1+ ’3ln(X3/X2)
H 0 : 3 - 4 = 0 H0: = 0
H1: 0
tc(0.05, 931)= 1.96
(R2R – R2UR) / m
F* = =0
(1-R UR) / n - k
2 => not reject H0
57
F (0.05, 4, 931) =2.37
c
5- Test for Functional Form
58
(MacKinnon, White, Davidson)
MWD Test for the functional form (Gujarati(2003) pp.280)
H0: linear model;
H1:log-linear model
^
1. Run OLS on the linear model, obtain Y
^ ^ + ^ X + ^ X
Y =
^
1 2 2 3 3
^
and check t-statistic of ’3
60
MWD TEST: TESTING the Functional form of regression
Example: 8.5
Step 1:
Run the linear model
and obtain
^
Y
^
CV1 = _ = 1583.279
Y 24735.33
= 0.064
61
Step 2:
Run the log-log model
and obtain
^
lnY
fitted
or
estimated
^
0.07481
CV2 = _ = = 0.0074
Y 10.09653
62
Step 4:
H0 : true model
is linear
tc0.05, 11 = 1.796
tc0.10, 11 = 1.363
t* < tc at 5%
=> not reject H0
t* > tc at 10%
=> reject H0
63
Step 6:
H0 : true model is
log-log model
tc0.025, 11 = 2.201
tc0.05, 11 = 1.796
tc0.10, 11 = 1.363
65
Compare two different functional form models:
Model 1 Model 2
linear model log-log model
^
/ Y of model 1 2.1225/89.612 0.0236
Coefficient Variation = =
(C.V.) ^
/ Y of model 2 0.0217/4.4891 0.0048
H0 : no structural change
H1 : yes
Procedures:
67
2. Run OLS on two sub-sample groups separately and
obtain the RSS1, and RSS2
68
Structural stability:CHOW TEST
69
Scatter plot of Income and Savings
70
Structural stability : Ho: Var(u1) = Var(u2) = 2
Whole sample
RSSR
71
Y = 1 + 2X + u1
Sub-sample n1
RSS1
72
Y = 1 + 2X +u2
Sub-sample n2
RSS2
73
Empirical Results:
2
Dep. Constant Indep. V R SEE RSS n
variable X
Y 624226 0.0376 0.7672 31.12 23248.3 26
(70-95) (4.89) (8.89)
75
THE END
76