Professional Documents
Culture Documents
Allchap1 PDF
Allchap1 PDF
Marsh
PowerPoint Slides
for
Undergraduate Econometrics
by
Lawrence C. Marsh
The Role of
Econometrics
in Economic Analysis
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
1.2
The Role of Econometrics
Using Information:
economic theory
economic data } economic
decisions
c = f(i)
qs = f( p, pc, pf ) supply
p = own price; pc = price of competitive products;
ps = price of substitutes; pf = price of factor inputs
Copyright 1996 Lawrence C. Marsh
1.7
How much ?
1. unknown
and
2. unobservable
Copyright 1996 Lawrence C. Marsh
1.9
The Statistical Model
c = f(i) + e
Systematic part provides prediction, f(i),
but actual will miss by random error, e.
Copyright 1996 Lawrence C. Marsh
The Consumption Function 1.11
c = f(i) + e
Need to define f(i) in some way.
To make consumption, c,
a linear function of income, i :
f(i) = β1 + β2 i
c = β1 + β2 i + e
Copyright 1996 Lawrence C. Marsh
1.12
The Econometric Model
y = β1 + β2 X2 + β3 X3 + e
Controlled (experimental)
vs.
Uncontrolled (observational)
• economic model
economic variables and parameters.
• statistical model
sampling process with its parameters.
• data
observed values of the variables.
Copyright 1996 Lawrence C. Marsh
1.15
The Practice of Econometrics
“Skippy”
Copyright 1996 Lawrence C. Marsh
2.1
Chapter 2
Some Basic
Probability
Concepts
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
2.2
Random Variable
random variable:
A variable whose value is unknown until it is observed.
The value of a random variable results from an experiment.
Examples:
Gross national product (GNP)
money supply
interest rates
price of eggs
household income
expenditure on clothing
Copyright 1996 Lawrence C. Marsh
2.6
Dummy Variable
die x f(x)
one dot 1 1/6
two dots 2 1/6
three dots 3 1/6
four dots 4 1/6
five dots 5 1/6
six dots 6 1/6
Copyright 1996 Lawrence C. Marsh
2.8
A discrete random variable X
has pdf, f(x), which is the probability
that X takes on the value x.
f(x) = P(X=x)
0.4
f(x) 0.3
0.2
0.1
0 1 2 3 X
number, X, on Dean’s List of three roommates
Copyright 1996 Lawrence C. Marsh
2.10
A continuous random variable uses
area under a curve rather than the
height, f(x), to represent probability:
f(x)
red area
green area 0.1324
0.8676
. .
$34,000 $55,000 X
P[X=a] = P[a<X<a]=0
b
P[a<X<b]= ∫ f(x) dx
a
n
Rule 1: Σ xi = x1 + x2 + . . . + xn
i=1
n n
Rule 2: Σ axi = a Σ xi
i=1 i=1
n n n
Rule 3: Σ (xi + yi) = Σ xi + Σ yi
i=1 i=1 i=1
n x1 + x2 + . . . + xn
Rule 5: x = n Σ xi =
1
i=1 n
n
Notation: Σ f(xi) = Σi f(xi) = Σ f(xi)
x i=1
n m n
Rule 7: Σ Σ f(xi,yj) = Σ [ f(xi,y1) + f(xi,y2)+. . .+ f(xi,ym)]
i=1 j=1 i=1
1. Empirically:
The expected value of a random variable, X,
is the average value of the random variable in an
infinite number of repetitions of the experiment.
In other words:
Analytical mean:
n
E[X] = Σ xi f(xi)
i=1
EX = Σ xi f(xi)
i=1
Σ
2 2
EX = xi f(xi)
i=1
It is important to notice that f(xi) does not change!
Σ
3 3
EX = xi f(xi)
i=1
Copyright 1996 Lawrence C. Marsh
2.22
EX = 0 (.1) + 1 (.3) + 2 (.3) + 3 (.2) + 4 (.1)
= 1.9
2 2 2 2 2 2
EX = 0 (.1) + 1 (.3) + 2 (.3) + 3 (.2) + 4 (.1)
= 0 + .3 + 1.2 + 1.8 + 1.6
= 4.9
3 3 3 3 3 3
EX = 0 (.1) + 1 (.3) + 2 (.3) + 3 (.2) +4 (.1)
= 0 + .3 + 2.4 + 5.4 + 6.4
= 14.5
Copyright 1996 Lawrence C. Marsh
2.23
n
E [g(X)] = Σ
i=1
g(xi) f(xi)
E(X+a) = E(X) + a
Multiplying by constant will multiply
its expected value by that constant:
E(bX) = b E(X)
Copyright 1996 Lawrence C. Marsh
2.26
Variance
2
var(X) = E [(X - EX) ]
Copyright 1996 Lawrence C. Marsh
2
2.27
var(X) = E [(X - EX) ]
2
var(X) = E [(X - EX) ]
2 2
= E [X - 2XEX + (EX) ]
2 2
= E(X ) - 2 EX EX + E (EX)
2 2 2
= E(X ) - 2 (EX) + (EX)
2 2
= E(X ) - (EX)
2 2
var(X) = E(X ) - (EX)
Copyright 1996 Lawrence C. Marsh
2.28
variance of a discrete
random variable, X:
n
var ( X) = ∑(xi - EX ) f (xi ) 2
i=1
Z = a + cX
var(Z) = var(a + cX)
2
= E [(a+cX) - E(a+cX)]
2
= c var(X)
2
var(a + cX) = c var(X)
Copyright 1996 Lawrence C. Marsh
2.31
Joint pdf
E(XY) = Σ Σ xi yj f(xi,yj)
i j
E(XY) = (0)(1)(.45)+(0)(2)(.15)+(1)(1)(.05)+(1)(2)(.35)=.75
Copyright 1996 Lawrence C. Marsh
2.34
Marginal pdf
f(x,y) f(x,y)
f(x|y) = f(y|x) =
f(y) f(x)
Copyright 1996 Lawrence C. Marsh
2.37
conditonal
Y=1 Y=2
f(Y=1|X = 0)=.75 f(Y=2|X= 0)=.25
.75 .25
X=0 .45 .15 .60
f(X=0|Y=1)=.90 .90 .30 f(X=0|Y=2)=.30
f(X=1|Y=1)=.10 .10 .70 f(X=1|Y=2)=.70
X=1 .05 .35 .40
.125 .875
f(Y=1|X = 1)=.125 f(Y=2|X = 1)=.875
.50 .50
Copyright 1996 Lawrence C. Marsh
2.38
Independence
cov(X,Y) = E(XY) - EX EY
Copyright 1996 Lawrence C. Marsh
Y=1 Y=2 2.42
covariance
.50 .50 cov(X,Y) = E(XY) - EX EY
EY=1(.50)+2(.50)=1.50 = .75 - (.40)(1.50)
= .75 - .60
EX EY = (.40)(1.50) = .60
= .15
E(XY) = (0)(1)(.45)+(0)(2)(.15)+(1)(1)(.05)+(1)(2)(.35)=.75
Copyright 1996 Lawrence C. Marsh
2.43
Correlation
cov(X,Y)
ρ(X,Y) =
var(X) var(Y)
Correlation is a pure number falling between -1 and 1.
Copyright 1996 Lawrence C. Marsh
Y=1 Y=2 2.44
EX=.40
2 2 2
EX=0(.60)+1(.40)=.40
X=0 .45 .15 .60 2 2
var(X) = E(X ) - (EX)
2
= .40 - (.40)
.05 .35 .40 = .24
X=1
cov(X,Y) = .15
Y~ 2
N(β,σ )
1 - (y - β) 2
f(y) = exp
2 π σ2 2 σ2
f(y)
β y
Copyright 1996 Lawrence C. Marsh
2.49
The Standardized Normal
Z = (y - β)/σ
Z ~ N(0,1)
1 - z2
f(z) = exp
2π
2
Copyright 1996 Lawrence C. Marsh
2.50
Y~ N(β,σ2)
f(y)
β
a
y
β
a b
y
a-β Y-β b-β
P[a<Y<b] = P < < σ
σ σ
a-β b-β
= P <Z<
σ σ
Copyright 1996 Lawrence C. Marsh
2.52
Linear combinations of jointly
normally distributed random variables
are themselves normally distributed.
W ~ N[ E(W), var(W) ]
Copyright 1996 Lawrence C. Marsh
2.53
Chi-Square
E[V] = E[ χ(m) ] = m
2
mean:
E(y|x) = β1 + β2 x
Copyright 1996 Lawrence C. Marsh
3.4
f(y|x=480)
f(y|x=480)
µy|x=480 y
µy|x=480 µy|x=800 y
∆E(y|x)
∆E(y|x) β2=
∆x
∆x
β1{
x (income)
re
itu
nd
pe
ex
.
.
y
t
re
it u
n d
pe
e x
.
.
.
x1 x2 x3 income xt
Figure 3.3+. The variance of yt increases
as household income, x t , increases.
Copyright 1996 Lawrence C. Marsh
3.9
Assumptions of the Simple Linear
Regression Model - I
1. The average value of y, given x, is given by
the linear regression:
E(y) = β1 + β2x
2. For each value of x, the values of y are
distributed around their mean with variance:
var(y) = σ2
3. The values of y are uncorrelated, having zero
covariance and thus no linear relationship:
cov(yi ,yj) = 0
4. The variable x must take at least two different
values, so that x ≠ c, where c is a constant.
Copyright 1996 Lawrence C. Marsh
3.10
One more assumption that is often used in
practice but is not required for least squares:
y ~ N [(β1+β2x), σ2 ]
Copyright 1996 Lawrence C. Marsh
3.11
The Error Term
y is a random variable composed of two parts:
y3 .} e3
y2 e2 {
.
y1 }
. e1
x1 x2 x3 x4 x
Figure 3.5 The relationship among y, e and
the true regression line.
Copyright 1996 Lawrence C. Marsh
y 3.13
. y4
^e { ^y = b + b x
4
.^y 1 2
^y 4
.} ^e3
3
y2 .
^e { . y3
2 .
y^2
^y
1.
} ^e
y1 . 1
x1 x2 x3 x4 x
Figure 3.7a The relationship among y, ^e and
the fitted regression line.
Copyright 1996 Lawrence C. Marsh
y 3.14
.y4 ^y = b + b x
y^*2
^y*
. 3
{.
^e*
4
1 2
^y*= b* + b* x
^y* 1 2
y^*1. . ^e*3 { 4
^e* { y .
{
2 . 2 y
3
^e*
1
y1.
x1 x2 x3 x4 x
Figure 3.7b The sum of squared residuals
from any other line will be larger.
Copyright 1996 Lawrence C. Marsh
f(.) 3.15
f(e) f(y)
0 β1+β2x
e t = y t - β1 - β2x t
T
S(β1,β2) = Σ(y t - β1 - β2x t )2 (3.3.4)
t =1
∂S(.)
= - 2 Σ (y t - β1 - β2x t )
∂β1
∂S(.)
= -2Σ x t (y t - β1 - β2x t )
∂β2
Set each of these two derivatives equal to zero and
solve these two equations for the two unknowns: β1 β2
Copyright 1996 Lawrence C. Marsh
Minimize w. r. t. β1 and β2 : 3.21
T
S(.) = Σ
t =1
(y t - β1 - β2x t )2
S(.)
S(.)
∂S(.) < .
0
∂βi ∂S(.) =
0
.∂S(.)
∂βi ∂βi
>0
.
bi βi
Copyright 1996 Lawrence C. Marsh
To minimize S(.), you set the two 3.22
derivatives equal to zero to get:
∂S(.)
= - 2 Σ (y t - b1 - b2x t ) = 0
∂β1
∂S(.)
= -2Σ x t (y t - b1 - b2x t ) = 0
∂β2
When these two terms are set to zero,
β1 and β2 become b1 and b2 because they no longer
represent just any value of β1 and β2 but the special
values that correspond to the minimum of S(.) .
Copyright 1996 Lawrence C. Marsh
3.23
- 2 Σ (y t - b1 - b2x t ) = 0
-2Σ x t (y t - b1 - b2x t ) = 0
Σ y t - Tb1 - b2 Σ x t = 0
Σ x t y t - b1 Σ x t - b2 Σ xt
2
= 0
Tb1 + b2 Σ x t = Σ yt
Σ xt + b2 Σ xt Σ xtyt
2
b1 =
Copyright 1996 Lawrence C. Marsh
3.24
Tb1 + b2 Σ x t = Σ yt
Σ xt + b2 Σ xt Σ xtyt
2
b1 =
T Σ x t yt -
Σ xt Σ yt
b2 =
TΣ t - ( Σ t)
2
x x 2
b1 = y - b2 x
Copyright 1996 Lawrence C. Marsh
3.25
elasticities
percentage change in y ∆y/y ∆y x
η = = =
percentage change in x ∆x/x ∆x y
∆y x ∂y x
η = lim =
∆x→ 0 ∆x y ∂x y
Copyright 1996 Lawrence C. Marsh
3.26
applying elasticities
E(y) = β1 + β2 x
∂E(y)
= β2
∂x
∂E(y) x x
η = = β2
∂x E(y) E(y)
Copyright 1996 Lawrence C. Marsh
3.27
estimating elasticities
∂y x x
η =
^
= b2
∂x y y
y^t = b1 + b2 x t = 4 + 1.5 x t
x = 8 = average number of years of experience
y = $10 = average wage rate
η = b2
^ x 8
= 1.5 = 1.2
y 10
Copyright 1996 Lawrence C. Marsh
3.28
Prediction
Estimated regression equation:
y^t = 4 + 1.5 x t
xt = years of experience
y^t = predicted wage rate
^
If xt = 2 years, then yt = $7.00 per hour.
^
If xt = 3 years, then yt = $8.50 per hour.
Copyright 1996 Lawrence C. Marsh
3.29
log-log models
ln(y) = β1 + β2 ln(x)
∂ln(y) ∂ln(x)
= β2
∂x ∂x
1 ∂y 1 ∂x
= β2
y ∂x x ∂x
Copyright 1996 Lawrence C. Marsh
3.30
1 ∂y 1 ∂x
= β2
y ∂x x ∂x
x ∂y
= β2
y ∂x
elasticity of y with respect to x:
x ∂y
η = = β2
y ∂x
Copyright 1996 Lawrence C. Marsh
4.1
Chapter 4
Properties of
Least Squares
Estimators
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
4.2
Simple Linear Regression Model
yt = β1 + β2 x t + ε t
1. yt = β1 + β2x t + ε t
2. E(ε t) = 0 <=> E(yt) = β1 + β2x t
3. var(ε t) = σ 2 = var(yt)
4. cov(ε i,ε j) = cov(yi,yj) = 0
5. xt ≠ c for every observation
6. ε t~N(0,σ 2) <=> yt~N(β1+ β2x t,σ 2)
Copyright 1996 Lawrence C. Marsh
4.4
The population parameters β1 and β2
are unknown population constants.
Eb1 = β1
Copyright 1996 Lawrence C. Marsh
4.11
Equivalent expressions for b2:
var(b2) = σ2
Σ(x t − x) 2
Given b1 = y − b2x
the variance of the estimator b1 is:
Σx t 2
var(b1) = σ 2
Τ Σ(x t − x) 2
Copyright 1996 Lawrence C. Marsh
4.14
Covariance of b1 and b2
−x
cov(b1,b2) = σ2
Σ(x t − x)2
σ2 Σx t2
b1 ~ N β1 ,
Τ Σ(x t − x)2
σ2
b2 ~ N β2 ,
Σ(x t − x)2
Copyright 1996 Lawrence C. Marsh
4.20
yt and ε t normally distributed
The least squares estimator of β2 can be
expressed as a linear combination of yt’s:
b2 = Σ wt yt
(x t − x)
where wt =
Σ(x t − x)2
b1 = y − b2x
^e = yt − b1 − b2 x t
t
Σ
T
^
e 2
t
^2
σ = t =1
T− 2
σ
^ 2 is an unbiased estimator of σ 2
Copyright 1996 Lawrence C. Marsh
4.24
The Least Squares
Predictor, ^yo
^y = b + b x
o 1 2 o (4.7.2)
Copyright 1996 Lawrence C. Marsh
5.1
Chapter 5
Inference
in the Simple
Regression Model
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
5.2
Assumptions of the Simple
Linear Regression Model
1. yt = β1 + β2x t + ε t
2. E(ε t) = 0 <=> E(yt) = β1 + β2x t
3. var(ε t) = σ 2 = var(yt)
4. cov(ε i,ε j) = cov(yi,yj) = 0
5. xt ≠ c for every observation
6. ε t~N(0,σ 2) <=> yt~N(β1+ β2x t,σ 2)
Copyright 1996 Lawrence C. Marsh
5.3
Probability Distribution
of Least Squares Estimators
σ2 Σx2t
b1 ~ N β1 ,Τ Σ(x t − x)2
σ2
b2 ~ N β2 ,
Σ(x t − x) 2
Copyright 1996 Lawrence C. Marsh
5.4
Error Variance Estimation
Σ ^ 2
et
σ
^2 =
Τ−2
b2 − β2
Ζ = ∼ Ν(0,1)
var(b2)
Copyright 1996 Lawrence C. Marsh
5.7
Simple Linear Regression
yt = β1 + β2x t + ε t where E ε t = 0
yt ~ N(β1+ β2x t , σ 2)
since Eyt = β1 + β2x t
ε t = yt − β1 − β2x t
Therefore, ε t ~ N(0,σ 2) .
Copyright 1996 Lawrence C. Marsh
5.8
Create a Chi-Square
Σt =1(ε t /σ)2 =
(ε1 /σ)2 + (ε 2 /σ)2 +. . .+ (ε T /σ)2
Z
t= ~ t(m)
V/m
where Z ~ N(0,1)
and V ~ χ 2
(m)
Copyright 1996 Lawrence C. Marsh
5.12
Z
t = ~ t(m)
V / ( T− 2)
(b2 − β2)
where Z =
var(b2)
σ2
and var(b2) =
Σ( xi − x )2
Copyright 1996 Lawrence C. Marsh
5.13
Z
t = (T−2) σ
^2
V =
V / (T-2) σ 2
(b2 − β2)
var(b2)
t =
(T−2) σ
^2
( T− 2)
σ 2
σ
Copyright 19962 Lawrence C. Marsh
5.14
var(b2) =
Σ( xi − x )2
(b2 − β2)
t =
se(b2)
Copyright 1996 Lawrence C. Marsh
5.16
Student’s t - statistic
(b2 − β2)
t = ~ t (T−2)
se(b2)
(1−α)
α/2 α/2
-tc 0 tc t
red area = rejection region for 2-sided test
Copyright 1996 Lawrence C. Marsh
5.18
probability statements
P(-tc ≤ t ≤ tc) = 1 − α
(b2 − β2)
P(-tc ≤ ≤ tc) = 1 − α
se(b2)
Copyright 1996 Lawrence C. Marsh
5.19
Confidence Intervals
Two-sided (1−α)x100% C.I. for β1:
b1 − tα/2[se(b1)], b1 + tα/2[se(b1)]
Hypothesis Tests
1. A null hypothesis, H0.
2. An alternative hypothesis, H1.
3. A test statistic.
4. A rejection region.
Copyright 1996 Lawrence C. Marsh
5.22
Rejection Rules
1. Two-Sided Test:
If the value of the test statistic falls in the critical region in either
tail of the t-distribution, then we reject the null hypothesis in favor
of the alternative.
2. Left-Tail Test:
If the value of the test statistic falls in the critical region which lies
in the left tail of the t-distribution, then we reject the null
hypothesis in favor of the alternative.
2. Right-Tail Test:
If the value of the test statistic falls in the critical region which lies
in the right tail of the t-distribution, then we reject the null
hypothesis in favor of the alternative.
Copyright 1996 Lawrence C. Marsh
5.23
Type II error:
We make the mistake of failing to reject the null
hypothesis when it is false.
β = P(failing to reject H0 when it is false).
Copyright 1996 Lawrence C. Marsh
5.26
Prediction Intervals
A (1−α)x100% prediction interval for yo is:
y^ o ± tc se( f )
f = y^ o − yo se( f ) = ^ f)
var(
1 ( x − x )2
var( f ) = σ 1 +
^ ^ 2 + o
Τ Σ(x t − x)2
Copyright 1996 Lawrence C. Marsh
6.1
Chapter 6
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
6.2
Explaining Variation in yt
t=1
Σyt − Tb1 = 0
T T
Σ Σ(yt − β1)
T
2 2
et = t=1
t=1 t=1
b1 = y
∂Σ
T
2
et T
t=1
= −2 Σ(yt − b1) = 0
∂β1 t=1 Why not y?
Copyright 1996 Lawrence C. Marsh
6.3
Explaining Variation in yt
^
yt = b1 + b2xt + et
^
Explained variation: yt = b1 + b2xt
Unexplained variation:
^e = y − y
^ = y −b −b x
t t t t 1 2 t
Copyright 1996 Lawrence C. Marsh
6.4
Explaining Variation in yt
^ ^
yt = yt + et using y as baseline
yt − y = yt − y + et Why not y?
^ ^
T cross
Σ(yt−y) = Σ(yt−y) +Σe
T T
^
2 2 ^2 product
t=1 t=1 t=1
t
term
drops
SST = SSR + SSE out
Copyright 1996 Lawrence C. Marsh
6.5
Total Variation in yt
T
SST = Σ(yt − y) 2
t=1
Copyright 1996 Lawrence C. Marsh
6.6
Explained Variation in yt
et = yt−yt = yt − b1 − b2xt
^ ^
^
SSE measures variation of yt around y t
T T
SSE = Σ(yt − ^
y)t
2
=Σ ^e 2
t
t=1 t=1
Copyright 1996 Lawrence C. Marsh
6.8
Analysis of Variance Table
0≤ R ≤1 2
2 SSR
R = SST
Copyright 1996 Lawrence C. Marsh
6.10
Coefficient of Determination
SST = SSR + SSE
SST SSR SSE
Dividing = +
by SST SST SST SST
1 = SSR + SSE
SST SST
R = 2 SSR
SST = 1− SSE
SST
Copyright 1996 Lawrence C. Marsh
6.11
Coefficient of Determination
R 2 is only a descriptive measure.
^ =Σ
T
var(X) t=1
(x t − x) /( T−1)
2
Σ(yt − y) /(T−1)
T
^ =
var(Y) 2
t=1
Σ(xt − x)(yt − y)
T
t=1
r=
Σ(xt − x) Σ(yt − y)
T T
2 2
t=1 t=1
Copyright 1996 Lawrence C. Marsh
2 6.15
Correlation Analysis and R
se(b1) = ^ 1)
var(b = 490.12 = 22.1287
se(b2) = ^ 2)
var(b = 0.0009326 = 0.0305
b1 40.7676
t = = = 1.84
se(b1) 22.1287
b2 0.1283
t = se(b2)
= = 4.20
0.0305
Copyright 1996 Lawrence C. Marsh
6.18
Regression Computer Output
^2
SSE /(T-2) = σ = 1429.2455
SSR
2
R = = 1− SSE
= 0.317
SST SST
Copyright 1996 Lawrence C. Marsh
6.20
Reporting Regression Results
yt = 40.7676 + 0.1283xt
(s.e.) (22.1387) (0.0305)
yt = 40.7676 + 0.1283xt
(t) (1.84) (4.20)
Copyright 1996 Lawrence C. Marsh
6.21
Reporting Regression Results
2
R = 0.317
This R2 value may seem low but it is
typical in studies involving cross-sectional
data analyzed at the individual or micro level.
yt = β1 + β2xt + exp(β3xt) + et
Copyright 1996 Lawrence C. Marsh
Linear vs. Nonlinear 6.27
y nonlinear
relationship
food
expenditure between food
expenditure and
income
0 income x
Copyright 1996 Lawrence C. Marsh
6.28
Useful Functional Forms
1. Linear
Look at 2. Reciprocal
each form
3. Log-Log
and its
slope and 4. Log-Linear
elasticity 5. Linear-Log
6. Log-Inverse
Copyright 1996 Lawrence C. Marsh
6.29
Useful Functional Forms
Linear
yt = β1 + β2xt + et
xt
slope: β2 elasticity: β2 y
t
Copyright 1996 Lawrence C. Marsh
6.30
Useful Functional Forms
Reciprocal
yt = β1 + β2 xt + et 1
slope: elasticity:
1
− β2 2 1
− β2 x y
xt t t
Copyright 1996 Lawrence C. Marsh
6.31
Useful Functional Forms
Log-Log
ln(yt)= β1 + β2ln(xt) + et
yt
slope: β2 x elasticity: β2
t
Copyright 1996 Lawrence C. Marsh
6.32
Useful Functional Forms
Log-Linear
ln(yt)= β1 + β2xt + et
Linear-Log
yt= β1 + β2ln(xt) + et
slope: β2 _
1
elasticity: β
_
1
xt 2 yt
Copyright 1996 Lawrence C. Marsh
6.34
Useful Functional Forms
Log-Inverse
ln(yt) = β1 - β2 x + et 1
t
yt 1
slope: β2 2 elasticity: β2 x
xt t
Copyright 1996 Lawrence C. Marsh
6.35
Error Term Properties
1. E (et) = 0
2. var (et) = σ 2
3. cov(ei, ej) = 0
4. et ~ N(0, σ )
2
Copyright 1996 Lawrence C. Marsh
6.36
Economic Models
1. Demand Models
2. Supply Models
3. Production Functions
4. Cost Functions
5. Phillips Curve
Copyright 1996 Lawrence C. Marsh
6.37
Economic Models
1. Demand Models
* quality demanded (yd) and price (x)
* constant elasticity
ln(yt )= β1 + β2ln(x)t + et
d
Copyright 1996 Lawrence C. Marsh
6.38
Economic Models
2. Supply Models
* quality supplied s
(y )
and price (x)
* constant elasticity
ln(yt )= β1 + β2ln(xt) + et
s
Copyright 1996 Lawrence C. Marsh
6.39
Economic Models
3. Production Functions
* output (y) and input (x)
* constant elasticity
Cobb-Douglas Production Function:
ln(yt)= β1 + β2ln(xt) + et
Copyright 1996 Lawrence C. Marsh
6.40
Economic Models
yt = β1 + β2 x 2
t + et
Copyright 1996 Lawrence C. Marsh
6.41
Economic Models
5. Phillips Curve
nonlinear in both variables and parameters
* wage rate (wt) and time (t)
wt − wt-1
% ∆wt = w = γα + γη u
1
t-1 t
unemployment rate, ut
Copyright 1996 Lawrence C. Marsh
7.1
Chapter 7
The Multiple
Regression Model
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
Two Explanatory Variables 7.2
yt = β1 + β2xt2 + β3xt3 + et
yt = β1 + β2xt2 + β3xt3 + et
yt = output xt2 = capital xt3 = labor
1. E(et) = 0
2. var(et) = σ2
3. cov(et , es) = 0 for t ≠ s
4. et ~ N(0, σ2)
Copyright 1996 Lawrence C. Marsh
7.6
Statistical Properties of yt
1. yt = β1 + β2xt2 +. . .+ βKxtK + et
2. E (yt) = β1 + β2xt2 +. . .+ βKxtK
3. var(yt) = var(et) = σ2
4. cov(yt ,ys) = cov(et ,es) = 0 t≠s
5. The values of xtk are not random
6. yt ~ N(β1+β2xt2 +. . .+βKxtK, σ2)
Copyright 1996 Lawrence C. Marsh
7.8
Least Squares Estimation
yt = β1 + β2xt2 + β3xt3 + et
T
2
S ≡ S(β1, β2, β3) = Σ(yt − β1 − β2xt2 − β3xt3)
t=1
Define: y*t = yt − y
x*t2 = xt2 − x2
x*t3 = xt3 − x3
Copyright 1996 Lawrence C. Marsh
7.9
Least Squares Estimators
b1 = y − b1 − b2x2 − b3x3
Σ ^ 2
e t
σ
^2 =
Τ−Κ
var(b2) = σ 2
When r23 = 0
(1− r23)
2
Σ(xt2 − x2) 2 these reduce
to the simple
var(b3) =
σ 2
regression
(1− r23)
2
Σ(xt3 − x3) 2
formulas.
2
4. The correlation is close to zero: r23 0.
Copyright 1996 Lawrence C. Marsh
7.15
Covariances
yt = β1 + β2xt2 + β3xt3 + et
− r23 σ2
cov(b2,b3) =
Σ(xt2 − x2) Σ(xt3 − x3)
2
(1− r23) 2 2
yt = β1 + β2xt2 + β3xt3 + et
Since bk is a linear
function of the yt’s: bk ~ N βk, var(bk)
bk − βk
z = ~ N(0,1) for k = 1,2,...,K
var(bk)
Copyright 1996 Lawrence C. Marsh
Student-t 7.19
bk − βk bk − βk
t = =
^ )
var(b se(bk)
k
Hypothesis Testing
and
Nonsample Information
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
Chapter 8: Overview 8.2
1. Student-t Tests
2. Goodness-of-Fit
3. F-Tests
4. ANOVA Table
5. Nonsample Information
6. Collinearity
7. Prediction
Copyright 1996 Lawrence C. Marsh
Student - t Test 8.3
H0: β3 ≤ 0 b3
t= ~ t (T−K)
H1: β3 > 0 se(b3)
df = T− K
= T− 4
(1 − α) α
0 tc
Copyright 1996 Lawrence C. Marsh
Two Tail Test 8.5
H0: β2 = 0 b2
t= ~ t (T−K)
H1: β2 ≠ 0 se(b2)
df = T− K
= T− 4
α/2 (1 − α) α/2
-tc 0 tc
Copyright 1996 Lawrence C. Marsh
Goodness - of - Fit 8.6
Coefficient of Determination
2 SSR =
Σ ^
(y t− y) 2
R = SST
t=1
T
Σ(yt − y)
t=1
2
0≤ R ≤12
Copyright 1996 Lawrence C. Marsh
Adjusted R-Squared 8.7
b2 −6.642
t= = = −2.081
se(b2) 3.191
Copyright 1996 Lawrence C. Marsh
Reporting Your Results 8.9
Reporting t-statistics:
F =
(SSER − SSEU)/J H0: β2 = 0
SSEU/(T−K)
H1: β2 ≠ 0
(1964.758 − 1805.168)/1
=
1805.168/(52 − 3) dfn = J = 1
= 4.33 dfd = T− K = 49
By definition this is the t-statistic squared:
t = − 2.081 F = t2 = 4.33
Copyright 1996 Lawrence C. Marsh
Multiple Restriction F-Test 8.11
(SSER − SSEU)/J
F =
SSEU/(T−K)
(1 − α) α
0 Fc F
Copyright 1996 Lawrence C. Marsh
F-Test of Entire Equation 8.13
yt = β1 + β2Xt2 + β3Xt3 + et
2 SSR = 11776.18
=
R = SST 13581.35
0.867
Copyright 1996 Lawrence C. Marsh
Nonsample Information 8.15
yt = β1 + β2Xt2 + β3Xt3 + et
Given a set of values for the explanatory
variables, (1 X02 X03), the best linear
unbiased predictor of y is given by:
^y = b + b X + b X
0 1 2 02 3 03
Extensions
of the Multiple
Regression Model
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
9.2
Topics for This Chapter
1. Intercept Dummy Variables
2. Slope Dummy Variables
3. Different Intercepts & Slopes
4. Testing Qualitative Effects
5. Are Two Regressions Equal?
6. Interaction Effects
7. Dummy Dependent Variables
Copyright 1996 Lawrence C. Marsh
9.3
Intercept Dummy Variables
Dummy variables are binary (0,1)
yt = β1 + β2Xt + β3Dt + et
yt = speed of car in miles per hour
Xt = age of car in years
Dt = 1 if red car, Dt = 0 otherwise.
H0: β3 = 0
Police: red cars travel faster.
H1: β3 > 0
Copyright 1996 Lawrence C. Marsh
9.4
yt = β1 + β2Xt + β3Dt + et
red cars: yt = (β1 + β3) + β2xt + et
other cars: yt = β1 + β2Xt + et
yt
β1 + β3
β1 β2
red c
miles ars
per β2 othe
r car
hour s
0 age in years Xt
Copyright 1996 Lawrence C. Marsh
Slope Dummy Variables 9.5
yt = β1 + β2Xt + β3DtXt + et
Stock portfolio: Dt = 1 Bond portfolio: Dt = 0
yt yt = β1 + (β2 + β3)Xt + et
value
of β stocks
β2 + 3
porfolio bonds
β2
β1 yt = β1 + β2Xt + et
β1 = initial
investment 0 years Xt
Copyright 1996 Lawrence C. Marsh
9.6
Different Intercepts & Slopes
yt = β1 + β2Xt + β3Dt + β4DtXt + et
“miracle” seed: Dt = 1 regular seed: Dt = 0
yt yt = (β1 + β3) + (β2 + β4)Xt + et
harvest
weight β “miracle”
β2 + 4
of corn yt = β1 + β2Xt + et
β1 + β3 regular
β2
β1
rainfall Xt
Copyright 1996 Lawrence C. Marsh
yt = β1 + β2 Xt + β3 Dt + et 9.7
For men: Dt = 1.
For women: Dt = 0.
yt Men
yt = (β1+ β3) + β2 Xt + et
wage
Women
rate
β2 yt = β1 + β2 Xt + et
β1+ β3 . β2 Testing for H0: β3 = 0
discrimination
β1 . in starting wage H1: β3 > 0
0 years of experience Xt
yt = β1 + β5 Xt + β6 Dt Xt + et
Copyright 1996 Lawrence C. Marsh
9.8
For men Dt = 1.
For women Dt = 0.
yt
yt = β1 + (β5 + β6 )Xt + et Men
wage
β5 + β6 Women
rate
yt = β1 + β5 Xt + et
β5
Men and women have the same
starting wage, β1 , but their wage rates
β1 increase at different rates (diff.= β6 ).
Yt = β 1 + β 2 Xt + β 3 Dt + β 4 Dt Xt + et
H0: β3 ≤ 0 vs. Η1: β3 > 0 intercept
Testing for b3 − 0
discrimination in tn−4
starting wage. Est. Var b3 ˜
H0: β4 ≤ 0 vs. Η1: β4 > 0 slope
Testing for b4 − 0
tn−4
discrimination in
wage increases. Est. Var b 4 ˜
Copyright 1996 Lawrence C. Marsh
Testing: Ho: β3 = β4 = 0 9.12
H1 : otherwise
( SSE R − SSE U ) / 2 2
∼ F T−4
SSE U / ( T − 4 )
T
SSE U =∑(yt −b1−b2Xt −b3 Dt −b4 Dt Xt )
2
t=1
and intercept and slope
T
∑ ( yt − b 1 − b 2 X t ) 2
SSE R =
t =1
Copyright 1996 Lawrence C. Marsh
9.13
Are Two Regressions Equal?
variations of “The Chow Test”
1. Interaction Dummies
2. Polynomial Terms
(special case of continuous interaction)
20 30 40 50 60 70 80 90 Xt
People retire at different ages or not at all.
Copyright 1996 Lawrence C. Marsh
9.18
Polynomial Regression
yt = income; Xt = age
yt = β1 + β2 X t + β3 X t
2
+ β4 X t
3
+ et
yt = β1 + β2 Zt + β3 Bt + β4 Zt Bt + et
2. Probit Model
3. Logit Model
Copyright 1996 Lawrence C. Marsh
9.23
Linear Probability Model
1 quits job
yi =
0 does not quit
yt = 0
total hours of work each week Xi2
Copyright 1996 Lawrence C. Marsh
Linear Probability Model 9.25
∫−∞
zi −0.5u2
F(zi) = P[ Z ≤ zi ] = 1
e du
2π
Copyright 1996 Lawrence C. Marsh
9.27
Probit Model
Since zi = β1 + β2 Xi2 + . . . , we can
substitute in to get:
pi = P[ Z ≤ β1 + β2Xi2 ] = F(β1 + β2Xi2)
yt = 1
yt = 0
total hours of work each week Xi2
Copyright 1996 Lawrence C. Marsh
9.28
Logit Model
pi is the probability of quitting the job.
1
Define pi : pi = − (β + β X + . . .)
1 +e 1 2 i2
1
pi =
1 + e − (β1 + β2 Xi2 + . . .)
yt = 1
yt = 0
total hours of work each week Xi2
Copyright 1996 Lawrence C. Marsh
9.30
Maximum Likelihood
Heteroskedasticity
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
10.2
The Nature of Heteroskedasticity
Heteroskedasticity is a systematic pattern in
the errors where the variances of the errors
are not constant.
Ordinary least squares assumes that all
observations are equally reliable.
For efficiency (accurate estimation/prediction)
reweight observations to ensure equal error
variance.
Copyright 1996 Lawrence C. Marsh
10.3
Regression Model
yt = β1 + β2xt + et
heteroskedasticity: var(et) = σt 2
Copyright 1996 Lawrence C. Marsh
10.4
Homoskedastic pattern of errors
consumption
yt
.
. . .
. . . .
. . . . . .
. . . . .
. . . .
. . . .
. . .
.
. . . . .
.. .
.
.
income xt
Copyright 1996 Lawrence C. Marsh
10.5
The Homoskedastic Case
f(yt) yt
o n
p ti
u m
s
c o n .
.
.
.
x1 x2 x3 x4 xt
income
Copyright 1996 Lawrence C. Marsh
10.6
Heteroskedastic pattern of errors
consumption
.
yt .
. .
. . . .
. .
. . . .
. . . . .
. . . .. . . . .
.
. . . . . . .
. . .
. . . . .
. .
income xt
Copyright 1996 Lawrence C. Marsh
10.7
The Heteroskedastic Case
f(yt)
y
t
o n
p ti
u m
ns
c o
.
. rich people
.
poor people
x1 x2 x3 income xt
Copyright 1996 Lawrence C. Marsh
10.8
Properties of Least Squares
var(b2) = σ 2
Σ (xt − x )2
correct formula for least squares variance:
Σ t ( t )
σ 2 x −x 2
var(b2) =
[Σ ( t ) ]
x − x 2 2
Copyright 1996 Lawrence C. Marsh
10.10
Hal White’s Standard Errors
^
Σ t ( t )
e 2 x −x 2
est.var(b2) =
[Σ (xt − x )2]2
1. Proportional Heteroskedasticity.
(continuous function(of xt, for example))
2. Partitioned Heteroskedasticity.
(discrete categories/groups)
Copyright 1996 Lawrence C. Marsh
10.12
Proportional Heteroskedasticity
yt = β1 + β2xt + et
The variance is
assumed to be
where σt 2 = σ 2 xt proportional to
the value of xt
Copyright 1996 Lawrence C. Marsh
10.13
std.dev. proportional to xt
yt = β1 + β2xt + et
variance: var(et) = σt 2 σt 2 = σ 2 xt
standard deviation: σt = σ xt
var(e*t ) = σ 2
et is heteroskedastic, but e*t is homoskedastic.
Copyright 1996 Lawrence C. Marsh
10.15
Generalized Least Squares
These steps describe weighted least squares:
1. Decide which variable is proportional to the
heteroskedasticity (xt in previous example).
2. Divide all terms in the original model by the
square root of that variable (divide by xt ).
3. Run least squares on the transformed model
which has new yt*, x*t1 and x*t2 variables
but no intercept.
Copyright 1996 Lawrence C. Marsh
10.16
Partitioned Heteroskedasticity
yt = β1 + β2xt + et
yt = bushels per acre of corn t = 1, ,100 ...
yt 1 xt et
σ1 = β1 σ1 + β2 σ1 + σ1 t = 1, . . . ,80
yt 1 xt et
σ2 = β1 σ2 + β2 σ2 + σ2
t = 81, . . . ,100
Copyright 1996 Lawrence C. Marsh
10.18
σ
^ 2 provides estimator of σ 2 using
2 2
the 20 observations on “sweet” corn.
Copyright 1996 Lawrence C. Marsh
10.19
Detecting Heteroskedasticity
Determine existence and nature of heteroskedasticity:
Ho: σ1 2 = σ2 2
H1: σ1 2 > σ2 2 Use F
Table
^
σ 2
Goldfeld-Quandt 1
Test Statistic
GQ = ~ F[T1-K1, T2-K2]
σ
^
2
2
Autocorrelation
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
11.2
σ 2
var(et) = σ e2 = ν 2 corr(et, et−k) = ρk for k > 0
1− ρ
Copyright 1996 Lawrence C. Marsh
11.7
Autocorrelation creates some
Problems for Least Squares:
yt = β1 + β2xt + ρ et−1 + νt
(continued)
Copyright 1996 Lawrence C. Marsh
11.9
yt = β1 + β2xt + ρ et−1 + νt
yt = β1 + β2xt + et
(continued)
Copyright 1996 Lawrence C. Marsh
11.10
yt = β1 + β2xt + ρ(yt−1 − β1 − β2xt−1) + νt
y1 = β1 + β2x1 + e1
with error variance: var(e1) = σe2 = σν2 /(1-ρ2).
Σ
T
^
e ^
e
t t-1
^ρ = t=2
Σ
T
^
e 2
t=2
t-1
Copyright 1996 Lawrence C. Marsh
11.18
Durbin-Watson Test
Ho: ρ = 0 vs. H1: ρ ≠ 0 , ρ > 0, or ρ < 0
Σ ( ^ ^
− t-1)
T 2
e t e
t=2
d =
Σ t
T
^
e 2
t=1
Copyright 1996 Lawrence C. Marsh
11.19
Testing for Autocorrelation
The test statistic, d, is approximately related to ^
ρ as:
d ≈ 2(1−ρ)
^
^y ^ ^ ^ ~
T+1 = β1 + β2xT+1 + ρ eT
^ ^
where β1 and β2 are generalized least squares
~
estimates and eT is given by:
~ ^ ^
e =y −β − β x
T T 1 2 T
Copyright 1996 Lawrence C. Marsh
11.21
For h periods ahead, the best predictor is:
^y ^ ^ ^ ~
T+h = β1 + β2xT+h + ρ eT
h
^ ~
Assuming | ρ | < 1, the influence of ρ eT
^ h
Pooling
Time-Series and
Cross-Sectional Data
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
12.2
If left unrestricted,
this model requires different equations
for each firm in each time period.
Copyright 1996 Lawrence C. Marsh
12.3
Each firm gets its own coefficients: β1i , β2i and β3i
but those coefficients are constant over time.
Copyright 1996 Lawrence C. Marsh
12.4
Two-Equation SUR Model
The investment expenditures (INV) of General Electric (G)
and Westinghouse(W) may be related to their stock market
value (V) and actual capital stock (K) as follows:
i = G, W t = 1, . . . , 20
Copyright 1996 Lawrence C. Marsh
12.5
Estimating Separate Equations
We make the usual error term assumptions:
E(eGt) = 0 E(eWt) = 0
var(eGt) = σG2 var(eWt) = σ2W
cov(eGt, eGs) = 0 cov(eWt, eWs) = 0
σG = σW
2 2
Ηο: σGW = 0
σ
^ 2
λ = T rGW 2
rGW =
2
GW
σ σ
^ 2 ^2
λ ∼ χ(1)
G W asy. 2
Copyright 1996 Lawrence C. Marsh
12.14
Start with
the residuals
^eGt and ^eWt
from each
σ
^
GW =
1
T
Σ ^eGte^ Wt
equation
estimated
separately. σ
^ 2
G = 1
T
Σe
^ 2
Gt
σ
^ 2
σ
^ 2
= 1
Σ ^e 2
rGW =
2 GW W T Wt
σ σ
^ 2 ^2
G W
λ = T rGW 2
λ ∼ χ(1)
asy. 2
Copyright 1996 Lawrence C. Marsh
12.15
Fixed Effects Model
yit = β1it + β2itx2it + β3itx3it + eit
(SSER − SSEU) / J J
F= ~ F(NT − K)
SSEU / (NT − K)
β1i = β1 + µi
E(µi) = 0 var(µi) = σµ
2
Simultaneous
Equations
Models
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
13.2
Keynesian Macro Model
c = β1 + β2 y
y=c+i
Copyright 1996 Lawrence C. Marsh
13.4
The Statistical Model
The consumption equation:
ct = β1 + β2 yt + et
yt = ct + it
The Simultaneous Nature
Copyright 1996 Lawrence C. Marsh
13.5
of Simultaneous Equations
2. 1.
ct = β1 + β2 yt + et
5.
Since yt
contains et
3.
they are
4. yt = ct + it correlated
Copyright 1996 Lawrence C. Marsh
13.6
The Failure of Least Squares
yt ct et
ct
et yt it
Copyright 1996 Lawrence C. Marsh
13.8
Deriving the Reduced Form
ct = β1 + β2 yt + et
yt = ct + it
ct = β1 + β2(ct + it) + et
(1 − β2)ct = β1 + β2 it + et
Copyright 1996 Lawrence C. Marsh
13.9
Deriving the Reduced Form
(1 − β2)ct = β1 + β2 it + et
β1 β2 1
ct = + it + et
(1−β2) (1−β2) (1−β2)
ct = π11 + π21 it + νt
The Reduced Form Equation
Copyright 1996 Lawrence C. Marsh
13.10
Reduced Form Equation
ct = π11 + π21 it + νt
β1 β2
π11 = (1−β ) π21 = (1−β )
2 2
1
and νt = (1−β2) + et
Copyright 1996 Lawrence C. Marsh
13.11
yt = ct + it
where ct = π11 + π21 it + νt
yt = π11 + (1+π21) it + νt
It is sometimes useful to give this equation
its own reduced form parameters as follows:
yt = π12 + π22 it + νt
ct = π11 + π21 it + νt
Copyright 1996 Lawrence C. Marsh
13.12
yt = π12 + π22 it + νt
Since ct and yt are related through the identity:
yt = ct + it , the error term, νt, of these two
equations is the same, and it is easy to
show that:
β1
π =π =11 12
(1−β2)
π22 = (1−π21) = 1
(1−β2)
Copyright 1996 Lawrence C. Marsh
13.13
Identification
The structural parameters are β1 and β2.
π11 ^ ^ π21 ^
β1 = β2 =
^
^y = ^π + ^π x + ^π x
t1 11 21 t1 31 t2 yt1 = ^yt1 + ^νt1
^yt2 = ^π12 + ^π22 xt1 + ^π32 xt2 yt2 = ^yt2 + ^νt2
Copyright 1996 Lawrence C. Marsh
2SLS: Stage II 13.19
Nonlinear
Least
Squares
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh
14.2
Review of Least Squares Principle
(minimize the sum of squared errors)
(A.) “Regression” model with only an intercept term:
yt = α + et ∂ SSE = − 2 Σ (y − ^α) = 0
∂α t
et = yt − α Σ yt − Σ ^α = 0
2 ^ = 0
Σ yt − Τ α
Σ et = Σ (yt − α)
2
et = yt − βxt
^
Σ xtyt − Σ βxt = 0
2
^
Σxt yt − β Σxt = 0
2
Σ et = Σ (yt − βxt)
2 2
^β Σx2 = Σ x y
2 t t t
SSE = Σ (yt − βxt)
This yields an exact ^β = Σ xtyt
analytical solution: Σx2t
Copyright 1996 Lawrence C. Marsh
Review of Least Squares 14.4
(C.) Regression model with both an intercept and a slope:
2
yt = α + βxt + et SSE = Σ (yt − α − βxt)
∂ SSE = − 2 Σ x (y − ^α − ^βx ) = 0
∂β t t t
∂ SSE = − 2 Σ (y − ^α − ^βx ) = 0
∂α t t
yt = xtβ + et
PROBLEM: An exact
2 analytical solution to
SSE = Σ (yt − β
xt ) this does not exist.
^ β
β
Copyright 1996 Lawrence C. Marsh
14.7
Conclusion
The least squares principle
is still appropriate when the
model is nonlinear, but it is
harder to find the solution.
Copyright 1996 Lawrence C. Marsh
14.8
Optional Appendix
Nonlinear least squares
optimization methods:
yt = f(Xt,b) + εt for t = 1, . . . , n.
^
Now call this b value b(1) as follows:
1. replacement form:
2. updating form:
y t = b1 + b2 Xt 2 + . . . + bK Xt K + εt for t = 1, . . . , n.
E u t = 0 , E u 2t = su2, E u t u s = 0 for s ≠ t.
Therefore, u t is nonautocorrelated and homoskedastic.
y t = b1 + b2 X t 2 + b3 X t 3 + εt for t = 1, . . . , n.
^ α^ ^α ^α ^ ^α
Given OLS estimates: α1 2 3 4 α5 6
^ − ^α3 − ^α
^ρ = α
^
ρ= ρ=
^ 5
^α2 ^α 4 6
∂ yt ∂ yt ∂ yt ∂ yt
f’(Xt,b) = [ ]
∂b1 ∂b2 ∂b 3 ∂ρ
∂ yt ∂ yt
= (1 − ρ) = (X t, 2 − ρ X t-1,2)
∂b1 ∂b2
∂ yt
= (X t, 3 − ρ X t-1,3)
∂b 3
∂ yt
= ( - b1 - b2Xt-1,2 - b3Xt-1,3+ y t-1 )
∂ρ
Copyright 1996 Lawrence C. Marsh
14.22
^β = [ f’(X,b(m))T f’(X,b(m))]-1 f’(X,b(m))T y∗(m)
(m+1)
Distributed
Lag Models
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh15.2
Economic action
at time t
Copyright 1996 Lawrence C. Marsh15.3
Unstructured Lags
yt = α + β0 xt + β1 xt-1 + β2 xt-2 + . . . + βn xt-n + et
Step 3: Define zt .
yt = α + γ zt + et
Copyright 1996 Lawrence C. Marsh15.8
Arithmetic Lag Structure
βi
β0 = (n+1)γ .
β1 = nγ .
β2 = (n-1)γ . linear
. lag
. structure
.
βn = γ .
0 1 2 . . . . . n n+1
i
Copyright 1996 Lawrence C. Marsh15.9
Polynomial Lag Structure
proposed by Shirley Almon (1965)
n = the length of the lag
the lag weights fit a polynomial
p = degree of polynomial
yt = α + γ0 z t0 + γ1 z t1 + γ2 z t2 + et
^ ‘s in terms of ^γ , ^γ , and ^γ .
Step 5: Express βi 0 1 2
β^0 = ^γ0
^
β1 = ^γ0 + ^γ + ^γ
1 2
^
β2 = ^γ +
0 2γ^ + 4γ
1
^
2
^
β3 = ^γ0 + ^ + 9γ
3γ 1
^
2
^
β 4 = ^γ +
0
^ + 16γ
4γ 1
^
2
Copyright 1996 Lawrence C. Marsh
15.13
Polynomial Lag Structure
βi β2
β0
. . .β
β1
3
. β4
.
0 1 2 3 4 i
Figure 15.3
Copyright 1996 Lawrence C. Marsh
15.14
Geometric Lag Structure
infinite distributed lag model:
yt = α + β0 xt + β1 xt-1 + β2 xt-2 + . . . + et
∞
yt = α + iΣ
=0
βi x t-i + et (15.3.1)
β0 = β
β1 = βφ
Substitute βi = β φi β2 = β φ2
β3 = β φ3
..
.
infinite geometric lag:
yt = α + β(xt + φ xt-1 + φ2 xt-2 + φ3 xt-3 + . . .) + et
Copyright 1996 Lawrence C. Marsh
15.16
Geometric Lag Structure
yt = α + β(xt + φ xt-1 + φ2 xt-2 + φ3 xt-3 + . . .) + et
impact multiplier :
β
β + β φ + β φ2
long-run multiplier :
β
β(1 + φ + φ2 + φ3 + . . . ) = 1− φ
Copyright 1996 Lawrence C. Marsh
15.17
Geometric Lag Structure
βi β0 = β .
geometrically
β1 = β φ . declining
β2 = β φ2 . weights
β3 = β φ3
β4 = β φ4
. .
0 1 2 3 4 i
Figure 15.5
Copyright 1996 Lawrence C. Marsh
15.18
Geometric Lag Structure
Problem:
How to estimate the infinite number
of geometric lag coefficients ???
Answer:
Use the Koyck transformation.
Copyright 1996 Lawrence C. Marsh
15.19
The Koyck Transformation
yt = δ1 + δ2 yt-1 + δ3xt + νt
Copyright 1996 Lawrence C. Marsh
15.21
The Koyck Transformation
yt = α(1− φ) + φ yt-1 + βxt + (et − φ et-1)
yt = δ1 + δ2 yt-1 + δ3xt + νt
^β = ^δ
The original structural 3
parameters can now be ^φ = ^δ
estimated in terms of 2
these reduced form
^α = ^δ / (1− ^δ )
parameter estimates. 1 2
Copyright 1996 Lawrence C. Marsh
15.22
Geometric Lag Structure
^ + φ^ x + φ^2 x + ^φ3 x + . . .) + ^e
yt = ^α + β(x t t-1 t-2 t-3 t
^ ^
β0 = β
^ ^ ^
β1 = β φ
^β = ^β ^φ2
2
^β = ^β ^φ3
3
.
.
.
^ ^ ^ ^
yt = α + β0 xt + β1 xt-1 + β2 xt-2 + β3 xt-3 + . . . + ^et
^
Copyright 1996 Lawrence C. Marsh
15.23
Durbin’s h-test
for autocorrelation
Estimates inconsistent if geometric lag model is autocorrelated,
but Durbin-Watson test is biased in favor of no autocorrelation.
T−1
h= 1− d
2 1 − ( T − 1)[se(b2)]2
yt = α + β x*t + et
adjust expectations
based on past realization:
rearrange to get:
or
yt = β1 + β2yt-1+ β3xt-1 + ut
^ ^
and we get:
β1 ^ β3
^
α= β=
^ ^ ^
λ = (1− β2)
^
(1− β2) (1− β2)
Copyright 1996 Lawrence C. Marsh
15.30
Partial Adjustment
y*t = α + β xt + et
Solving for yt :
yt = β1 + β2yt-1+ β3xt + νt
^ ^
^
^γ = (1− β β1 ^ β3
2) ^α = β=
^
^
(1− β2) (1− β2)
Copyright 1996 Lawrence C. Marsh16.1
Chapter 16
Time
Series
Analysis
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh16.2
1. autoregressive (AR)
δ is the intercept.
θ1 is parameter generally between -1 and +1.
et is an uncorrelated random error with
mean zero and variance σe 2 .
Copyright 1996 Lawrence C. Marsh16.8
δ is the intercept.
θi’s are parameters generally between -1 and +1.
et is an uncorrelated random error with
mean zero and variance σe 2 .
Copyright 1996 Lawrence C. Marsh16.9
positive
negative
2/ T
0
k
−2/ T
^y ^ ^ ^
T+1 = δ + θ y
1 T + θ2 yT-1
^y ^ ^ ^
T+2 = δ + θ1 T+1 + θ2 yT
y
^y ^ ^ ^
T+1 = δ + θ y
1 T + θ2 yT-1
µ is the intercept.
αi‘s are unknown parameters.
et is an uncorrelated random error with
mean zero and variance σe 2 .
Copyright 1996 Lawrence C. Marsh
16.15
An MA(1) process:
yt = µ + et + α1et-1 (16.2.2)
T T
S(µ,α1) = Σ et = t=1Σ(yt - µ - α1et-1)
2 2
(16.2.3)
t=1
Copyright 1996 Lawrence C. Marsh
16.16
Stationary vs. Nonstationary
stationary:
A stationary time series is one whose mean, variance,
and autocorrelation function do not change over time.
nonstationary:
A nonstationary time series is one whose mean,
variance or autocorrelation function change over time.
Copyright 1996 Lawrence C. Marsh
16.17
yt = z t - z t-1
2/ T
0
k
−2/ T
θ1 = 1 nonstationary process
∆zt = θ1zt -1 +
*
µ + et + α1et -1 (16.3.3)
H0: θ1 = 0
*
vs. H1: θ1 < 0
*
(16.3.4)
Dickey-Fuller Test
Phillips-Perron Test
Copyright 1996 Lawrence C. Marsh
16.26
1. extension of AR model.
2. all variables endogenous.
3. no structural (behavioral) economic model.
4. all variables jointly determined (over time).
5. no simultaneous equations (same time).
Copyright 1996 Lawrence C. Marsh
16.30
The random error terms in a VAR model
may be correlated if they are affected by
relevant factors that are not in the model
such as government actions or
national/international events, etc.
yt = θ0 + θ1yt-1 + φ1xt-1 + et
xt = δ0 + δ1yt-1 + α1xt-1 + ut
(continued)
Copyright 1996 Lawrence C. Marsh
16.37
Error Correction Model
θ0 =
*
θ0 + γ1β1 φ1 δ1
γ1 =
- α1
1 α1 - 1
β2 =
δ1
δ0
*
= δ0 + γ2β1 γ2 = δ1
Copyright 1996 Lawrence C. Marsh
16.38
Step 1:
Estimate by least squares:
yt-1 = β1 + β2 xt-1 + εt-1
to get the residuals:
^ε ^ ^
t-1 = yt-1 - β1 - β2 xt-1
Copyright 1996 Lawrence C. Marsh
16.39
Step 2:
∆yt = θ0 + γ1 ^ε
*
t-1 + et
∆xt = δ0 + γ2 ^ε
*
t-1 + ut
Copyright 1996 Lawrence C. Marsh
16.40
Guidelines for
Research Project
Copyright © 1997 John Wiley & Sons, Inc. All rights reserved. Reproduction or translation of this work beyond
that permitted in Section 117 of the 1976 United States Copyright Act without the express written permission of the
copyright owner is unlawful. Request for further information should be addressed to the Permissions Department,
John Wiley & Sons, Inc. The purchaser may make back-up copies for his/her own use only and not for distribution
or resale. The Publisher assumes no responsibility for errors, omissions, or damages, caused by the use of these
programs or from the use of the information contained herein.
Copyright 1996 Lawrence C. Marsh17.2
What Book Has Covered
ð Formulation
economic ====> econometric.
ð Estimation
selecting appropriate method.
ð Interpretation
how the xt’s impact on the yt .
ð Inference
testing, intervals, prediction.
Copyright 1996 Lawrence C. Marsh17.3
Topics for This Chapter
1. Types of Data by Source
2. Nonexperimental Data
3. Text Data vs. Electronic Data
4. Selecting a Topic
5. Writing an Abstract
6. Research Report Format
Copyright 1996 Lawrence C. Marsh17.4
Types of Data by Source
i) Experimental Data
from controlled experiments.
ii) Observational Data
passively generated by society.
iii) Survey Data
data collected through interviews.
Copyright 1996 Lawrence C. Marsh17.5
Time vs. Cross-Section
Micro Data:
data collected on individual economic
decision making units such as individuals,
households or firms.
Macro Data:
data resulting from a pooling or aggregating
over individuals, households or firms at the
local, state or national levels.
Copyright 1996 Lawrence C. Marsh17.7
Flow Data:
outcome measured over a period of time,
such as the consumption of gasoline during
the last quarter of 1997.
Stock Data:
outcome measured at a particular point in
time, such as crude oil held by Chevron in
US storage tanks on April 1, 1997.
Copyright 1996 Lawrence C. Marsh17.8
Quantitative Data:
outcomes such as prices or income that may
be expressed as numbers or some transfor-
mation of them (e.g. wages, trade deficit).
Qualitative Data:
outcomes that are of an “either-or” nature
(e.g. male, home owner, Methodist, bought
car last year, voted in last election).
Copyright 1996 Lawrence C. Marsh17.9
International Data
http://econwpa.wustl.edu/EconFAQ/EconFAQ.html