Professional Documents
Culture Documents
dr Anna Staszewska-Bystrova
General Information
I Materials:
1. Lecture notes
2. Adkins L. C., Using gretl for Principles of Econometrics 4th Edition,
free e-book, 2014 (available at:
www.learneconometrics.com/gretl.html)
I Software: Gretl
q s = f (p, p c , p f )
q s = f (p, p c , p f ) + ε
I The random error ε accounts for the many factors that affect supply
that we have omitted from this simple model, and it also reflects the
intrinsic uncertainty in economic activity.
I To complete the specification of the econometric model, we must
also say something about the form of the algebraic relationship
among our economic variables.
We can assume that the systematic part of the supply relation is linear
f (p, p c , p f ) = α0 + α1 p + α2 p c + α3 p f
q s = α0 + α1 p + α2 p c + α3 p f + ε
The Data
We’ll use the phrase statistical inference a lot. By this we mean we want
to ”infer” or learn something about the real world by analyzing a sample
of data. The ways in which statistical inference is carried out include:
I Estimating economic parameters, such as elasticities, using
econometric methods.
I Predicting economic outcomes, such as the enrollment in Lódź
University for the next 10 years.
I Testing economic hypotheses, such as the question of whether
newspaper advertising is better than store displays for increasing
sales.
Types of econometric Models
I Single equation
I Multi-equation
I Linear
I Nonlinear
yt = β0 + β1 x1t + β2 x2t + εt
To make this statistical model complete, assumptions about the
probability distribution of the random errors, εt , need to be made.
1. E (εt ) = 0. Each random error has a probability distribution with zero
mean.
2. var (εt ) = σ 2 . The variance σ 2 is an unknown parameter and it
measures the uncertainty in the statistical model. It is the same for each
observation. Errors with this property are said to be homoskedastic.
3. cov (εt , εs ) = 0. The covariance between the two random errors
corresponding to any two different observations is zero. Thus, any pair of
errors is uncorrelated.
4. We will sometimes further assume that the random errors εt have
normal probability distributions. That is, εt ∼ N(0, σ 2 ).
In addition to the above assumptions about the error term, we make two
assumptions about the explanatory variables.
I The first is that the explanatory variables are not random variables.
I The second assumption is that no explanatory variable is an exact
linear function of any of the other explanatory variables. This
assumption is equivalent to assuming that no variable is redundant.
As we will see, if this assumption is violated, a condition called
“exact multicollinearity”, the least squares procedure fails.
The statistical properties of yt follow from those of εt
1. E (yt ) = β0 + β1 x1t + β2 x2t . This assumption says that the average
value of yt changes for each observation and is given by the regression
function
2. var (yt ) = var (εt ) = σ 2
3. cov (yt , ys ) = cov (εt , εs ) = 0
4. yt ∼ N[(β0 + β1 x1t + β2 x2t ), σ 2 ] is equivalent to assuming that
εt ∼ N(0, σ 2 )
Multiple Regression in Matrix Notation
y = X β + ε,
yt = β0 + β1 xt + εt
1 1
X = ... ...
1 1
so that xt = 1. The matrix X has rank 1 for its 2 columns are identical.
It will be impossible to say anything about β0 and β1 separately, for the
model says
yt = β0 + β1 + εt
yt = 2 + εt
Because distinct values have identical implications for the behaviour of y
there is no possibility of learning from the value of y which of the β
combinations is the correct one.
Generally if X is of less than full rank then corresponding to the true
value β there will be a β ∗ (actually infinitely many) such that
X β = X β∗
∑
n
b =
S(β) (yt − (βb0 + βb1 x1t + ... + βbk xkt ))2
t=1
The algebra is best done in matrix notation. Write the sum of squares
b as
function S(β)
b = (y − X β)
S(β) b ′ (y − X β)
b
b is
based on the fact that the t-th component of the vector (y − X β)
∂S(β)b ∂ ′
= (y y − βb′ X ′ y − y ′ X βb + βb′ X ′ X β)
b
∂ βb ∂ βb
∂ ′
= (y y − 2βb′ X ′ y + βb′ X ′ X β)
b
∂ βb
= −2X ′ y + 2X ′ X βb
∂S(β)b
=0
∂ βb
gives a set of linear equations
−X ′ y + X ′ X βb = 0.
These equations are called the normal equations. When X has full rank
the matrix X ′ X is nonsingular and can be inverted. We can write
βb = (X ′ X )−1 X ′ y
When X does not have full rank, X ′ X is singular and there are infinitely
many solutions to the equation −X ′ y + X ′ X βb = 0.
Properties of the least squares estimators
βb = (X ′ X )−1 X ′ y = (X ′ X )−1 X ′ (X β + ε)
= (X ′ X )−1 X ′ X β + (X ′ X )−1 X ′ ε
= I β + (X ′ X )−1 X ′ ε = β + (X ′ X )−1 X ′ ε
w = c + Az
E βb = β + (X ′ X )−1 X ′ E ε = β
where I have used the assumptions concerning var (ε), the rule for the
transpose of a product and exploited the fact that (X ′ X )−1 is symmetric.
I Summing up, we have established two points
E βb = β ; var (β)
b = σ 2 (X ′ X )−1