Professional Documents
Culture Documents
Any estimation method has certain quality requirements which depend on certain
conditions being met. As regards the ordinary least squares method (OLS
method), certain estimation properties are known, provided that the standard
assumptions of the multiple regression model are met. If the disturbance variable on
average is equal to 0 and its variance is constant, then the OLS estimators for the
regression coefficients are best linear unbiased estimators (BLUE) in the
absence of autocorrelation, with appeal to the Gauss-Markov theorem.
In addition, the OLS method is consistent. Consistency means that with increasing
sample size the estimated regression coefficients better and better approach the
unknown population regression coefficients.
● Unbiasedness
follows.
Now the expected value of β̂ can be determined:
E(βˆ ) E β ( X'X)1 X'u) E(β) E ( X'X)1 X'u
β E( X'X) 1 X' u
With E (u) = 0 one finally obtains
(2.34) E(βˆ ) β
This proves the unbiasedness of the OLS estimator β̂ . Making use of the
extensive notation of (2.34)
E(ˆ 1 ) 1
ˆ
E( 2 ) 2
,
E(ˆ )
k k
it can be seen that the mean values of the OLS estimators ˆ 1, ˆ 2 , ..., ˆ k
conform with the unknown regression coefficients of the econometric model.
● Efficiency
● Consistency
When an estimator is required to increasingly approach the values of a
parameter vector with increasing sample sizes, then is in particular its limiting
behavior for n -> ∞ of interest, i.e. a property relating to large samples. In
looking for consistency we are concerned with the question as to whether an
estimator will finally coincide with the parameter vector β if sample size
increases beyond all bounds:
lim E(βˆ ) β and lim Cov(βˆ ) 0
n n
2.4 Coefficient of determination and multiple correlation coefficient
● Coefficient of determination
- Evaluates global goodness of fit of the regression model
Variance decomposition:
with
1 1 1
(2:38) s2ŷ ( ŷ t y)2 ŷ2t y2 yˆ 'yˆ y2 because of yˆ y
n n n
(2:39) s 2û
1 2 1
û t uˆ 'uˆ because of uˆ 0
n n
Coefficient of determination:
s 2ŷ s 2û
2 2
(2.40a) R or (2.40b) R 1 2
s 2y sy
Value range: 0 ≤ R2 ≤ 1
Formulas:
or
y t 24034,68
X'y
y'y y 2t 30778325
t t
x y 33960226,83
we calculate
ˆ ˆ βˆ 'X'y 30777206
y'y
_
With n = 19 and y 1264,98 the coefficient of determination is then given by:
- Simple Regression:
(2.42) R 2 ryx
2
s 2ŷ ˆ 22 s 2x
In the simple regression model, the square root of the coefficient of determination
therefore depicts the absolute value of the coefficient of correlation between
the regressand y and the regressor x:
(2.43) ryx R 2
- Multiple Regression:
The coefficient of determination is then the square root of the multiple correlation
coefficient ryŷ, indicating the correlation between the dependent variable y and
all independent variables x1, x2, ..., xk (=correlation between the dependent variable y t
and the regression values ŷ t ).
(2:44) R 2 ry2ŷ r 2
yx1, 2,...k
The sign of the multiple correlation coefficient ryx1, 2,...k remains undefined,
since positive and negative effects of the regressors can occur simultaneously:
One can avoid this by taking account of the number of degrees of freedom in the
calculation of the coefficient of determination. In contrast to R² the adjusted
coefficient of determination
1
uˆ ' uˆ
2
R 1 nk
(2.46)
1
y ' y ny 2
n 1
can decline in the case of the inclusion of irrelevant regressors.
(2:47) R2 1
n 1
nk
1 R2