Professional Documents
Culture Documents
Under OLS estimation given in the above simple linear regression, there is constant term α,
therefore we should take in to account that the value of one (1) in the first column of X value,
since the assumption of OLS estimation the sum of error term is equal to zero(0).
x1
[1
]x
[1 [ ¿ 2 ]¿
[ . .. ]. . .
[1 ] xn
Let ¿
i
i 0 and x
i
i 0
A Show that the least squares normal equations simply
X ' (Y −X β^ )=0
X ' e=0
For every column of xk of X, this implies xk'e = 0. If the first column of X is a column of
n
x '1 e=i' e= ∑ ei =0
1s, then, the least squares residuals sum to zero. this follows from i=1
n
∑ x i e i=0
and also imply i=1
n
∑ e i =0
i=1
Using the least square sum of residuals to zero from the first normal equation
e= y i −α−βx i
n n
∑ e i =∑ ( y i −α−βx i )=0
i=1 i=1
n n
∑ y i =∑ α + β ∑ x i
i=1 i=1
∑ y i =nα+β ∑ x i
∑ yi =α+β
n
∑ xi
ȳ=α+β x̄ ⇒α= ȳ −β x̄
n n
∑ x̄ ei = x̄ ∑ ei =0 ∑ (x i − x̄ )( y i−α−bx i )
Since i=1 i=1 ,Substituting ei=yi--bxi to get
or
From question b above we have got ȳ=α+β x̄ from this α= ȳ−b x̄ where b= then
∑ (x i − x̄ )( y i− ȳ−b( x i − x̄ ))=0
substitute it on α
n
∑ ( xi − x̄ )( y i − ȳ )
n n b= i=1 n
∑ ( x i − x̄ )( y i− ȳ )=b ∑ ( x i− x̄ )( x i− x̄ )) ∑ ( x i − x̄ )2
i=1 i=1 so i =1
D Prove that these two values uniquely minimize the sum of squares by showing that the
diagonal elements of the second derivatives matrix of the sum of squares with respect to the
parameters are both positive and that the determinant is
i =1
∑ xi 2∑ xi
[n [ 2n
[¿ n ]¿ 2 X ' X=¿ [¿ ]¿
[ ∑ xi ] ∑ x 2i [2 ∑ x i ]2 ∑ xi2
i=1
X’X= ¿ and ¿
The diagonal elements are 2n and 2 ∑ x2i which clearly shows both of them are positive.
n 2
=
4n
[(∑ ) ]
i=1
x i − x̄
this the result that we want.
2. Suppose that b is the least squares coefficient vector in the regression of y on X and that C is
any other Kx1 vector. Prove that the difference in the two sums of squared residuals is
3.Explain what effect the following problems have on the properties of the least squares
estimates of the coefficients and their standard errors. How would you detect whether each
problem was present:
The formulas used by classical ordinary least squares (OLS) regression to estimate the
population parameters in a regression model will be unbiased, be efficient, have minimum mean
square error (MSE) and be consistent, if the following assumptions hold true:
1. The model is correctly specified, e.g., all relevant explanatory variables are included in the
regression.
If the above assumptions are “violated” the classical regression formulas(OLS) may not be
unbiased, efficient, have minimum mean square error (MSE) or be consistent.
a.Heteroscedasticity.
Heteroscedasticity occurs when the error variance has non-constant variance. In this case, we can
think of the disturbance for each observation as being drawn from a different distribution with a
different variance. Stated equivalently, the variance of the observed value of the dependent
variable around the regression line is non-constant. We can think of each observed value of the
dependent variable as being drawn from a different conditional probability distribution with a
different conditional variance. A general linear regression model with the assumption of
heteroscedasticity can be expressed as follows:
Yi = 1 + 2 Xi2 + … + k Xik + εi
If the error term has non-constant variance, then the consequences of using the OLS estimator to
obtain estimates of the population parameters are:
3. The estimated variances and covariances of the OLS estimates are biased and inconsistent.
Detection of heteroscedasticity
(b) Serial correlation.
Autocorrelation occurs when the errors are correlated. In this case, we can think of the
disturbances for different observations as being drawn from different distributions that
are not explanatory distributions.
Consequences of Autocorrelation on OLS and Standard error
The consequences are the same as heteroscedasticity. That is:
1. The OLS estimator is still unbiased.
2. The OLS estimator is inefficient; that is, it is not BLUE.
3. The estimated variances and covariances of the OLS estimates are biased and
inconsistent. If there is positive autocorrelation, and if the value of a right-hand side
variable grows over time, then the estimate of the standard error of the coefficient
estimate of this variable will be too low and hence the t-statistic too high.
4. Hypothesis tests are not valid.
Detection of autocorrelation.
There are several ways to use the sample data to detect the existence of autocorrelation.
Plot the residuals
One way to detect autocorrelation is to estimate the equation using OLS, and then plot the
residuals against time.
The Durbin-Watson d test: The most often used test for first-order autocorrelation is the
Durbin-Watson d test. It is important to note that this test can only be used to test for
first-order autocorrelation, it cannot be used to test for higher-order autocorrelation. Also,
this test cannot be used if the lagged value of the dependent variable is included as a
right-hand side variable.
The Breusch-Godfrey Lagrange Multiplier Test: The Breusch-Godfrey test is a general
test of autocorrelation. It can be used to test for firstorder autocorrelation or higher-order
autocorrelation. This test is a specific type of Lagrange multiplier test.
Economists usually test for positive autocorrelation because negative serial correlation is
highly unusual when using economic data. The null and alternative hypotheses are
H 0: 1 = 2 = 0 vs H 1: At least one is not zero
(f) Exact multicollinearity: One of the assumptions of CLR model is that there are
no exact linear relationships between the independent variables and that there are at
least as many observations as the dependent variables (Rank of the regression). If
either of these is violated it is impossible to estimate OLS and the estimating
procedure simply breaks down.
Consequences of multicollinearity
1. Although BLUE, the OLS estimators have larger variances making precise
estimation difficult. OLS are BLUE because near collinearity does not affect the
assumptions made.
2. The confidence intervals tend to be much wider, leading to the acceptance of the
null hypothesis
4. The OLS estimators and their standard errors could be sensitive to small changes in
the data.
Detection of multicollinearity.
(d) Nonnormality
If the error terms are not normally distributed, inferences about the regression
coefficients (using t-tests) and the overall equation (using the F-test) will become
unreliable. However, as long as the sample sizes are large (namely the sample size
minus the number of estimated coefficients is greater than or equal to 30) and the
error terms are not extremely different from a normal distribution, such tests are
likely to be robust. Whether the error terms are normally distributed can be assessed
by using methods like the normal probability plot. The formal tests to detect non-
normal errors one can estimate the values of skewness and kurtosis. These values can
be obtained from the descriptive statistics.
4.Suppose you face the demand curve Q p . In the past, you have set the following
prices and sold the accompanying quantities
Suppose that your marginal cost is 10. Based on the least squares regression, compute a 95%
confidence interval for the expected value of the profit maximizing output.
Solution
Since the mean value of error term is zero, that is E(ε) =0 .The expected value of the demand
function,E(Q)=q=+P or the expected value of the inverse demand function, E(P)=-/+(1/)q
=p.then total revenue is TR= QP=q(-/+(1/)q)=(- /)q+(1/)q2. From this by differentiating
total revenue with respect to Q a linear demand curve, marginal revenue is MR = d(pq)/dq=(-α/β)
+ (2/β)q. The profit maximizing output is that at which marginal revenue equals marginal cost, or
α 2q α
− + =10⇒ q eqm= +5 β
10. Equating MR to 10 and solving for q produces MR=MC , β β 2
, q at equilibrium ,so we require a confidence interval for this combination of the parameters.
α 20 .769
q eqm= +5 β= −5∗0 .841=6 .182
2 2
Now ^ β^ P and
ε^ =Q−α− ∑ ^ε2=204 .6142
^2
^δ 2= ∑ ε = 204 .6142 =15 .74
N−2 15−2
2
(11. 067 )2
var( α^ )=δ
1
+2 p
[
N ∑ ( P−p )2
1
=15. 74∗ +
]
15 278. 93 [
= 7. 96 ]
2
δ 15 . 74
var( β^ )= = = 0 . 056436
∑ ( P− p )2 278 . 93
The estimated covariance matrix of the
2 '
coefficients is obtained as δ P P=−0 .624559
−0 . 6246
[ 7 . 96 [¿ ]¿
[−0 . 6246 ] 0 . 056436
¿
The estimate of the variance of q^ is (1/4)7.96 + 25(.056436) + 5(-.06246) or 0.278415, so the
estimated standard error is 0.5276
The degree of freedom for this data is N-2=15-2 =13 .Thus the 95% cutoff value for a t
distribution with 13 degrees of freedom is 2.161.
eqm
q ±t α SE ( q )
, n−2
2
5 For the classical normal regression model y p with no constant term and K
regressors, assuming that the true value of is zero, what is the exact expected value of
R2
F K, N K k
1 R 2
N K
b' X ' Xb
K
ee e
n−K
e=Y − Xb
=Y −X ( X ' X )−1 X ' Y
¿ [( I−X ( X ' X )−1 X ' ) ] Y
¿ MY =M ( Xβ+ε )=MX β + Mε=Mε
¿ ε ' M ' Mε=ε ' M ' Mε=ε ' Mε
Where
y1t and y2 t are endogenous, x1t x2 t and x3t are exogenous variables, and
1t and 2 t
are
NID 0,
random disturbances.
a. Discuss the identifiability of each equation of the system in terms of the order and
rank conditions for identification.
b. Explain why the ordinary least squares estimator of
12 , 11 is inconsistent
c. What are the two-stage least squares estimators of the coefficients in the two
equations? Describe the procedure step by step
Solution :a Order condition for the identification of equation j
where K*j is the number of exogenous variables excluded from equation j and Mj is the number
of endogenous variables included in equation j
know from our two simultaneous equations the first equation over identified since there
are two K*1 which are X2t and X3t and one M1 which is Y2 (that is K*j>Mj , 2>1).
Therefore the equation is identified, we call it over identified
for the second equation since K*j=Mj the equation identified.
Inshort :In this model y1t and y2t are endogenous and X1t X2t and X3t are exogenous
(predetermined). The first equation excludes two variables X 2t and X3t and hence overidentified
by the order condition.however the second equation excludes exactly one variable X1t , and hence
by the order condition it is exactly identified. Therefore, the model as a whole is
identified. The order condition is only a necessary condition not sufficient condition for
identification.
rank [ Π ¿j ]= M j
Consider system of simultaneous equations
1 1 -12 -11 0 0