Professional Documents
Culture Documents
1
1 (e) Consider the case where X is n × (p + 1) with a constant term
and p independent variables defining the regression model, i.e.,
1 x1,1 x1,2 · · · x1,p
1 x2,1 x2,2 · · · x2,p
X= .
.. .. .. .
.. . . . .
.
1 xn,1 xn,2 · · · xp,n
G as follows:
Define
1 −x̄1 −x̄2 · · · −x̄p
0
1 0 ··· 0
G=
0 0 1 ··· 0
.. .
. . . . .
. . . . 0
0 0 ··· 1
Pn
where x̄j = i=1 xi,j /n, for j = 1, 2, . . . , p.
By (d), the regression model with X 0 = XG is equivalent to
the original regression model in terms of having the same fitted
values ŷ and residuals ˆ
• If β = (β0 , β1 , . . . , βp )T is the regression parameter for X,
show that
β 0 = G−1 β is the regression parameter for X 0 .
Solve for G−1 and provide explicit formulas for the elements
of β 0 .
• Show that:
0Tp
0T 0 n
[X X ] =
0p X T X
x1,1 − x̄1 x1,2 − x̄2 · · · x1,p − x̄p
x2,1 − x̄1 x2,2 − x̄2 · · · x2,p − x̄p
where X = .. .. .. .
. . . ..
xn,1 − x̄1 xn,2 − x̄2 · · · xp,n − x̄p
• Prove the following formula for elements of the projection/hat
matrix:
Hi,j = n1 + (xi − x̄)T [X T X ]−1 (xj − x̄)
where xi = (xi,1 , xi,2 , . . . xi,p )T is the vector of independent
variable values for case i, and x̄ = (x̄1 , x̄2 , . . . , x̄p )T .
The leverage of case i, Hi,i , increases with the second term,
the squared M ahalanobis distance between xi and the mean
vector x̄.
2
Case Deletion Influence Measures
3
A common issue arising in a regression analysis with p explanatory
variables is whether just the first k (< p) explanatory variables (given
by the first k columns of X) enter in the regression model. This can
be expressed as an hypothesis about the regression parameter β,
H0 : βk+1 = βk+2 = · · · = βp ≡ 0.
βˆI
3 (a) Consider the estimate βˆ0 = where
0p−k
β̂I = (X T −1 T
I XI ) XI y
XI = X[1] X[2] · · · X[k]
Show that βˆ0 is the constrained least-squares estimate of β cor-
responding to the hypothesis H0 , i.e.,
β̂0 minimizes: SS(β) = (y − Xβ)T (y − Xβ)
subject to
β̂j = 0, j = k + 1, k + 2, . . . , p.
3 (b) Show that the QR-decomposition of XI is XI = QI RI , where QI
is the matrix of the first k columns of Q and RI is the upper-left
k × k block of R. Furthermore, verify that:
β̂I = RI−1 QTI y, and
ŷI = HI y,
where HI = QI QTI , the n × n projection/Hat matrix under the
null hypothesis.
3 (c) From the lecture notes, recall the definition of
QT
A= , where
WT
• A is an (n × n) orthogonal matrix (i.e. AT = A−1 )
• Q is the column-orthonormal matrix in a Q-R decomposition
of X
Note: W can be constructed by continuing the Gram-Schmidt
Orthonormalization process (which was used to construct Q
from X) with X ∗ = [ X I n ].
Then, consider
T
Q y zQ (p × 1)
z = Ay = T =
W y zW (n − p) × 1
Prove the following relationships for the unconstrained regression
model:
4
• yT y = y12 + y22 + · · · + yn2
= z12 + z22 + · · · + zn2
= zT z
T
• ŷ ŷ = ŷ12 + ŷ22 + · · · + ŷn2
= z12 + z22 + · · · + zk2 + · · · + zp2
• ˆT ˆ = ˆ21 + ˆ22 + · · · + ˆ2n
2
= zp+1 2
+ zp+2 + · · · + zn2
Prove the following relationships for the constrained regression
model:
• y T y = y12 + y22 + · · · + yn2
= z12 + z22 + · · · + zn2
= zT z
• ŷIT ŷI = (ŷI )21 + (ŷI )22 + · · · + (ŷI )2n
= z12 + z22 + · · · + zk2
• ˆT ˆI = ˆ(I )2 + ˆ(I )2 + · · · + ˆ(I )2
I 1 2 n
2
= zk+1 2
+ · · · + zp2 + zp+1 2
+ zp+2 + · · · + zn2
3 (d) Under the assumption of a normal linear regression model, the
lecture notes
detail ho
w the distribution
is
of z = Ay
zQ Rβ
z= ∼ Nn , σ2I n
zW O n−p
=⇒
z Q ∼ Np [(Rβ), σ 2 I p ]
z W ∼ N(n−p) [(O (n−p) , σ 2 I (n−p) ]
and z Q and z W are independent.
• For the unconstrained (and the constrained) model, deduce
that:
SSERROR = ˆT ˆ ∼ σ 2 × χn−p 2
5
SSREG(k+1,...,p|1,2,...,k) /(p−k)
Fˆ = SSERROR /(n−p)
ANOVA Table
ŷ T ŷ−ŷIT yˆI MS
Regression on ‘k + 1, . . . , p’ ŷ T ŷ − ŷIT ŷI (p-k) M S0 = (p−k) F̂ = M S 0
Error
Adjusting for ‘1, 2, . . . , k’
ˆT ˆ
Error ˆT ˆ (n − p) M SError = (n−p)
Total yT y n
6
MIT OpenCourseWare
http://ocw.mit.edu
For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.