Professional Documents
Culture Documents
Irene Cozzolino
Ass.1 −→ It says that the expected value of the error term is zero,
which means that, on average, the regression line should be correct.
Under Ass.1 it holds that
0
E (yi ) = xi β
It is still possible to perform the linear regression but there can be some
problems −→ instability of estimates & huge standard errors.
0 −1 P
P 0 0
N N
β̂ OLS = i=1 xi xi i=1 xi yi −→ β̂ OLS = (X X)−1 X y
0
Fitted values: ŷi = xi β̂ OLS .
E (β̂ OLS ) = β
N
!−1
X 0
0 −1
V (β̂ OLS ) = σ 2 xi xi = σ2 X X (13)
i=1
β̂ MLE = β̂ OLS .
2 = N1 N ˆ2i .
P
σ̂MLE i=1
Most important property:
0
β̂ MLE ∼ N(β, σ 2 (X X)−1 )
(
H0 : βj = 0 −→ The j-th regressor is not statistically significant in explaining the response variable
H1 : βj 6= 0 −→ The j-th regressor is statistically significant
(
H0 : β1 = β2 = · · · = βK = 0 −→ None of the K regressors is statistically significant
H1 : ∃j > 0 : βj 6= 0 −→ Exist at least one regressor, different from the intercept, stat. significant
p-value
p > α −→ Accept H0
p < α −→ Reject H0
We can then select the best model out of all of the models that we
have considered.
Various statistics can be used to judge the quality of a model, these
include Akaike information criterion (AIC).
Unfortunately, there are a total of 2K models that contain subsets of
K variables. This means that even for moderate K, trying out every
possible subset of the predictors is infeasible. For instance, we saw
that if K = 2, then there are 22 = 4 models to consider. But if
K = 30, then we must consider 230 = 10 0730 7410 824 models! This is
not practical.
It is natural to quantify the extent to which the model fits the data.
R 2 Statistic: It provides a measure of goodness of fit of the model.
It takes the form of a proportion - the proportion of variance
explained - and so it always takes on a value between 0 and 1.
R 2 measures the proportion of variability in y that can be explained
using X. An R 2 statistic that is close to 1 indicates that a large
proportion of the variability in the response has been explained by the
regression. A number near 0 indicates that the regression did not
explain much of the variability in the response; this might occur
because the linear model is wrong.
However, it can still be challenging to determine what is a good R 2
value.
Estimate the parameters of each of the models and calculate the Rj2 .
Rj2 = 1 −→ Xj is a perfect linear combination of the remaining K − 1
variables. We have a problem of collinearity. We should delete it since
its information is perfectly contained in the other K − 1 variables.
Rj2 ≥ 0.8 −→ approximate collinearity.
Rj2 = 0 −→ No problems of collinearity.
1
VIFj = 1−Rj2
. Usually, those variables with VIFj > 10 are deleted
from the model.