Professional Documents
Culture Documents
𝑌𝑖 = 𝛽1 + 𝛽2 𝑋𝑖 + 𝜇𝑖 … … … (1)
𝑌𝑖 = 𝛽̂1 + 𝛽̂2 𝑋𝑖 + 𝜇̂ 𝑖
𝑌𝑖 = 𝑌̂𝑖 + 𝜇̂ 𝑖
𝜇̂ 𝑖 = 𝑌𝑖 − 𝑌̂𝑖
Now differentiating (3) partially with respect to 𝛽̂1 and 𝛽̂2 we obtain
𝛿
∑ 2
̂1 ( 𝜇̂ 𝑖 )
𝛿𝛽
= −2 ∑(𝑌𝑖 − 𝛽̂1 − 𝛽̂2 𝑋𝑖 )……………………………(4)
𝛿
∑ 2
̂2 ( 𝜇̂ 𝑖 )
𝛿𝛽
= −2 ∑(𝑌𝑖 − 𝛽̂1 − 𝛽̂2 𝑋𝑖 )𝑋𝑖 ………………………..(5)
⇒ ∑ 𝑌𝑖 − ∑ 𝛽1 − ∑ 𝛽̂2 𝑋𝑖 =0
⇒ ∑ 𝑌𝑖 − 𝑛𝛽̂1 − 𝛽̂2 ∑ 𝑋𝑖 =0
1. The OLS estimators are expressed solely in terms of the observable quantities. Therefore, they can
be easily computed.
2. They are point estimators
3. Once the OLS estimates are obtained from the sample data, the regression line can be easily
obtained. The regression line thus obtained has the following properties:
i. It passes through the sample means of Y and X
ii. The mean value of the estimated 𝑌 = 𝑌̂𝑖 is equal to the mean value of the actual Y for
𝑌̂𝑖 = 𝛽̂1 + 𝛽̂2 𝑋𝑖
= (𝑌̅ − 𝛽̂2 𝑋̅) + 𝛽̂2 𝑋𝑖
= 𝑌̅ + 𝛽̂2 (𝑋𝑖 − 𝑋̅)
Summing both sides of this last equality over the sample values and dividing through by n give
𝑌̂𝑏𝑎𝑟 = 𝑌̅
Linearity:
𝑌𝑖 = 𝛽1 + 𝛽2 𝑋𝑖 + 𝜇𝑖 … … … (1)
𝑌𝑖 = 𝛽̂1 + 𝛽̂2 𝑋𝑖 + 𝜇̂ 𝑖
∑ 𝑋𝑖 ∑ 𝑌𝑖
∑ 𝑌𝑖 𝑋𝑖 − ∑(𝑋𝑖 − ̅̅̅
𝑋)(𝑌𝑖 − 𝑌̅) ∑ 𝑦𝑖 𝑥𝑖
𝛽̂2 = 𝑛
2 = ∑(𝑋𝑖 − ̅̅̅
= … … … . (2)
2 (∑ 𝑋𝑖 ) 𝑋)2 ∑ 𝑥𝑖2
∑ 𝑋𝑖 −
𝑛
𝛽̂1 = 𝑌̅ − 𝛽̂2 𝑋̅
∑ 𝑦𝑖 𝑥𝑖
𝛽̂2 =
∑ 𝑥𝑖2
∑(𝑌𝑖 − 𝑌̅)𝑥𝑖
𝛽̂2 =
∑ 𝑥𝑖2
∑ 𝑌𝑖 𝑥𝑖 ∑ 𝑌̅𝑥𝑖
𝛽̂2 = −
∑ 𝑥𝑖2 ∑ 𝑥𝑖2
∑ 𝑌𝑖 𝑥𝑖
𝛽̂2 = = ∑ 𝑘𝑖 𝑌𝑖 … … … . (3)
∑ 𝑥𝑖2
The OLS estimator 𝛽̂2 is the linear function of the sample values 𝑌𝑖
Some properties of 𝒌𝒊 :
𝑥𝑖
i. ∑ 𝑘𝑖 = ∑ =0
∑ 𝑥𝑖2
𝑥𝑖 ∑ 𝑥𝑖 2 1
ii. ∑ 𝑘𝑖2 = ∑( 2
2) = 2 =∑
∑ 𝑥𝑖 (∑ 𝑥𝑖 )2 𝑥𝑖 2
𝛽̂2 = ∑ 𝑘𝑖 𝑌𝑖
= ∑ 𝑘𝑖 𝛽1 + 𝛽2 ∑ 𝑘𝑖 𝑋𝑖 + ∑ 𝑘𝑖 𝜇𝑖
= 𝛽1 ∑ 𝑘𝑖 + 𝛽2 ∑ 𝑘𝑖 𝑋𝑖 + ∑ 𝑘𝑖 𝜇𝑖
= 𝛽2 + ∑ 𝑘𝑖 𝜇𝑖
̂ 𝟏:
Unbaisedness of 𝜷
𝑌𝑖 = 𝛽1 + 𝛽2 𝑋𝑖 + 𝜇𝑖
⇒ ∑ 𝑌𝑖 = ∑ 𝛽1 + 𝛽2 ∑ 𝑋𝑖 + ∑ 𝜇𝑖
⇒ ∑ 𝑌𝑖 = 𝑁𝛽1 + 𝛽2 ∑ 𝑋𝑖 + ∑ 𝜇𝑖
∑ 𝑌𝑖 ∑ 𝑋𝑖 ∑ 𝜇𝑖
⇒ = 𝛽1 + 𝛽2 +
𝑁 𝑁 𝑁
⇒ 𝑌̅ = 𝛽1 + 𝛽2 𝑋̅ … … … … . . (2)
𝛽̂1 = 𝑌̅ − 𝛽̂2 𝑋̅
= 𝛽1 + 𝛽2 𝑋̅ − 𝛽̂2 𝑋̅
= 𝛽1 + (𝛽2 − 𝛽̂2 )𝑋̅
̂ 𝟐:
Variance of 𝜷
We know that
2
𝑉(𝛽̂2 ) = 𝐸[𝛽̂2 − 𝐸(𝛽̂2 )]
2
= 𝐸[𝛽̂2 − 𝛽2 ]
2
= 𝐸 [∑ 𝑘𝑖 𝜇𝑖 ]
= 𝐸(𝑘1 𝜇1 + 𝑘2 𝜇2 + ⋯ … … … +𝑘𝑛 𝜇𝑛 )2
𝜎2
𝑉(𝛽̂2 ) = 𝜎 2 ∑ 𝑘𝑖2 =
∑ 𝑥2
𝑖
̂ 𝟏:
Variance of 𝜷
𝜎2
𝑉(𝛽̂1 ) = 𝑉(𝑌̅ − 𝛽̂2 𝑋̅) = 𝑉(−𝛽̂2 𝑋̅ + 𝑌̅) = 𝑋̅ 2 𝑉(𝛽̂2 ) = 𝑋̅ 2
∑ 𝑥2
𝑖
𝛽̂2 = ∑ 𝑘𝑖 𝑌𝑖
Which shows that 𝛽̂2 is a weighted average of the Y’s, with 𝑘𝑖 serving as the weights.
𝛽2∗ = ∑ 𝑤𝑖 𝑌𝑖
Where 𝑤𝑖 are also weights, not necessarily equal to 𝑘𝑖 . Now
𝐸(𝛽2∗ ) = ∑ 𝑤𝑖 𝐸(𝑌𝑖 )
= ∑ 𝑤𝑖 (𝛽1 + 𝛽2 𝑋𝑖 )
= 𝛽1 ∑ 𝑤𝑖 + 𝛽2 ∑ 𝑤𝑖 𝑋𝑖
∑ 𝑤𝑖 = 0 and ∑ 𝑤𝑖 𝑋𝑖 = 1
𝑉(𝛽2∗ ) = 𝑉 (∑ 𝑤𝑖 𝑌𝑖 )
= ∑ 𝑤𝑖2 𝑉(𝑌𝑖 )
= 𝜎 2 ∑ 𝑤𝑖2
𝑥𝑖 𝑥𝑖 2
= 𝜎 2 ∑(𝑤𝑖 − + )
∑ 𝑥2
𝑖 ∑ 𝑥 2
𝑖
2
𝑥𝑖 2
𝑥𝑖 𝑥𝑖 𝑥𝑖
= 𝜎 2 ∑(𝑤𝑖 − ) + 𝜎2 ∑ ( ) + 2 𝜎 2 ∑ (𝑤𝑖 − )( )
∑ 𝑥2
𝑖
2
∑ 𝑥𝑖 2
∑ 𝑥𝑖 ∑ 𝑥2
𝑖
𝑥𝑖 2 2 1
= 𝜎 2 ∑(𝑤𝑖 − ) + 𝜎 ( 2)
∑ 𝑥2
𝑖 ∑ 𝑥𝑖
𝑥𝑖 2
𝑉(𝛽2∗ ) = 𝜎 2 ∑(𝑤𝑖 −
2
) + 𝑉(𝛽̂2 )
∑ 𝑥𝑖
𝑉(𝛽2∗ ) ≥ 𝑉(𝛽̂2 )
𝛽̂2 = ∑ 𝑘𝑖 𝑌𝑖
Since X’s are assumed to be fixed, or non stochastic, and since 𝑌𝑖 = 𝛽1 + 𝛽2 𝑋𝑖 + 𝜇𝑖 we can write,
𝛽̂2 = ∑ 𝑘𝑖 (𝛽1 + 𝛽2 𝑋𝑖 + 𝜇𝑖 )
Because, k. the beats and X are all fixed, 𝛽̂2 is ultimately a linear function of the random variable 𝜇𝑖 , which
is random by assumption. Therefore, the probability distribution of 𝛽̂2 will depend on the assumption made
about the probability distribution of 𝜇𝑖 .
𝜇𝑖 ~𝑁(0, 𝜎 2 )