0% found this document useful (0 votes)
323 views12 pages

OLS Method

Uploaded by

aditidocmoc
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
323 views12 pages

OLS Method

Uploaded by

aditidocmoc
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd

Study Notes

Ordinary Least Square


(OLS) Method
OLS Method

Ordinary Least Squares (OLS) Method

 Minimises the sum of the squares of the error term.


∑ ᷇ μ 2i =∑ ¿
 After minimization, you derive the normal equations.
 Using the normal equations, regression coefficients can be estimated.
 It is also possible to find the regression equations by using the following equation:
Y −Y =b YX ¿) or
X −X =b XY ¿)

Normal Equations

2
OLS Method

Derivation of Formulas for ^β 0 and ^β 1

3
OLS Method

4
OLS Method

Drawback of OLS Method

 Gives equal weightage to all the error term.


∑ ᷇ μ 2i =∑ ( μ21 + μ22 + …. + μ2n ¿2
 However, points closer to the SRF should get more weightage as compared to points
which are far away because the points which are close have contributed the most in the
construction of the SRF.
 This limitation is overcome by Weighted Least Square Method.

Properties of Sample Regression Line underlying OLS Method

1. It passes through the sample means of X and Y.


∑ Y i=n β^ 0 + β^ 1 ∑ X i (from first normal equation)
Divide both sides by n,

^β =Y − ^β X
n ( )
∑Y i ^ ^ ∑ X i
= β 0+ β 1
n
Y = ^β0 + β^ 1 X ⇒ 0 1

2. Mean value of estimatedY =Y^ i is equal to the mean value of actual Y.


Y^ i= β^ 0 + β^ 1 X i
¿ ( Y − ^β 1 X ) + ^β 1 X i
¿ Y + ^β ( X −X )
1 i

∑ Y^ i ∑ Y Σ ( X i−X )
= + ^β 1
n n n
Y^ =Y [ since , ∑ ( X i −X ) =0 ] .
3. Mean value of residuals is 0.

5
OLS Method

−2 ∑(Y i − ^β 0− ^β 1 ∑ X i)=0 [from first normal equation]


∑ ^μi=0
Hence ,
^μ=0
4. The residuals are uncorrelated with the predicted Y.
Σ 1 u^ i Y^ i=Σ ^μi ( ^β 0 + ^β 1 X i )
¿ ^β Σ ^μ + ^β Σ μ^ X
0 i 1 i i

¿ ^β 0 ( 0 ) + ^β 1 ( 0 ) [from normal equations]


¿0
5. The residuals are uncorrelated with X.
Σ μ^ i X i=0 [from second normal equation]

Gauss-Markov Theorem

 In the class of all linear and unbiased estimators of β, the OLS estimator ^β has the minimum
variance. In short, ^β is BLUE (best linear unbiased estimator).
 It is linear, that is, a linear function of a random variable, such as the dependent variable
Y in the regression model.
 It is unbiased, that is, its average or expected value, E ( ^β ), is equal to the true value, β.
 It has minimum variance in the class of all such linear unbiased estimators; an unbiased
estimator with the least variance is known as an efficient estimator.
 To see this, look at Precision or Standard Errors of Least Square Estimates

Precision or Standard Errors of least square estimates

6
OLS Method

7
OLS Method

8
OLS Method

9
OLS Method

Where Sxy denote the S.E. of estimate of regression equation of x on y.

Degrees of freedom

10
OLS Method

 The term number of degrees of freedom means the total number of observations in the
sample (= n) less the number of independent (linear) constraints or restrictions put on them.
 In other words, it is the number of independent observations out of a total of n observations.
 For example, before the RSS can be computed, ᷇ β 1 and ᷇ β 2 must first be obtained.
These two estimates therefore put two restrictions on the RSS (residual sum of squares).
Therefore, there are n− 2, not n, independent observations to compute the RSS.
 Following this logic, in the three-variable regression RSS will have n− 3 df, and for the k-
variable model it will have n− k df.
 The general rule is this: df = (n− number of parameters estimated).

Error term 𝞵 and method of least squares

The stochastic error term is a catchall that includes all those variables that cannot be readily
quantified. It may represent variables that cannot be included in the model for lack of data

Whatever the source of the random term 𝞵, it is assumed that the average effect of the error
availability, or errors of measurement in the data, or intrinsic randomness in human behaviour.

term on the regression is marginal at best.

Refer to the example discussed above. As per the regression line Y = -7.4 + 1.64X, the
calculated values of Y will be as follows:

Error Square of
Rainfall (X) Yield (Y) Y1 (calculated) (Y-Y1) error
12 14 12.28 1.72 2.96
9 8 7.36 0.64 0.41
8 6 5.72 0.28 0.08
10 9 9 0 0
11 11 10.64 0.36 0.13
13 12 13.92 -1.92 3.69
7 3 4.08 -1.08 1.17
70 63 8.43

At these values of a and b, the sum of squares of the errors (or residuals) is minimum.
Graphically, it can be depicted as:

11
OLS Method

Regression line fitted as per The line minimizes the sum


actual data of squared errors

12

You might also like