You are on page 1of 15

Two-Variable Regression

Model: The Problem of


Estimation
Gujarati 4e, Chapter 3
The Method of Ordinary Least
Squares (OLS)
 Carl Friedrich Gauss (German)
 OLS  very attractive statistical properties
that made OLS one of the most powerful and
popular methods
 PRF: Yi  1   2 X i  ui
 SRF: Yi  ˆ1  ˆ2 X i  uˆi
 Yˆi  uˆi

01/06/22 Prepared by Sri Yani Kusumastuti 2


Least squares criterion

Y
Yi SRF: Yˆi  ˆ1  ˆ2 X i

û3
û1
uˆi  Yi  Yˆi  Yi  ˆ1  ˆ2 X i
û4
û2 Minimizing the sum of the
residuals  zero

OLS  Minimizing the sum


of squares the residuals
X
X1 X2 X3 X4

01/06/22 Prepared by Sri Yani Kusumastuti 3


Least squares criterion
 
2
Minimizing  uˆ 2
i  Yi  ˆ1  ˆ2 X i

  ˆ
u 2

 i  2  1 Y  ˆ  ˆ X  0
ˆ

 i 1 2 i  
1

  ˆ
u 2

 i  2  1 Y  ˆ  ˆ X X  0
ˆ

 i 1 2 i i  
2

The normal equation:


 Y  nˆ  ˆ
i X 1 2  i

Y X i i  ˆ1  X i  ˆ2  X i2

01/06/22 Prepared by Sri Yani Kusumastuti 4


Least squares criterion
 Solving the normal equations simultaneously,
we obtain

ˆ
2 
n X iYi   X i  Yi

 X  X   Y Y    x y
i i i i i i

n X i    X i 
2 2
 X  X  i x i
2 2
i

ˆ1   i  Yi   X i  X iYi
X 2

 Y  ˆ2 X
n X    X i 
2 2
i

ˆ1 and ˆ2  the least-squares estimators

01/06/22 Prepared by Sri Yani Kusumastuti 5


The regression line properties

1. It passes through the sample mean of Yi


and Xi
2. The mean value of the estimated Yi is equal
to the mean value of actual Yi
3. The mean value of the residual is zero
4. The residuals are uncorrelated with the
predicted Yi
5. The residuals are uncorrelated with Xi

01/06/22 Prepared by Sri Yani Kusumastuti 6


The classical linear regression
model: the assumption of LS
1. Linear regression model
2. X values are fixed in repeated sampling
3. Zero mean value of disturbance ui
4. Homoscedasticity or equal variance of ui
5. No autocorrelation between the disturbances
6. Zero covariance between ui and Xi
7. The number of observation n must be greater than
the number of parameters to be estimated
8. Variability in X values
9. The regression model is correctly specified
10. There is no perfect multicollinearity

01/06/22 Prepared by Sri Yani Kusumastuti 7


What happen if the assumptions of
CLRM are violated?
Assumption number Type of violation
1 Nonlinearity in parameters
2 Stochastic regressor(s)
3 Nonzero mean of ui
4 Heteroscedasticity
5 Autocorrelated disturbance
6 Nonzero covariance between disturbance and
regressor
7 Sample observation less than the number of
regressor
8 Insufficient variability in regressors
9 Specification bias
10 multicollinearity

01/06/22 Prepared by Sri Yani Kusumastuti 8


Standard errors of Least-Squares
(LS) estimates
 
   
2
var ˆ2  dan se ˆ2 =
ix 2
i
x 2

  n x 
var ˆ1 
 i
X 2
2
dan  
se ˆ1 =
 i
X 2


2
i n x 2
i

2  i
ˆ
u 2

dan  i
ˆ
u 2

nk nk
  xy
2
  xi yi 
2

 uˆ   y  2  xi   yi   x 2 
ˆ
x  y  x
2 2 22 2 2 2 i i

  i  
i i i i 2
i

01/06/22 Prepared by Sri Yani Kusumastuti 9


Properties of LS estimators: the
Gauss-Markov Theorem
 An estimator is said to be a best linear
unbiased estimator (BLUE) if
1. Linear  a linear function of a random variable
2. Unbiased  its average or expected value, E ˆ2 is
equal to the true value, 2
 
3. Efficient or best  it has minimum variance

Gauss-Markov Theorem: given the assumption of CLRM,


the least-squares estimators, in the class of unbiased linear
estimators, have minimum variance, that is, they are BLUE

01/06/22 Prepared by Sri Yani Kusumastuti 10


The coefficient of determination r2: a
measure of “Goodness of Fit”
uˆi  due to regression
Y
Yi
SRF : ˆ1  ˆ2 X i

 Yi  Y   total Yˆi
 Yˆ  Y   due to regression
i

Y
r2  tells how well
the SRF fits the data

X
Xi

01/06/22 Prepared by Sri Yani Kusumastuti 11


The coefficient of determination r2: a
measure of “Goodness of Fit”
Yi  Yˆi  uˆi
in the deviation form
yi  yˆ i  uˆi
 i  i  i  2 yˆiuˆi
y 2
 ˆ
y 2
 ˆ
u 2

  yˆi2   uˆi2
 ˆ22  xˆi2   uˆi2
TSS  ESS  RSS
01/06/22 Prepared by Sri Yani Kusumastuti 12
The coefficient of determination r2: a
measure of “Goodness of Fit”
ESS RSS
1 
TSS TSS

 
2
ESS Yˆi  Y
r 
2

 Y Y 
2
TSS i

r2 measures the proportion or percentage of the total


variation in Y explained by the regression model
Two properties of r2:
1. It is a nonnegativity quantity
2. Its limits are 0  r2  1

01/06/22 Prepared by Sri Yani Kusumastuti 13


Coefficient of correlation r
 r is a measure of degree of association between
two variables

r   r2

r
x y i i

x y 
2
i
2
i

n X Y    X    Y 
i i i i
r
n X   X   n Y   Y  
      
2 2 2 2
i i i i 

01/06/22 Prepared by Sri Yani Kusumastuti 14


The properties of r
1. It can be positive or negative
2. It lies between -1 and +1, that is -1  r  +1
3. It is symmetrical in nature, rxy=ryx
4. It is independent of the origin and scale
5. If X and Y are statistically independent, the
correlation coefficient between them is zero, but if
r=0, it does not mean that two variables are
independent
6. It is a measure of linear association or linear
dependence only; it has no meaning for describing
nonlinear relations
7. It does not necessarily imply any cause-and-effect
relationship

01/06/22 Prepared by Sri Yani Kusumastuti 15

You might also like