You are on page 1of 6

FIN-403 Final Exam Sample Questions

Final Exam Date: January 16 2018, closed book

Question 1

Show that Hansen’s J-statistic converges to the χ2L−K distribution . In standard notation:
     
x01 z10 y
 1
 ..   ..   .. 
   
0 n×K n×L
y t = x t β + εt , X =  .  ∈ R ,Z =  .  ∈ R , y =  .  ∈ Rn ,
     
xn 0 zn 0 yn

n
1X Z 0ε
gn (b) = zt (yt − x0t b), (⇒ gn (β) = ).
n n
t=1

By model assumptions,
Z 0ε d p
√ → N (0, V ), Ŵ → V.
n
The efficient IV estimator is

X 0 Z Z 0 X −1 X 0 Z Z 0 y p
β̂IV = ( Ŵ ) Ŵ , where Ŵ → V −1 .
n n n n

Prove that
p
ng(β̂IV )Ŵ g(β̂IV ) → χ2(L−K) .

Why must it be the efficient IV estimator for this to hold?

Question 2

Consider data generating process (wt , t = 1, 2, · · · ) which is a parametric model with parameter
β ∈ RK and which satisfies strict stationarity and mixing properties. Suppose E[m(wt , β)] =
0 ∈ RL for some m(·, ·), L ≥ K. So that

β = arg min E[m(wt , b)]0 W E[m(wt , b)]0


b

for any positive definite W ∈ RL×L .

1
1. Define the nonlinear GMM estimator β̂ given by the moment condition E[m(wt , β)] = 0
and the weighting matrix W and derive the following asymptotic variance matrix of the
efficient non-linear GMM estimator:

d d
( E[ m(wt , b)|β ]0 Σ−1 E[ 0 m(wt , b)|β ] )−1 .
db0 db

Assume the moment conditions mt (wt , β) follows a martingale difference sequence, specify
the optimal weighting matrix Σ−1 that achieves efficiency.

2. Show that the asymptotic variance matrix of the efficient IV (linear GMM) estimator is
a special case of above expression.

3. Show that the (conditional-)hetereoskedasticity robust asymptotic variance matrix for


the OLS estimator is in turn a special case of the IV asymptotic variance matrix.1

Question 3

1. Provide a example of a regression encountered in class that suffers from endogeneity and
propose an instrument. Explain why endogeneity is present and justify that the proposed
instrument is exogenous.

2. Consider the linear model yi = βxi + εt where xi ∈ R. Suppose xi is being instrumented


by a dummy zi = 1{i∈S} where S is a subsample. Compute the IV estimate and compare
it with running OLS on the subsample S.

3. Consider the CAPM yt = α+βxt +εt . If, instead of running OLS, xt is being instrumented
by its lagged value xt−k for some k > 0. State the corresponding moment condition and
explain its economic meaning.

4. Suppose the econometrician would like to consider more lagged values xt−k−1 , · · · , xt−k−p
for the instrumented CAPM, how would she test for the joint exogeneity of her instru-
ments?
1
The sample counterpart of hetereoskedasticity robust asymptotic variance matrix for the OLS estimator is
the sandwich estimator.

2
Question 4

1. Provide an example of a financial time series that might be hetereoskedastic.

2. Define the ARCH(1) model (yt ), with intercept term. Your model should have 3 parameters—
e.g. (µ, α0 , α1 )0 .

3. Show that it is conditionally hetereoskedastic by computing the conditional variance


Var(yt |yt−1 ) explicitly in terms of model parameters.

4. Provide, and justify, moment conditions that would allow the model to be estimated by
GMM.

5. State appropriate Gaussianity assumptions that would allow the model to be estimated
by MLE and give the expression of log-likelihood function.

Question 5

1. Consider the family {f (w; θ)} of densities indexed by parameter θ from which data w
is drawn. Define the score and prove the Cramer-Rao information inequality:2 For any
R
unbiased estimator g(w), i.e. Eθ [g(w)] = g(w)f (w; θ)dw = θ for all θ,

∂2
Varθ (g) ≥ (−Eθ [ log f (w; θ)])−1 .
∂θ2

2. Show that the OLS estimator β̂ achieves the Cramer-Rao lower bound, under appropriate
Gaussianity assumptions. State precisely the distributional assumptions you make. In
terms of exogeneity for the error term, is it exogenous/strictly exogenous/or...?

3. Consider an M-estimator given by

θ̃ = arg max g(w; θ),


θ

where g(w; θ) is a sample mean of the form


n
1X
g(w; θ) = m(wt ; θ), (wt ) i.i.d.
n
t=1

Derive the asymptotic variance of θ̃ − θ and show MLE is asymptotically efficient.


2
You may freely differentiate under the integral sign as needed.

3
Question 6
i.i.d.
Consider the Gaussian AR(1) model yt = ρyt−1 + εt , εt ∼ N (0, σ 2 ), |ρ| < 1.

1. Give the expression of the conditional log-likelihood function for sample y1 , · · · , yn , con-
ditional on y1 and the corresponding score.

2. Show the FOC of the (conditional) MLE for this model are also moment conditions that
can be used in a GMM framework.

3. Explain why this is a case of quasi-MLE.

4. Provide another example of quasi-MLE. Look for the simplest example possible.

Question 7

Consider the static 1-dimensional Kalman filter: the latent state and the observed value are
given by random variables ξ and y = hx + aξ + w respectively, where E[w] = 0, Cov(ξ, w) = 0
and x is deterministic.

1. Show that the best linear predictor of ξ from observing y is3

aVar(ξ)
ξˆ = E[ξ] + (y − hx − aE[ξ])
a2 Var(ξ)+ Var(w)
aσξ
≡ µξ + (y − hx − aµξ ).
a2 σξ + σw

2. Assume ξ and w are jointly normal. Give the likelihood function f (ξ, y; θ) = f (y|ξ; θ)f (ξ; θ)
of this model, where θ = (a, h, µξ , σξ , σw )0 .

3. Let ξnext = F ξ + v, where E[v] = 0, Cov(v, w) = Cov(v, ξ) = 0, and F a constant. As-


sume all random variables are jointly normal. Compute the conditional mean E[ξnext |y]
and conditional variance Var(ξnext |y) of ξnext conditional on y.
3
By definition, the best linear predictor of ξ from observing y is

ξˆ = arg min E[(z − ξ)2 ].


z=a+by

4
4. Define the CAPM model where beta follows a random walk. State clearly the latent
state and observation variables. Provide economic or econometric reasons why one might
entertain such a model.

Question 8

1. Compare the GMM and MLE frameworks. Provide two advantages/disadvantages of


each.

2. If MLE is asymptotically efficient, why not always use MLE?

3. Show that MLE is always a special case of exactly identified GMM. On the other hand,
provide an example where GMM cannot be replaced by MLE. You may want to search
from the preceding questions.

Question 9

Consider the AR(1) model with no intercept term:

i.i.d.
yt = ρyt−1 + εt , εt ∼ (0, σ 2 ).

Consider forecasting k-periods ahead.

1. Show that the conditional mean E[yt+k |yt , yt−1 , · · · y1 ] = ρk yt .

2. Assume |ρ| < 1 and (yt ) is strictly stationary, show that E[yt ] = 0 for all t, and

kE[yt+k |yt , yt−1 , · · · ] − E[yt+k ]k2 = ρk kyt k2 → 0.

3. Based on part (2), explain why the process (yt ) exhibits mean-reversion by reasoning
along the same line as how the mixing inequality means mixing processes exhibit mean-
reversion.

4. Assume ρ = 1, explain why (yt ) cannot be weakly stationary (therefore a fortiori not
strictly stationary either). Explain why (yt ) is a martingale.

5. Provide an example of a financial time series that may be a unit root process (i.e. ρ = 1).
Provide underlying economic reason(s) why your example may be a unit root process.

5
Question 10

Consider the AR(1) model with no intercept term:

i.i.d.
yt = ρyt−1 + εt , εt ∼ (0, σ 2 ).

Let ρ̂T denote the OLS estimate from a sample of size T .

1. Assume |ρ| < 1 and (yt ) is strictly stationary, show that


√ d
T (ρ̂T − ρ) → N (0, 1 − ρ2 ).

Argue that ρ̂T is therefore consistent.

2. Assume ρ = 1 , show that


1
d (W 2 − 1)
T (ρ̂T − 1) → 2R 1 1 ,
2
0 Wr dr
where Wt is a standard Brownian motion on [0, 1]. Argue that ρ̂T is therefore consistent.
Does |ρ| < 1 or ρ = 1 give rise to faster rate of convergence of ρ̂T ?

3. Under H0 : |ρ| < 1, show the t-statistic

ρ̂ − ρ d
qT → N (0, 1).
1
σ̂ · PT 2
t=1 yt−1

4. Under H0 : ρ = 1, show that the Dickey-Fuller t-statistic


1 2
ρ̂ − 1 d 2 (W1 − 1)
q →q .
σ̂ · PT 1
R1
2 W 2 dr
t=1 yt−1 0 r

5. Explain how one can test whether the series you proposed in Question 9 part (6) is indeed
unit root.

You might also like