Professional Documents
Culture Documents
Exercise 1
(a) First population moment can be calculated as
Z 1 Z 1 1
θ−1 θ θ
θxθ dx = θ+1
E(X) = xθx dx = x = .
0 0 θ+1 0 θ + 1
Equate it to the first sample moment and solve the equation to obtain the MME θ̃:
θ̃ X̄
= X̄ ⇒ θ̃ = .
θ̃ + 1 1 − X̄
Equate it to the first sample moment and solve the equation to obtain the MME θ̃:
θ̃ + 1 1
= X̄ ⇒ θ̃ = .
θ̃ X̄ − 1
x
(c) The pdf corresponds to a GAM(1/θ, 2) distribution since f (x; θ) = θ2 xe−θx = (1/θ)12 Γ(2) xe− 1/θ .
We can find the population moment E(X) = θ2 from Table B.2. Equate it to the first sample
moment and solve the equation to obtain the MME θ̃:
2 2
= X̄ ⇒ θ̃ = .
θ̃ X̄
Exercise 3
Qn Qn θ−1
(a) The likelihood function is L(θ) = i=1 f (xi ; θ) = θn ( i=1 xi ) and the associated log-
likelihood is
n
X
ln L(θ) = n ln θ + (θ − 1) ln(xi ).
i=1
The first and second derivative of the log-likelihood are:
n
d n X d2 n
ln L(θ) = + ln(xi ), 2
ln L(θ) = − 2 < 0.
dθ θ i=1 dθ θ
1
The second derivative is negative for all values of θ. We can thus solve the first order
condition to find the estimator θ̂ = − Pn nln(Xi ) .
i=1
Qn Qn −θ−2
(b) The likelihood function is L(θ) = i=1 f (xi ; θ) = (θ+1)n ( i=1 xi ) and the associated
log-likelihood is
n
X
ln L(θ) = n ln(θ + 1) − (θ + 2) ln(xi ).
i=1
The first and second derivative of the log-likelihood with respect to θ are
n
d n X d2 n
ln L(θ) = − ln(xi ), 2
ln L(θ) = − < 0.
dθ θ + 1 i=1 dθ (θ + 1)2
The second derivative is negative for all values of θ. We can thus solve the first order
condition to find the estimator θ̂ = Pn nln(Xi ) − 1.
i=1
Qn Qn Pn
(c) We have the likelihood function L(θ) = i=1 f (xi ; θ) = θ2n ( i=1 xi ) e−θ i=1 xi
and log-
likelihood function
n
X
ln L(θ) = 2n ln(θ) + ln(xi ) − θnx̄,
i=1
Pn
where we used x̄ = n1 i=1 xi . The first derivative is dθd
ln L(θ) = 2n
θ − nx̄. Solving the
2n
Pn
first order condition, i.e. θ̂ − i=1 xi = 0, gives the candidate solution θ̂ = 2/x̄. For the
second derivative we have
d2 2n
ln L(θ) = − 2 < 0,
dθ2 θ
for all θ. We conclude that the maximum likelihood estimator is θ̂ = 2/X̄.
Exercise 5
For the given pdf, the likelihood function equals
n n
!−3
Y Y
n 2n
L(θ) = f (xi ; θ) = 2 θ xi , 0 < θ ≤ x1:n = min xi .
1≤i≤n
i=1 i=1
Exercise 7
All the quantities in exercises (a)-(c) are transformations of p. We will thus first derive the MLE
for the parameter p and subsequently use the invariance property. First the derivation of the
likelihood function
n
Y Pn
xi −n
L(p) = f (xi ; p) = pn (1 − p) i=1 = pn (1 − p)n(x̄−1) ,
i=1
2
The first derivative of the log-likelihood function is
d n n(x̄ − 1)
ln L(p) = − = 0.
dp p 1−p
We can find a candidate solution for the MLE by setting this first derivative equal to zero, that
is
n n(x̄ − 1) 1
− =0 ⇒ p̂ = .
p̂ 1 − p̂ x̄
It remains to verify whether
our candidate solution indeed gives a maximum. For this we should
d2
show that dp2 ln L(p) < 0. By differentiation we find
p=p̂
d2 x̄3
n n(x̄ − 1)
2
ln L(p) = − 2
− 2
= −n <0
dp
p=p̂ p (1 − p) p=p̂
x̄ − 1
since x ∈ {1, 2, 3, . . .} hence x̄ > 1 (we rule out the case when we observe a sample of only ones
1
because this gives no information about the parameter). The ML estimator is thus p̂ = X̄ . The
subquestions are now quick to answer using the Invariance Property of MLEs (Theorem 9.2.2 on
page 298 of B&E).
(a) τ (p) = E(X) = p1 , hence the MLE is τ (p̂) = 1
p̂ = X̄
1−p 1−p̂
(b) τ (p) = Var(X) = p2 , hence the MLE is τ (p̂) = p̂2 = X̄(X̄ − 1)
1 k
(c) τ (p) = P(X > k) = (1 − p)k , hence the MLE is τ (p̂) = (1 − p̂)k = 1 − X̄
for arbitrary
k = 1, 2, . . .
Exercise 15
(a) If X ∼ BIN(n, p), then E(X) = np and Var(X) = np(1 − p) (see Table B.2). We have
X2 E(X 2 )
X X c c
E cp̂(1 − p̂) = E c 1− = E X− = E(X) −
n n n n n n
2
np(1 − p) + (np)2
c Var(X) + (E(X)) c
= E(X) − = np −
n n n n
c c n−1
= np − p(1 − p) − np2 = np(1 − p) − p(1 − p) = c p(1 − p).
n n n
n
E cp̂(1 − p̂) = p(1 − p) will hold when c = n−1 .
(b) Note that Var(X) = np(1 − p). In view of the previous exercise we obtain the unbiased
n2
estimator n−1 p̂(1 − p̂).
(c) We now have a random sample X1 , . . . , XN ∼ BIN(n, p). The fact that E(X) = np suggest
PN
the estimator p̂∗ = nN1
i=1 Xi . The following calculation shows that this is indeed an
unbiased estimator:
N
! N
∗ 1 X 1 X 1
E(p̂ ) = E Xi = E(Xi ) = N (np) = p.
nN i=1 nN i=1 nN
3
Similarly, an unbiased estimator for Var(X) = np(1 − p) is easily constructed using the
\ = 1 PN n2 Xi (1 − Xi ), we have
answer to part (a). Defining the estimator as Var(X) N i=1 n−1 n n
N ! N
1 X n2 Xi
2
\ =E
Xi 1 X n Xi Xi
E Var(X) 1− = E 1− = Var(X).
N i=1 n − 1 n n N i=1 n−1 n n
Exercise 17
(a) Since X ∼PUNIF(θ − 1, θ + 1), we have E(X) = θ−1+θ+1 2 = θ (see Table B.2) and also
n
E(X̄) = n1 i=1 E(Xi ) = θ. X̄ is thus an unbiased estimator for θ.
(b) The pdf of the uniform distribution UNIF(θ − 1, θ + 1) is
1
f (x; θ) = θ−1<x<θ+1
2
and the CDF is
0 x≤θ−1
x−θ+1
F (x; θ) = 2 θ−1<x<θ+1
1 x≥θ+1
(see page 109). Using Theorem 6.5.2 (page 217 of B&E), we see that the pdfs of the order
statistics X1:n and Xn:n are
n
g1 (x) = n(1 − F (x))n−1 f (x) = (θ + 1 − x)n−1 θ−1<x<θ+1
2n
n
gn (x) = n(F (x))n−1 f (x) = (x − θ + 1)n−1 θ − 1 < x < θ + 1,
2n
respectively. First, we calculate E(X1:n ):
Z θ+1 Z 2
n n
E(X1:n ) = x n (θ + 1 − x)n−1 dx = (θ + 1 − y) n y n−1 dy
θ−1 2 0 2
Z 2
n n 2n 2
= (θ + 1) − n
y dy = (θ + 1) − = (θ − 1) + ,
0 2 n+1 n+1
by changing the integration variable to y = θ + 1 − x. In other words, X1:n is on average
2
n+1 higher than the lower bound θ − 1. Second, for E(Xn:n ), we have
Z θ+1 Z 2
n n−1 n
E(Xn:n ) = x n (x − θ + 1) dx = (z + θ − 1) n z n−1 dz
θ−1 2 0 2
Z 2
n n 2n 2
= (θ − 1) + n
z dz = (θ − 1) + = (θ + 1) − ,
0 2 n+1 n+1
after changing the integration variable to z = x − θ + 1. We see that Xn:n is lower than
2
the upper bound θ + 1 by n+1 (the same quantity as before). Finally, by linearity of the
expectation, we have
2 2
(θ − 1) + n+1 + (θ + 1) − n+1
X1:n + Xn:n E(X1:n ) + E(Xn:n )
E = = = θ,
2 2 2
and we see that the “midrange” is indeed an unbiased estimator for θ.
4
Exercise 21
(a) If X ∼ BIN(1, p), then E(X) = p and Var(X) = p(1−p) (see Table B.2). For the numerator
of the CRLB, we have τ (p) = p and thus τ 0 (p) = 1. The following calculations can be used
to evaluated the expectation in the denominator:
f (x; p) = px (1 − p)1−x
ln f (x; p) = x ln p + (1 − x) ln(1 − p)
∂ x 1−x x−p
ln f (x; p) = − =
∂p p 1−p p(1 − p)
2 2
E(X − p)2
∂ X −p Var(X) 1
E ln f (X; p) = E = 2 2
= 2 2
=
∂p p(1 − p) p (1 − p) p (1 − p) p(1 − p)
The CRLB is now obtained as
[τ 0 (p)]2 1 p(1 − p)
2 = n = .
n
∂ p(1−p)
nE ∂p ln f (X; p)
(b) Only the numerator of the CRLB will change. We now have τ (p) = p(1 − p) such that
τ 0 (p) = 1 − 2p. The CRLB is
[τ 0 (p)]2 (1 − 2p)2 p(1 − p)(1 − 2p)2
2 = n = .
n
∂ p(1−p)
nE ∂p ln f (X; p)
(c) Looking at your answer for part (a) you should recognize that the CRLB coincides with
Var(X)/n. As an educated guess we therefore try p̂ = X̄. First, from E(X) = p, we see
that !
n n
1X 1X
E(p̂) = E Xi = E(Xi ) = p
n i=1 n i=1
and conclude that X̄ is an unbiased estimator for p. The variance from this estimator, i.e.
n
! n
1X 1 X p(1 − p)
Var(p̂) = Var Xi = 2 Var(Xi ) =
n i=1 n i=1 n
5
The CRLB is now obtained as
[τ 0 (µ)]2 1 9
2 = n = .
n
∂ 9
n E ∂µ ln f (X; µ)
Exercise 23
(a) We first have to derive the MLE for θ. We proceed with the usual steps. The likelihood
n 2
P
i=1 xi
Qn −n/2
function is L(θ) = i=1 f (xi ; θ) = (2πθ) exp − 2θ and therefore
Pn
n n x2
ln L(θ) = − ln(2π) − ln(θ) − i=1 i .
2 2 2θ
We subsequently compute the first two derivatives as
Pn
d n x2
ln L(θ) = − + i=12 i
dθ 2θ P 2θ
n 2
d2 n i=1 xi
2
ln L(θ) = 2
− 3
.
dθ 2θ θ
If P
we equate the first derivate to zero and solve for the estimator, then we find θ̂ =
1 n 2
n i=1 xi . The second order condition is fulfilled because
d2
n nθ̂ n
2
ln L(θ) = − =− < 0.
dθ θ=θ̂ 2θ̂ 2 θ̂ 3 2θ̂2
Pn
The MLE for θ is thus θ̂ = n1 i=1 Xi2 . From
n
! n n
1X 2 1X 1X
E(θ̂) = E Xi = E(Xi2 ) = Var(Xi ) = θ,
n i=1 n i=1 n i=1
we see that this estimator is unbiased for θ.
6
2
(b) Starting from f (X; θ) = √1 exp − x2θ , we find
2πθ
1 1 X2
ln f (X; θ) = − ln(2π) − ln(θ) − ,
2 2 2θ
and by steps similar to those in part (a), the second derivatie w.r.t. θ becomes
∂2 1 X2
ln f (X; θ) = − .
∂θ2 2θ2 θ3
√ 2
Noteh that X/ θ i∼ N(0, 1), and
thus Xθ ∼ χ2 (1). Using the latter result, we have
∂2 1 X2
− E ∂θ 2 ln f (X; θ) = θ 2 E θ − 2θ12 = 2θ12 and a CRLB evaluating to
−1
∂2 2θ2
−n E ln f (X; θ) = .
∂θ2 n
The variance of θ̂ is
n
! n 2 !
θ2 θ2 2θ2
1X 2 θX X
Var(θ̂) = Var X = Var √i = Var(Yn ) = 2 (2n) = ,
n i=1 i n i=1 θ n2 n n
Pn Pn Xi 2
where we made use of the random variable Yn = θ1 i=1 Xi2 = i=1
√
θ
. Since
√
Xi / θ ∼ N(0, 1), we know that this Yn is the sum of squared (and independent) stan-
dard normal random variables. Therefore, Yn ∼ χ2 (n) and Var(Yn ) = 2n. The estimator
θ̂ is thus unbiased (see (a)) and its variance attains the CRLB. We conclude that θ̂ is an
UMVUE for θ.
Exercise 26
(a) The likelihood function is
n
Y 1
L(θ) = f (xi ; θ) = , 0 < x1:n = min xi , xn:n = max xi ≤ θ.
i=1
θn 1≤i≤n 1≤i≤n
θ̃
X̄ = ⇒ θ̃ = 2X̄.
2
7
The pdf of the order statistic Xn:n , cf. Theorem 6.5.2 (page 217 of B&E), is thus equal to
n n−1
gn (x) = n(F (x))n−1 f (x) = x , 0 < x ≤ θ,
θn
(and zero elsewhere). We can now find E(Xn:n ) by integration,
θ θ
n xn+1
Z
n n n
E(θ̂) = E(Xn:n ) = x dx = = θ 6= θ,
0 θn θ n n + 1 0 n+1
We continue with the method of moments estimator θ̃. This estimator was unbiased such
θ2
that MSE(θ̃) = Var(θ̃). Since Var(X) = 12 , we can easily compute this variance by
exploiting the standard properties of variances, namely
n
! n
2X 4 X θ2
MSE(θ̃) = Var(θ̃) = Var(2X̄) = Var Xi = 2 Var(Xi ) = .
n i=1 n i=1 3n
The MSE of both estimator scales linearly with θ2 (which is expected because this is the
scale parameter). More interesting is the behavior as a function of n. The MLE will have
a smaller (or equal) MSE, that is MSE(θ̂) ≤ MSE(θ̃), when
2θ2 θ2
≤ ⇒ n2 − 3n + 2 = (n − 1)(n − 2) ≥ 0.
(n + 1)(n + 2) 3n
The parabolic function f (x) = (x − 1)(x − 2) intersect the x-axis in the points x = 1 and
x = 2. We therefore conclude that the MLE for θ has an MSE that is never higher than
the MSE of the MME (for any sample size n = 1, 2, . . .).
Exercise 31
The MSEs for θ̂ and θ̃ were computed in Exercise 26. According to Definition 9.4.2 (page 312 in
B&E) we only have to take the limit n → ∞.
8
2θ 2
(a) limn→∞ MSE(θ̂n ) = limn→∞ (n+1)(n+2) = 0, hence θ̂n = Xn:n is MSE consistent for θ.
θ2
(b) limn→∞ MSE(θ̃n ) = limn→∞ 3n = 0, hence θ̃n = 2X̄n is MSE consistent for θ.
Exercise 32 Rx x
In Exercise 5 we have seen that θ̂n = X1:n . From θ 2θ2 t−3 dt = −θ2 t−2 θ = 1 − θ2 x−2 , we find
the following CDF for θ̂:
0 x≤θ
F (x; θ) =
1 − θ2 x−2 θ < x.
and zero otherwise. We can now compute the probability stated in Definition 9.4.1 (page 311 in
B&E) explicitly:
Z θ+ε
2nθ2n x−2n−1 dx
P |θ̂n − θ| < ε = P X1:n − θ < ε = P X1:n < θ + ε =
θ
2n
θ+ε θ
= −θ2n x−2n θ = 1 − θ2n (θ + ε)−2n = 1 − .
θ+ε
θ
Since 0 < θ+ε < 1 for any > 0, we obtain limn→∞ P |θ̂n − θ| < ε = 1 thereby showing that
the MLE θ̂n = X1:n is (simply) consistent for θ.
1 The
Qn
CDF for X1:n is P(X1:n ≤ x) = 1−P(X1:n > x) = 1−P(X1 > x, X2 > x, . . . , Xn > x) = 1− i=1 P(Xi >
x) = 1 − [1 − F (x)]n . By differentiation w.r.t. x we find g1 (x) = n(1 − F (x))n−1 f (x).
9
Exercise 33
(a) If X ∼ POI(µ), then E(X) = Var(X) = µ (see Table B.2). For the numerator of the
CRLB, τ (µ) = µ yields τ 0 (µ) = 1. The following calculations can be used to evaluated the
expectation in the denominator:
e−µ µx
f (x; µ) =
x!
ln f (x; µ) = −µ + x ln µ − ln(x!)
∂ x x−µ
ln f (x; µ) = −1 + =
∂µ µ µ
2 2
E(X − µ)2
∂ x−µ Var(X) 1
E ln f (X; p) = E = = = .
∂µ µ µ2 µ2 µ
[τ 0 (µ)]2 1 µ
2 = n = .
n
∂ µ
nE ∂µ ln f (X; µ)
(b) The denominator of the CRLB remains unchanged. But now θ = τ (µ) = e−µ , hence
τ 0 (µ) = −e−µ . The new CRLB is thus
(c) The CRLB for µ has the form Var(X)/n. The sample mean is therefore a promising
candidate for an UMVUE. We compute mean and variance of our candidate estimator
µ̂ = X̄ and find
n
! n
1X 1X
E(µ̂) = E Xi = E(Xi ) = µ,
n i=1 n i=1
n
! n
1X 1 X µ
Var(µ̂) = Var Xi = 2 Var(Xi ) = .
n i=1 n i=1 n
From E(µ̂) we see that µ̂ is an unbiased estimator for µ and its variance attains the CRLB.
We conclude that µ̂ = X̄ is an UMVUE for µ.
(d) We use the invariance property for the transformation θ = τ (µ) = e−µ . The MLE for θ is
thus θ̂ = τ (µ̂) = e−X̄ .
Pn
(e) Let us define the new random variable Yn = i=1 Xi . It can by shown (using for instance
the properties of moment generating functions) that Yn ∼ POI(nµ). We have
1 1
1
1 nµ e− n −1 n 1−e− n
E(θ̂) = E e−X̄ = E e− n Yn = MYn −
=e =θ 6= θ,
n
10
(f) θ̂ is asymptotically unbiased for θ if limn→∞ E(θ̂) = θ. The latter expectation was already
calculated in the previous part so it remains to take the limit. Using the rules of limits, we
have 1 1
n 1−e− n limn→∞ n 1−e− n
lim E(θ̂) = lim θ =θ .
n→∞ n→∞
−1/n
The limit in the exponent can be written as limn→∞ 1−e1/n which at first sight gives the
indeterminate form 00 . As a possible solution one may realize that 1/n becomes small as
n → ∞ which motivates the use of a Taylor series for the exponential (that is exp(x) =
P ∞ xn
n=0 n! ). This results in
h
−n 1
i 1 1 1 1
n 1−e =n 1− 1− + 2 − + − ...
n 2n 3!n3 4!n4
1 1 1
=1− + − + · · · → 1 for n → ∞,
2n 3!n2 4!n3
1
n 1−e− n
and hence limn→∞ E(θ̂) = limn→∞ θ = θ. θ̂ is asymptotically unbiased for θ.
Pni=1 Xi Yn
(g) Using the previously defined Yn , we have θ̃ = n−1
n = n−1n . Then
( n−1
n ) −1
Yn !
log
n−1
Yn log( n−1 )
n−1 nµ e
E(θ̃) = E =E e n = MYn log =e
n n
n−1
= enµ( n −1 ) = e−µ = θ,
and find
2 µ
µ
Var(θ̃) = E θ̃2 − E(θ̃) = e−µ(2− n ) − [e−µ ]2 = e−2µ+ n − e−2µ = e−2µ e n − 1 .
1
µe−2µ
We should compare this expression to the CRLB which was found in part (b), that is n .
Another application of the Taylor series for the exponential results in
µ2 µ3
µ µ
Var(θ̃) = e−2µ e n − 1 = e−2µ 1+ + 2 + + . . . − 1
n 2n 3!n3
µe−2µ µ2
µ
= 1+ + + ... .
n 2n 3!n2
2
µ µ
Note that the higher order terms like 2n , 3!n 2 et cetera will all positive be positive. The
11
Exercise 34
Qn
(a) The likelihood and log-likelihood are L(p) = i=1 p(1 − p)xi = pn (1 − p)nx̄ and
ln L(p) = n ln(p) + nx̄ ln(1 − p),
respectively. The first and second derivative of this log-likelihood with respect to the
parameter p are
d n nx̄ n − np(1 + x̄)
ln L(p) = − =
dp p 1−p p(1 − p)
2
d n nx̄
ln L(p) = − 2 − .
dp2 p (1 − p)2
1
Equating the first derivative to zero yields the candidate solution p̂ = 1+x̄ . The following
calculation shows that this indeed gives a maximum
d2 n(1 + x̄)2
= −n(1 + x̄)2 −
2
ln L(p) ,
dp
p=p̂ x̄
x̄
where we used 1 − p̂ = 1+x̄ and rule out the case where x̄ = 0 (this can happen with finite
probability yet will not give us information regarding p). We conclude that the MLE for p
is p̂ = 1+1X̄ .
1−p
(b) Apply the invariance property to θ = τ (p) = p . The MLE is obtained as
1
1− 1+X̄
θ̂ = τ (p̂) = 1 = X̄.
1+X̄
(c) If X has pdf f (x; p) = p(1 − p)x for x = 0, 1, . . ., then Y = 1 + X has the GEO(p)
distribution. By linearity of the expectation we get E(X) = E(Y ) − 1 = p1 − 1 (see Table
B.2). This expectation will be needed later on. For the numerator of the CRLB, τ (p) = 1−p p
yields τ 0 (p) = − p12 . The following calculations can be used to evaluated the expectation in
the denominator of the CRLB:
∂2 1 x
ln f (x; p) = − 2 −
∂p2 p (1 − p)2
1
p −1
2
∂ 1 E(X) 1 1
E 2
ln f (X; p) = − 2
− 2
= − 2
− 2
=− 2
∂p p (1 − p) p (1 − p) p (1 − p)
The CRLB equals
1
[τ 0 (p)]2 p4 1−p
= n = .
−n E ∂2
ln f (X; p) p2 (1−p) np2
∂p2
Pn
(d) We compute the mean and variance of θ̂. The expectation is E(θ̂) = E(X̄) = E n1 i=1 Xi =
1
Pn 1−p
n i=1 E(Xi ) = p = θ. Because X is shifted version of Y , we have Var(X) = Var(Y ) =
1−p
p2 . The variance of θ̂ is therefore
n
! n
1X 1 X 1−p
Var(θ̂) = Var(X̄) = Var Xi = Var(Xi ) = .
n i=1 n2 i=1 np2
The estimator θ̂ is unbiased for θ and attains the CRLB. We conclude that θ̂ = X̄ is an
UMVUE for θ.
12
1−p
(e) limn→∞ MSE(θ̂n ) = limn→∞ Var(θ̂n ) = limn→∞ np2 = 0 which shows that the MLE θ̂n =
X̄n of θ is MSE consistent for θ.
(f) Under regularity conditions we know that the asymptotic distribution of MLEs is normal
with mean θ and the variance being equal to the CRLB (see page 316 of B&E). Thus, for
large n, approximately
1−p
θ̂ ∼ N θ, .
np2
√ d
It is mathematically neater to write n θ̂ − θ − → N 0, 1−p
p 2 .
13