Professional Documents
Culture Documents
15 Mvue PDF
15 Mvue PDF
1 / 28
Bias-Variance Trade-Off
Recall that
b = Bias2 ()
b + Var().
b
MSE()
In general, the minimum MSE estimator has non-zero bias and
non-zero variance.
We can reduce bias only at a potential increase in variance.
Conversely, modifying the estimator to reduce the variance may
lead to an increase in bias.
2 / 28
Example:
Let
xn = A + wn
wn N 0, 2
N
X
e=
xn
A
N n=1
N
1 X
xn ,
N n=1
then
e =
A
SN
3 / 28
Example: (cont.)
Lets find the value of that minimizes the MSE.
e = Var (SN ) = Var (SN ) =
Var A
h i
e = E A
e A = E [ SN ] A =
A=
Bias A
Thus the MSE is
e =
MSE A
4 / 28
#
"
N
N
X
X
1
2 1
= E
x n A + A2
xi xj 2E
N 2 i,j=1
N n=1
N
N
1 X
1 X
E[x
x
]
2
E[xn ]A + A2
i
j
N 2 i,j=1
N n=1
2
= 2 A2 +
2A2 + A2
N
2 2
2
=
+ ( 1) A2
|
{z
}
N
| {z }
2
e)
Bias
A
(
e
Var(A)
5 / 28
6 / 28
7 / 28
b3 is MVUE
no MVUE exists!
8 / 28
Example:
Suppose we observe a single scalar realization x of
X Unif (0, 1/) , > 0.
An unbiased estimator of does not exist. To see this, note that
p (x|) = I[0,1/] (x) .
If b is unbiased, then
h i
> 0, = E b =
=
=
But if this is true for all , then we have b (x) = 0, which is not an
unbiased estimator.
9 / 28
10 / 28
12 / 28
Rao-Blackwell Theorem
Rao-Blackwell Theorem
Let Y , Z be random variables and define the function
g(z) := E[Y |Z = z].
Then
E[g(Z)] = E[Y ]
and
Var(g(Z)) Var(Y )
with equality iff Y = g(Z) almost surely.
Note that this version of Rao-Blackwell is quite general and has
nothing to do with estimation of parameters. However, we can
apply it to parameter estimation as follows.
15 / 28
17 / 28
=x1
E[b
] =
Var [b
] =
The MSE is therefore:
e = (x1 + x2 )
2
averages the two observations together.
18 / 28
=
=
19 / 28
p(x1 , t)
p(t)
p(x1 , t) =
p(t) =
E(t) =
Var(t) =
20 / 28
p(x1 |t)
=
=
=
=
x1 |t
1
2 2
1
4 2
1
2
2
2
(x1 ) + (t x1 ) (t 2) /2
exp
2 2
1 2
1
x 2x1 + 2 + t2 2x1 t + x21 2t+
exp 22 1
+2x1 + 2 t2 /2 + 4t/2 42 /2
2
1
1
2
2
2x
2x
t
+
t
/2
exp
1
1
2 2
2
1
(x1 t/2)2
exp
2
2
=E[b
|t] =
Var( ) =
MSE( ) =
21 / 28
Lehmann-Scheffe Theorem
If t is complete, there is at most one unbiased estimator that is a
function of t.
22 / 28
Proof
Suppose
E[b1 ] = E[b2 ] =
b1 (X) := g1 (T (X))
b2 (X) := g2 (T (X)).
Define
(t) := g1 (t) g2 (t).
Then
E[(t)] =
By definition of completeness, we have
In other words
b1 = b2 with probability 1.
23 / 28
24 / 28
25 / 28
Proof (cont.)
26 / 28
xi Unif[0, ], i = 1, . . . , N.
What is an unbiased estimator of ?
N
2 X
b
1 =
xi
N
i=1
27 / 28
Example: (cont.)
From the Fisher-Neyman factorization theorem,
p(X|) =
N
Y
1
i=1
I[0,] (xi )
1
I
() I(,mini xi ] (0)
N [maxi xi ,)
{z
}
|
{z
} |
a(X)
b (t)
we see that
T = max xi
i