You are on page 1of 15

Ch.

1 Introduction to Estimation

1/15

An Example Estimation Problem: DSB Rx


S( f )

M( f )

f o

fo f

s(t; f o ,o ) = m(t ) cos(2f o t + o )


BPF
&
Amp
Electronics Adds
Noise w(t)
(usually white)

x ( t ) = s ( t ) + w( t )

Est. Algo.

Audio
Amp

cos(2fo t + o )

fo & o

Oscillator
w/ fo & o

Goal: Given x (t ) = s (t; f o ,o ) + w(t )


Find Estimates
(that are optimal in some sense)

M ( f )

f
Describe with
Probability Model:
PDF & Correlation
2/15

Discrete-Time Estimation Problem


These days, almost always work with samples of the observed
signal (signal plus noise): x[n ] = s[n; f , ] + w[n ]
o

Our Thought Model: Each time you observe x[n] it


contains same s[n] but different realization of noise w[n],
so the estimate is different each time.
fo & o are RVs
Our Job: Given finite data set x[0], x[1], x[N-1]
Find estimator functions that map data into estimates:
fo = g1 ( x[0], x[1], , x[ N 1]) = g1 ( x )

o = g 2 ( x[0], x[1], , x[ N 1]) = g 2 ( x )

These are RVs


Need to describe w/
probability model

3/15

PDF of Estimate
Because estimates are RVs we describe them with a PDF
Will depend on:
1. structure of s[n]
2. probability model of w[n]
3. form of est. function g(x)

p ( fo )

Mean measures centroid


Std. Dev. & Variance
measure spread

fo

fo

Desire:

E{ fo } = f o

2
2f = E fo E{ fo } = small
o
4/15

1.2 Mathematical Estimation Problem


General Mathematical Statement of Estimation Problem:
For Measured Data x = [ x[0] x[1] x[N-1] ]
Unknown Parameter = [1 2 p ]
is Not Random
x is an N-dimensional random data vector
Q: What captures all the statistical information needed for an
estimation problem ?
A: Need the N-dimensional PDF of the data, parameterized by
In practice, not given PDF!!!
Choose a suitable model
Captures Essence of Reality
Leads to Tractable Answer

p ( x; )
Well use p(x;) to find = g ( x )

5/15

Ex. Estimating a DC Level in Zero Mean AWGN


Consider a single data point is observed x[0] = + w[0]
~ N(,

2)

Gaussian
zero mean
variance 2

So the needed parameterized PDF is:


p(x[0]; ) which is Gaussian with mean of
So in this case the parameterization changes the data PDF mean:
p(x[0];1)

x[0]

p(x[0];2)

x[0]

p(x[0];3)

3 x[0]

6/15

Ex. Modeling Data with Linear Trend


See Fig. 1.6 in Text
Looking at the figure we see what looks like a linear trend
perturbed by some noise
So the engineer proposes signal and noise models:

x[n ] = [%
A
"+
$Bn
"
#] + w[n ]
s[ n; A, B ]

Signal Model: Linear Trend

Noise Model: AWGN


w/ zero mean

AWGN = Additive White Gaussian Noise


White = x[n] and x[m] are uncorrelated for n m

E ( w w )( w w )T = 2 I
7/15

Typical Assumptions for Noise Model


W and G is always easiest to analyze
Usually assumed unless you have reason to believe otherwise
Whiteness is usually first assumption removed
Gaussian is less often removed due to the validity of Central Limit Thm

Zero Mean is a nearly universal assumption


Most practical cases have zero mean
But if not w[n ] = w zm [n ] +
Non-Zero Mean of

Zero Mean

Now group into signal model

Variance of noise doesnt always have to be known to make an


estimate
BUT, must know to assess expected goodness of the estimate
Usually perform goodness analysis as a function of noise variance (or
SNR = Signal-to-Noise Ratio)
Noise variance sets the SNR level of the problem
8/15

Classical vs. Bayesian Estimation Approaches


If we view (parameter to estimate) as Non-Random
Classical Estimation
Provides no way to include a priori information about
If we view (parameter to estimate) as Random
Bayesian Estimation
Allows use of some a priori PDF on
The first part of the course: Classical Methods
Minimum Variance, Maximum Likelihood, Least Squares
Last part of the course: Bayesian Methods
MMSE, MAP, Wiener filter, Kalman Filter
9/15

1.3 Assessing Estimator Performance


Can only do this when the value of is known:
Theoretical Analysis, Simulations, Field Tests, etc.
Recall that the estimate = g ( x ) is a random variable
Thus it has a PDF of its own and that PDF completely displays
the quality of the estimate.
p()
Illustrate with 1-D
parameter case

Often just capture quality through mean and variance of = g ( x )


Desire:

m = E{} =

= E E{ } = small

If this is true:
say estimate is
unbiased
10/15

Equivalent View of Assessing Performance


Define estimation error: e =
RV

RV

( = + e )
Not RV

Completely describe estimator quality with error PDF: p(e)


p(e)

Desire:

me = E{e} = 0

e2 = E (e E{e})2 = small

If this is true:
say estimate is
unbiased

11/15

Example: DC Level in AWGN


Model: x[n ] = A + w[n ], n = 0, 1, , N 1
Gaussian, zero mean, variance 2
White (uncorrelated sample-to-sample)

PDF of an individual data sample:


p ( x[i ]) =

1
2 2

( x[i ] A) 2
exp

2
2

Uncorrelated Gaussian RVs are Independent


so joint PDF is the product of the individual PDFs:
( x[n ] A) 2
1
1
exp
p(x) =
=

2
2 N /2
2
2

n =0 2

( 2 )
N 1

exp

N 1

( x[n] A)

n =0

2 2

( property: prod of exps gives sum inside exp )

12/15

Prob. Theory

Each data sample has the same mean (A), which is the thing we
are trying to estimate so, we can imagine trying to estimate
A by finding the sample mean of the data:
1 N 1
A =

Statistics

x[ n ]

n =0

Lets analyze the quality of this estimator


Is it unbiased?
1
E{ A } = E
N

1
x
[
n
]
E{x[i ]}
= %

#
n =0
N n $
=A
E{ A } = A
Yes! Unbiased!
N 1

Due to Indep.
(white & Gauss.
Indep.)

Can we get a small variance?


1

var( A) = var
N

N 1

1
x
n
[
]
=

2
N
n =0

var( A ) =

2
N

N 1

N 1

var( x[n]) = N 2

n =0

n =0

N 2
N2

Can make var small by increasing N!!!

13/15

Theoretical Analysis vs. Simulations


Ideally wed like to be always be able to theoretically
analyze the problem to find the bias and variance of the
estimator
Theoretical results show how performance depends on the problem
specifications

But sometimes we make use of simulations


to verify that our theoretical analysis is correct
sometimes cant find theoretical results

14/15

Course Goal = Find Optimal Estimators


There are several different definitions or criteria for optimality!
Most Logical: Minimum MSE (Mean-Square-Error)

See Sect. 2.4


To see this result:
2

mse( ) = E

mse( ) = E

[(

) (

)]

= E E{} + E{}

= var{} + b 2 ( )

Bias

= E E{ } + b( ) E E{} + b 2 ( )
%"$
" "#
"

b( ) = E{}

=0

= var{} + b 2 ( )

Although MSE makes sense, estimates usually rely on


15/15

You might also like