You are on page 1of 26

Chapter 3

Wiener Filters

1. Introduction
l

Wiener filters are a class of optimum linear filters which involve linear
estimation of a desired signal sequence from another related sequence.

In the statistical approach to the solution of the linear filtering problem,


we assume the availability of certain statistical parameters (e.g. mean
and correlation functions) of the useful signal and unwanted additive
noise. The problem is to design a linear filter with the noisy data as
input and the requirement of minimizing the effect of the noise at the
filter output according to some statistical criterion.

A useful approach to this filter-optimization problem is to minimize the


mean-square value of the error signal that is defined as the difference
between some desired response and the actual filter output. For
stationary inputs, the resulting solution is commonly known as the
Weiner filter.

The Weiner filter is inadequate for dealing with situations in which


nonstationarity of the signal and/or noise is intrinsic to the problem. In
such situations, the optimum filter has to be assumed a time-varying
form. A highly successful solution to this more difficult problem is
found in the Kalman filter.

2. Linear Estimation with Mean-Square Error Criterion


l

Fig. 3.1 shows the block schematic of a linear discrete-time filter W(z)
for estimating a desired signal d(n) based on an excitation x(n)
d(n)
x(n)

W(z)

y(n)

e(n)

Fig. 3.1
l

We assume that both x(n) and d(n) are random processes (discrete-time
random signals). The filter output is y(n) and e(n) is the estimation error.

To find the optimum filter parameters, the cost function or performance


function must be selected.
In choosing a performance function the following points have to be
considered
(1) The performance function must be mathematically tractable.
(2) The performance function should preferably have a single
minimum so that the optimum set of the filter parameters could be
selected unambiguously.

The number of minima points for a performance function is closely


related to the filter structure. The recursive (IIR) filters, in general,
result in performance function that may have many minima, whereas the
non-recursive (FIR) filters are guaranteed to have a single global
minimum point if a proper performance function is used.

In Weiner filter, the performance function is chosen to be


= E[ e(n) ]
2

This is also called mean-square error criterion

3. Wiener Filter the Transversal, Real-Valued Case

Fig. 3.2 shows a transversal filter with tap weights w0, w1, , wN-1 .
x(n)

x(n-1)

z-1

w0

z-1

z-1
wN-1

w1

y(n)
-

e(n)

d(n)

Fig. 3.2

Let

W = [ w0
X ( n) = [ xn

w1 ... wN 1 ]T
xn 1 ... x n N +1 ]T

....(1)

N 1

The output is y (n) = wi x(n i)


i =0

= W T X (n )
= X T (n )W

....(2)

Thus we may write


e(n) = d (n) y( n)
= d (n) X T (n)W ...(3)

The performance function, or cost function, is then given by


= E[e 2 (n )]
= E[(d (n ) W T X (n ))(d ( n) X T (n )W )]
= E[d 2 (n)] W T E[ X ( n)d ( n)] E[ X T (n) d (n)]W + W T E[ X (n) X T (n )]W ...(4)

Now we define the Nx1 cross-correlation vector


P E[ X (n)d (n)]
= [ p0

p1 ... p N 1 ]T

...(5)

and the NxN autocorrelation matrix


R E [ X (n) X T (n )]
r0, N 1
r1, N 1
M

L rN 1, N 1

...(6)

= E[d 2 ( n)] 2W T P + W T RW

...(7)

r00
r
10
=
M

rN 1,0

r01
r11
M
rN 1,1

Also we note that


E[d (n ) X T ( n)] = P T
W T P = P TW

Then we obtain

Equation (7) is a quadratic function of the tap-weight vector W with a


single global minimum.
Note that R has to be a positive definite matrix in order to have a unique
minimum point in the w-space.
l

Minimization of performance function


To obtain the set of tap weights that minimize the performance function,

= 0 , for i = 0,1,..., N 1
wi

we set

= 0

or

... (8)

where is the gradient vector defined as the column vector


=[

w0

w1

T
]
wN 1

and zero vector 0 is defined as N-component vector


0 = [0 0 L 0]T

Equation (7) can be expanded as


N 1

N 1 N 1

l =0

l =0 m=0

= E[d 2 (n) 2 p l wl + wl wm rlm

...(9)

N 1 N 1

and

w w
l =0 m=0

N 1 N 1

m lm

w w
l =0 m=0

can be expanded as
N 1 N 1

N 1

l =0 m=0
l i mi

m0
l i

r = wl wm rlm + wi wm rim + wi rii

m lm

...(10)

Then we obtain
N 1

= 2 p i + Wl (rli + ril )
wi
l =0

...(11)

for i = 0,1, 2,..., N 1

By setting
= 0 , we obtain
wi
N 1

W ( r
l

l =0

li

+ ril ) = 2 pi

...(12)

Note that
rli = E[ x (n l ) x (n i )] = xx (i l )

The symmetry property of autocorrelation function of real-valued signal,


we have the relation rli = ril
Equation (12) then becomes
N 1

r w
l =0

il

= pi

for i = 0,1,2,..., N 1

In matrix notation, we then obtain


RWop = P

...(13)

where Wop is the optimum tap-weight vector


Equation (13) is also known as the Wiener-Hopf equation,
which has the solution Wop = R 1 P
Assuming that R has inverse.
The minimum value of is
min = E [d 2 (n )] Wop P
T

= E[d 2 (n)] Wop RWop


T

...(15)

Equation (15) can also expressed as


min = E [d 2 (n )] P T R 1 P

...(16)

...(14)

Example 3.1 A Modelling problem

v(n)
Plant

d(n)
+

W(z)

x(n)

y(n)

e(n)

Assuming that
d (n ) = 2 x( n) + 3 x(n 1) + v (n)

and v(n) represents a stationary white process v2=0.1 ,


x(n) is also a white process, with unit variancex2=1
Note
By this assumption, the actual system function of the plant is 2+3z-1
Design an optimum filter (Weiner filter) to model the plant.
Question
How to choose the number of taps (N)
Once N is determined, the procedure to solve the problem is
straightforward.
If N=2 is chosen
p.55(textbook)
1 0

then R =

0 1
2
P=
3

and = 13.1 4 w0 6 w1 + w0 2 + w12


w

thus 0, 0 =
w0,1 3
min = 0.1

4. Principle of Orthogonality An Alternative Approach to the Design


of Wiener Filter
l

The cost function (or performance function) is given by


= E[e 2 (n )]

...(17)

e( n )
= E[2e(n )
]
w i
wi

...(18)

for i = 0,1,2,..., N 1

where e(n)=d(n)-y(n)
Since d(n) is independent of wi , we get
y (n)
e( n )
=
wi
wi
= x( n i )

...(19)

Then we obtain

= 2 E [e (n) x (n i)]
wi

...(20)

for i = 0,1,2,..., N 1
l

When the Wiener filter tap,


weights are set to their optimal values,

=0.
wi

Hence, if e0(n) is the estimation error when wi are set to their optimal
values, then equation (20) becomes
E[ e0 ( n ) x( n i )] = 0

for i = 0,1,..., N 1

...(21)

That is,
the estimation error is uncorrelated with the filter tap inputs, x(n-i).
This is known as the principles of orthogonality.
l

We can also show that the optimal filter output is also uncorrelated with
the estimation error.
That is, E[ e0 (n ) y 0 (n )] = 0

....(22)

This result indicates that the optimized Weiner filter output and the
estimation error are orthogonal.

5. Normalized Performance Function


l

If the optimal filter tap weights are expressed by w0,l , l=0,1,2, ,N-1.
The estimation error is then given by
N 1

e0 (n ) = d (n ) w0,l x (n l )

...(23)

l =0

and then
d ( n) = e 0 ( n ) + y 0 ( n)

...(24)

E[ d 2 ( n)] = E [e0 (n )] + E [ y 0 ( n)] + 2 E[ e0 ( n ) y 0 (n )]


2

We may note that


E[ e0 (n )] = min
2

...(26)

and we obtain
min = E[d 2 (n )] E [ y0 (n )]
2

...(27)

Define ? as the normalized performance function


=

E [d 2 (n )]

...(28)

?=1 when y(n)=0

? reaches its minimum value, ?min , when the filter tap-weights are
chosen to achieve the minimum mean-squared error.
This gives min = 1

E [ y0 (n )]
E [d 2 ( n )]

We have 0 < ?min < 1 .

...(29)

6. Wiener Filter Complex-Valued Case

In many practical applications, the random signals are complex-valued.


For example, the baseband signal of QPSK & QAM in data
transmission systems.

In the Wiener filter for processing complex-valued random signals, the


tap-weights are assumed to be complex variables.

The estimation error, e(n), is also complex-valued. We may write


= E [ e(n ) ] = E [e(n )e* (n )]
2

...(30)

The tap-weight wi is expressed by


wi=wi,R+jwi,I

(31)

The gradient of a function with respect to a complex variable w=wR+jwI


is defined as

+ j
wR
w I

...(32)

The optimum tap-weights of the complex-valued Wiener filter will be


obtained from the criterion:
cwi = 0

That is,

cw

for i = 0,1,2,..., N 1

= 0 and
=0
wi , R
wi ,R

...(33)

Since = E [e( n )e* (n )] , we have


cwi = E [e(n ) cwi e* (n ) + e * (n ) cwi e( n )]

...(34)

Noting that
N 1

e(n ) = d (n ) wk x (n k )

...(35)

cwi e (n ) = x( n i ) cwi wi

...(36)

cwi e * (n ) = x * (n i ) cwi w *i

...(37)

k =0

Applying the definition (32), we obtain

cwi wi =

wi
wi
+ j
= 1 + j ( j) = 0
wi , R
wi , I

and cw wi* =
i

wi*
wi*
+ j
= 1 + j( j ) = 2
wi ,R
wi , I

Thus, equation (34) becomes


cwi = 2 E [e(n ) x * (n i)]

...(38)

The optimum filter tap-weights are obtained when cw = 0 . This givens


i

E[ e0 ( n ) x * (n i )] = 0

for i = 0,1,2,..., N 1

...(39)

where e0(n) is the optimum estimation error.


l

Equation (39) is the principle of orthogonality for the case of


complex-valued signals in wiener filter.

The Wiener-Hopf equation can be derived as follows:


Define x (n ) [ x (n ) x( n 1) L x (n N + 1)]T
and

w ( n) [w0*

w1* L w*N 1 ]T

...(40)

...(41)

We can also write x ( n ) [ x * ( n ) x * ( n 1) L x * (n N + 1)]H


and

w ( n) [w0

w1

L wN 1 ]H

...(43)

where H denotes complex-conjugate transpose or Hermitian.


Noting that
e(n ) = d (n ) w0H x( n)

and E[e0 ( n) x (n )] = 0

...(44)

from equation(39)

...(45)

We have
E[ x ( n)(d * (n ) x H (n )w0 )] = 0

and then
Rw0 = p

...(47)

...(46)

...(42)

where R = E[ x (n ) x H (n )]
and p = E [ x (n )d * (n )]

Equation (47) is the Wiener-Hopf equation for the case of


complex-valued signals.

The minimum performance function is then expressed as


min = E [d 2 (n )] w0H R w0

...(48)

Remarks:
In the derivation of the above Wiener filter we have made assumption
that it is causal and finite impulse response, for both real-valued and
complex-valued signals.

7. Unconstrained Weiner Filters


l

The block diagram of a Wiener filter is shown in Fig. 3.1

d(n)
x(n)

W(z)

y(n)

e(n)

Fig. 3.1
We assume that the filter W(z) may be non-causal and/or IIR.
l

We consider only the case the signals and the system parameters are
real-valued.
Moreover, we assume that the complex variable z remains on the unit
circle, i.e. z = 1 , thus z*=z-1 , z = 1 .

7.1 Performance Function


l

The performance function is defined as


= E[ e(n) ]
2

Where e(n)=d(n)-y(n)
In terms of autocorrelation and cross-correlation functions, we have
= E[d 2 ( n)] + E[ y 2 (n )] 2 E[ y (n)d (n)]
= dd (0) + yy (0) 2 yd (0)

...(49)

Using the inverse Z-transform relations, we have


= dd (0) +

1
dz
1
dz
yy ( z ) 2
yd ( z )

2j c
z
2j c
z

...(50)

Since Y(z)=X(z)W(z), yd ( z ) = W ( z ) xd ( z)
If z is selected to be on the unit circle in the z-plane, then
yy ( z ) = W ( z ) xx ( z ),
2

W ( z ) = W ( z )W * ( z ) and W * ( z ) = W ( z 1 )
2

Using these relations in equation(50), we obtain


= dd (0) +
= dd (0) +

1
dz
1
dz
2
W ( z ) xx ( z ) 2
W ( z ) xd ( z )

2j c
z
2j c
z
1
dz
[W * ( z) xx ( z ) 2 xd ( z )]W ( z)

2j c
z

...(51)

Where the contour of integration, c , is the unit circle.


l

The performance function given in equation(51) covers IIR and FIR


This is a most general form of the Wiener filter performance function.
Example 3.2
From general expression of , eq.(51) to demonstrate the special case
of real-valued FIR system.
Consider the case where the Wiener filter is an N-tap FIR filter with
N 1

system function W ( z ) = wl z l

...(52)

l=0

Using eq.(52) in the expression of l , given by eq.(51), then


N 1
N 1
dz
dz
1
1
l
+m
= dd (0) +
( wl z )( wm z ) xx ( z) 2
( wl z l ) xd ( z )

z
z
2j c l =0
2j c l =0
N 1 N 1

= dd (0) + wl w m
l =0 m=0

N 1
1
1
m l 1
z
z
dz
wl

xd ( z ) z l 1 dz
(
)
2

xx

2j c
2j c
l =0

Using the inverse z-transform relation, this gives


N 1 N 1

N 1

l =0 m=0

l =0

= dd (0) + wl wm xx (m l ) 2 wl xd ( l)

Now using the notations


dd (0) = E [d 2 (n )]
xd (l ) = p l
xx (m l ) = xx (l m) = rlm

We obtain the expression


N 1

N 1 N 1

l =0

l =0 m=0

= E[d 2 ( n)] 2 pl wl + wl wm rlm

Note that
p = E[ x (n)d (n)]
Thus pl = E[ x( n l )d (n )]

By definition, xy ( k ) = E[ x (n) y * (n k )]
xd (l ) = E[ x (n)d (n + l )] = E[ x( n l )d ( n)]

Example 3.3
Modelling with an IIR filter
v(n)
G(z)
Plant
x(n)

1 w0 z 1
1 w1 z 1

d(n)
+
y(n)

e(n)

1 w0 z 1
The plant is modeled by an IIR filter with W ( z ) =
1 w1 z 1

We assume that all involved signals and system parameters are


real-valued.
x(n) : input sequence, white process with zero-mean, x2 = 1 .
v(n) : additive noise.
x(n) & v(n) are uncorrelated.
Thus, xx ( z ) = 1 and vx ( z ) = 0
G(z) : system function of the unknown plant.
xd ( z) = G( z 1 ) xx ( z )

Using equation(51) to obtain the performance function :


= dd (0) +

w1 w0 1 w 0 w1 w0
w w0
w

+
2[ 1
G( w11 ) + 0 G( )]
2
w1
w1
w1
w1
1 w1

We may find that there can be many local minima, and searching for the
global minimum of may not be a trivial task.

7.2 Optimum transfer function


l

From the principle of orthogonality, we have


E[e0 (n) x( n i)] = 0

where e0 (n) = d ( n)

for i = 0,1, 2,..., N 1

l =

0 ,l

x (n l )

...(52)

...(53)

Here we assume that all involved signals are real-valued.


l

Combining eq.(52)&eq.(53), we obtain

l =

0 ,l

E[ x( n l ) x( n i)] = E[d (n ) x(n i)]

...(54)

with E[ x( n l ) x( n i)] = xx (i l )
and

E[d (n ) x(n i)] = dx (i )

and taking Z-transform on both sides of equation(54), we get


xx ( z )W0 ( z) = dx ( z )

...(55)

This is referred to as the Wiener-Hopf equation of the unconstrained


Wiener filtering.
l

The optimum unconstrained Wiener filter is given by


W 0 ( z) =

dx ( z )
xx ( z )

...(56)

and
W0 (e jw ) =

dx (e jw )
xx (e jw )

...(57)

This is the frequency response of the Wiener filter

dx (e jw ) : cross-power spectral density of d(n) and x(n)


xx (e jw ) : power spectral density of x(n)

8. Application of Wiener Filters Modelling

Consider the modeling problem depicted in Fig. 3.8

0 (n)
(n)

G(z)
Plant

i(n)

x(n)

W(z)

d(n)
+

y(n)
-

e(n)

(n), 0(n), andi(n) are assumed to be stationary, zero-mean and


uncorrelated with one another.

The input to Wiener filter is given by


x (n ) = (n ) + i (n )

...(58)

and the desired output is given by


d ( n) = g n * (n ) + 0 (n )

...(59)

where gn is the impulse response sample of the plant.


l

The optimum unconstrained Wiener filter transfer function


W0 ( z ) =

dx ( z )
xx ( z )

...(60)

Note that
xx ( k ) = E[ x( n) x( n k )]
= E [( ( n ) + i (n ))( (n k ) + i (n k ))]
= E [ ( n ) ( n k )] + E [ (n )i ( n k )] + E [ i (n ) ( n k )] + E [ i (n ) i (n k )]
= (k ) + ii ( k )

...(61)

Taking Z-transform on both sides of eq.(61), we get


xx ( z ) = ( z ) + ii ( z )

...(62)

To calculate W0 ( z ) =

dx ( z )
, we must first find the expression for dx ( z) .
xx ( z )

We can show that


dx ( z ) = d ' ( z )

where d(n) is the plant output when the additive noise0(n) is


excluded from that.
Moreover, we have
d ' ( z ) = G( z) ( z )

Thus dx ( z ) = G ( z ) ( z )
And we obtain
W0 ( z ) =

( z )
( z ) + ii ( z )

G( z )

...(63)

We note that W0(z) is equal to G(z) only when ( z ) is equal to zero.


i i

That is, wheni(n) is zero for all values of n.


The noise sequencei(n) may be thought of as introduced by a
transducer that is used to get samples of the plant input.
l

Replacing z by ejw in equation(63), we obtain


(e jw )
W0 ( e ) =
G(e jw )
jw
jw
( e ) + ii (e )
jw

( e jw )
l Define K (e )
(e jw ) + ii ( e jw )
jw

We obtain
W0(ejw)=K(ejw)G(ejw)

...(64)

...(65)

With some mathematic manipulation, we can find the minimum


mean-square error, min , expressed by
min = 00 (0) +

1
2

(1 K (e jw )) (e jw ) G( e jw ) dw

...(66)

The best performance that one can expect from the unconstrained
Wiener filter is
min = 00 (0)

and this happens wheni(n)=0 .

The Wiener filter attempts to estimate that part of the target signal d(n)
that is correlated with its own input x(n) and leaves the remaining part
of d(n) (i.e. 0(n)) unaffected. This is known as the principles of
correlation cancellation .

9. Application of Wiener Filters Inverse Modelling

Fig. 3.9 depicts a channel equalization scenario.


d(n)=s(n)
?(n)
s(n)

y(n)

H(z)

x(n)

W(z)

+ e(n)
-

s(n) : data samples


H(z) : System function of communication channel
?(n) : additive channel noise which is uncorrelated with s(n)
W(z) : filter function used to process the received noisy signal samples
to recover the original data samples.
l

When the additive noise at the channel output is absent, the equalizer
has the following trivial solution:
W0 ( z ) =

1
H ( z)

...(67)

This implies that y(n)=s(n) and thus e(n)=0 for all n.


Demonstration: Fig. 3.10
l

When the channel noise, ?(n), is non-zero, the solution provided by


equation(67) may not be optimal.
x (n ) = hn * s(n ) + (n )

...(68)

and d ( n) = s( n ) ...(69)
where hn is the impulse response of the channel, H(z).
From equation(68), we obtain
xx ( z ) = ss ( z ) H ( z) + ( z )
2

...(70)

Also,
dx ( z ) = sx ( z ) = H ( z 1 ) ss ( z )

With z = 1 , we may also write


dx ( z) = H * ( z ) ss ( z )

...(72)

...(71)

and then
W0 ( z) =

H * ( z) ss ( z )

...(73)

ss ( z ) H ( z ) + ( z )
2

This is the general solution to the equalization problem when there is no


constraint on the equalizer length and, also, it may be let to be
non-causal.
l

Equation(73) can be rewritten as


W0 ( z ) =

1+

1
( z )
ss ( z ) H ( z )

1
H (z)

...(74)

Let z=ejw and define the parameter


(e )
jw

ss (e jw ) H ( e jw )

(e jw )

where ss ( e jw ) H (e jw )

...(75)

and (e jw ) are the signal power spectral

density and the noise power spectral density, respectively, at the channel
output.
We obtain
W0 (e jw ) =

(e jw )
1
jw
1 + (e ) H (e jw )

...(76)

Note that ?(ejw) is a non-negative quantity, since it is the signal-to-noise


power spectral density ratio at the equalizer input.
Also, 0

( e jw )
1
1 + (e jw )

...(77)

Cancellation of ISI and noise enhancement


Consider the optimized equalizer with frequency response given by
W0 (e jw ) =

(e jw )
1
jw
1 + (e ) H (e jw )

In the frequency regions where the noise is almost absent, the value of
is very large and hence
W0 (e jw )

1
H (e jw )

The ISI will be eliminated without any significant enhancement of noise.


On the other hand, in the frequency regions where the noise level is
high, the value of ?(ejw) is not large and hence the equalizer does not
approximate the channel inverse well. This is of course, to prevent noise
enhancement.

10. Noise Cancellation


l

Fig. 3.12 shows a typical noise canceller block diagram

s(n)

d(n)
Primary input

e(n)
-

H(z)
G(z)
?(n)

x(n) W(z)
Reference
input

s(n) : signal source


) : noise source
s(n) & ?(n) are uncorrelated
H(z) & W(z) are two system functions used to mixed the two input
signals to form d(n) & x(n).
d(n) : primary input
x(n) : Reference input
x(n)=?(n)+hn*s(n)
d(n)=s(n)+gn*?(n)
l

(78)
(79)

Since s(n) & ?(n) are uncorrelated with each other, we obtain
dx ( z ) = sdx ( z ) + dx ( z )

where

...(81)

is dx ( z) when ?(n)=0 for all n,

and dx ( z) is dx ( z) when s(n)=0 for all n.


This is because s(n) and ?(n) are uncorrelated with each other, their
contribution in dx ( z) can be considered separately.
Thus, we obtain
sdx ( z) = H * ( z ) ss ( z )

...(82)

and dx ( z) = G ( z ) ( z )

...(83)

Recall that z = 1 . Finally, we get


dx ( z) = H * ( z ) ss ( z) + G( z ) (z)

and W0 ( z ) =

...(84)

H * ( z ) ss ( z ) + G( z ) ( z )
( z) + ss ( z ) H ( z)

...(85)

Equation(85) for noise cancellation may be thought of as a


generalization of the results we obtained for the modeling and inverse
modeling scenarios, given by equation(63) and (73).

To minimize the mean-square value of the output error, we must strike a


balance between noise cancellation and signal cancellation at the output
of the noise canceller.
Cancellation of the noise ?(n) occurs when the Wiener filter W(z) is
chosen to be close to G(z), and cancellation of the signal s(n) occurs
when the Wiener filter W(z) is close to the inverse of H(z). In the sense,
we may note that the noise canceller treats s(n) and ?(n) without making
any distinction between them.

Define
pri (e jw ) : signal-to-noise PSD at primary input

: signal-to-noise PSD at reference input


out (e jw ) : signal-to-noise PSD at output
pri (e jw ) =

ss (e jw )
2

G (e jw ) ( e jw )

...(86)

ref ( e jw ) =

H (e jw ) ss (e jw )
(e jw )

...(87)

To calculate out (e jw ) , we note that s(n) reaches the canceller output

through two routes one direct and one through the cascade of H(z)
and W(z) :
2

see (e jw ) = 1 H (e jw )W (e jw ) ss (e jw )

...(88)

Similarly, ?(n) reaches the output through the routes G(z) and W(z) :
2

ee (e jw ) = G( e jw ) W (e jw ) ( e jw )

...(89)

Replacing W(ejw) by W0(ejw), we obtain


2

(e ) =
s
ee

jw

2
1 G( e jw )H ( e jw )
(e jw )
2

(e jw ) + H ( e jw ) ss (e jw )
2

ee

and (e ) =
jw

(e jw )

H ( e jw ) 1 G(e jw ) H (e jw ) 2ss ( e jw )
2

(e jw ) + H (e jw ) ss (e jw )

( e jw )

Hence, out (e jw ) can now be obtained as


out (e jw )
=
=

see (e jw )
ee (e jw )
(e jw )
2

H ( e jw ) ss ( e jw )
1
ref ( e jw )

...(90)

This is known as power inversion . ( Widrow et al, 1975)


** This result shows that the noise canceller works better when
ref ( e jw ) is low.

Example 3.6
Two Omni-directional Antenna in a Receiver
A

s(n)

d(n)

?0
?(n)

x(n)

B
0

90

~
x(n)
x (n)

desired signal
interferer signal
(jammer)

w0

y(n)

e(n)

w1

s(n)= a(n)cos nw0


?(n)= (n)cos nw0

s(n) arrives the receiver (antennas) in the direction perpendicular to the


line connecting A & B.
?(n) arrives at an angle ?0 with respect to the direction of s(n).

a(n) and (n) are narrowband baseband signals.


Thus s(n) & ?(n) can be treated as narrowband signals concentrated
around w=w0 .

since s(n) & ?(n) may be treated as single tones, and thus a filter with
two degrees of freedom is sufficient for optimal filtering.

Wiener filter coefficients can be determined as follows.


s(n) arrives at the same time at both omni-antennas. ?(n) arrives at B
first and arrives at A with a delay
0 =

l sin 0
c

the delay is normalized by the time step T which corresponds to one


increment of time index n. This gives
0 =

l sin0
cT

Thus
d ( n) = (n ) cos nw0 + ( n) cos[(n 0 )w0 ]
s (n ) = ( n) cos nw0 + (n ) cos nw0
x~( n) = (n ) sin nw0 + (n ) sin nw0
E [ x 2 ( n)]
E[ x( n ) x~(n )]
R=

~
E [ x~ 2 ( n )]
E [ x (n ) x ( n )]

With some mathematical manipulations, we obtain


R=

2 + 2 1 0
0 1
2

where 2 and 2 are variance of a(n) and (n), respectively.


2
2
E [d ( n) x( n)[ 1 + cos 0 w0
Also, P =

=
2
~
E [d (n ) x ( n )] 2 sin 0 w0

The Wiener-Hopf equation R W0 = P


( 2 + 2 cos 0 w0 )
1
2
2
2
sin 0 w0
+

gives W0 =

The signal-to-noise ratio at the output is equal to 2 / 2


while the signal-to-noise ratio at the reference input is equal to 2 / 2 .

You might also like