Professional Documents
Culture Documents
Lect11 PDF
Lect11 PDF
1
(11-4)
To determine, the limiting case FX ( x | Y = y ), we can let y1 = y
and y2 = y + y in (11-4).
2
This gives x y + y
x
FX ( x | y < Y y + y ) = y
f XY ( u , v ) dudv
f XY ( u , y ) du y
(11-5)
y + y
fY ( y ) y
y
f Y ( v ) dv
and hence in the limit x
F X ( x | Y = y ) = lim F X ( x | y < Y y + y ) =
f XY ( u , y ) du
. (11-6)
y 0 fY ( y )
(To remind about the conditional nature on the left hand
side, we shall use the subscript X | Y (instead of X) there).
Thus x
F X |Y ( x | Y = y ) =
f XY ( u , y ) du
.
(11-7)
fY ( y )
Differentiating (11-7) with respect to x using (8-7), we get
f XY ( x , y )
f X |Y ( x | Y = y ) = . (11-8)
fY ( y )
3
It is easy to see that the left side of (11-8) represents a valid
probability density function. In fact
f XY ( x , y )
f X |Y ( x | Y = y ) = 0 (11-9)
fY ( y )
and +
+ f XY ( x , y ) dx fY ( y )
f X |Y ( x | Y = y ) dx =
fY ( y )
=
fY ( y )
= 1, (11-10)
6
From (11-16) - (11-18), we get
f XY ( x , y ) 1
f X |Y ( x | y ) = = , 0 < x < y < 1, (11-19)
fY ( y ) y
and
f XY ( x , y ) 1
fY |X ( y | x ) = = , 0 < x < y < 1. (11-20)
f X ( x) 1 x
f ( | r ) = 2
= 2
f ( r | ) f ( ) d 1
( r ) 2 / 2
d
2
e
0
2 0
( r ) 2 / 2
= ( r )e 0 < < 2 ,
2
, (11-26)
where
2
(r) = 2
.
( r ) 2 / 2
d
2
e
0
9
Notice that the knowledge about the observation r is
reflected in the a-posteriori p.d.f of in Fig. 11.2 (b). It is
no longer flat as the a-priori p.d.f in Fig. 11.2 (a), and it
shows higher probabilities in the neighborhood of = r.
f ( ) f |r ( | r )
1
2
2 =r
(a) a-priori p.d.f of (b) a-posteriori p.d.f of
Fig. 11.2
Conditional Mean:
We can use the conditional p.d.fs to define the conditional
mean. More generally, applying (6-13) to conditional p.d.fs
we get
10
+
E ( g( X ) | B ) = g ( x ) f X ( x | B )dx. (11-27)
= E (( X X |Y ) 2 | Y = y ). (11-30)
we shall illustrate these calculations through an example.
11
Example 11.3: Let
1, 0 < | y |< x < 1,
f XY ( x , y ) = (11-31)
0, otherwise .
Determine E ( X | Y ) and E (Y | X ). y
Solution: As Fig. 11.3 shows, f XY ( x , y ) = 1
in the shaded area, and zero elsewhere.
From there x
x 1
f X ( x) = f XY ( x, y )dy = 2 x, 0 < x < 1,
x
and 1
fY ( y ) = 1 dx = 1 | y |, | y| < 1, Fig. 11.3
| y|
This gives
f XY ( x , y ) 1
f X |Y ( x | y ) = = , 0 < | y |< x < 1, (11-32)
fY ( y ) 1 | y |
and
f XY ( x , y ) 1 (11-33)
fY |X ( y | x ) = = , 0 < | y |< x < 1 .
f X ( x) 2x
12
1 x
Hence E(X |Y ) = x f X |Y ( x | y ) dx = | y| (1 | y |) dx
2 1
1 x 1 | y |2 1+ | y |
= = = , | y |< 1 . (11-34)
(1 | y |) 2 | y|
2 (1 | y |) 2
2 x
x y 1 y
E (Y | X ) = yf Y | X ( y | x ) dy = x 2 x dy =
2x 2
= 0, 0 < x < 1. (11-35)
x
g ( x ) f ( x | y ) dx
X |Y f Y ( y ) dy
E ( g ( X )|Y = y )
E ( g ( X ) | Y = y ) f Y ( y ) dy = E {E ( g ( X ) | Y = y )} .
+
=
(11-37) 13
Obviously, in the right side of (11-37), the inner
expectation is with respect to X and the outer expectation is
with respect to Y. Letting g( X ) = X in (11-37) we get the
interesting identity
E ( X ) = E { E ( X | Y = y )}, (11-38)
14
Conditional mean turns out to be an important concept in
estimation and prediction theory. For example given an
observation about a r.v X, what can we say about a related
r.v Y ? In other words what is the best predicted value of Y
given that X = x ? It turns out that if best is meant in the
sense of minimizing the mean square error between Y and
its estimate Y , then the conditional mean of Y given X = x,
i.e., E (Y | X = x ) is the best estimate for Y (see Lecture 16
for more on Mean Square Estimation).
15