Professional Documents
Culture Documents
E-mail: do-hong@hcmut.edu.vn
z-Transform.
∑ u (n)u
i = −∞
1 2 (n − i ) ↔ U1 ( z )U 2 ( z ) (1.4)
ROC of (1.4): intersection of ROC of U1(z) and ROC of U2(z).
LTI u(n)=h(n)
t=0 Filter
h(n)
When input sequence v(n) and output sequence u(n) are related
by difference equation of order N:
N N
∑ a u (n − j ) =∑ b v(n − j )
j =0
j
j =0
j
(1.8)
aj, bj: constant coefficients. Applying z-transform, obtaining:
N N
U ( z)
∑a z
j =0
j
−j
a ∏ (1 − c z
k
−1
)
H ( z) = = = 0 k =1
(1.9)
V ( z) N N
∑b z ∏ (1 − d
−j b0 −1
j k z )
j =0 k =1
a1 a2 … aM-1 aM
u(n)
+ + … + +
FIR filter
z-1
+ a1 u(n-1)
z-1
+ a2 u(n-2)
.
.
.
+ aM-1 u(n-M+1)
z-1
IIR filter
aM u(n-M)
∑ h( k ) < ∞
k = −∞
(1.11)
Im
Axioms of Probability.
Repeated Trials.
A ∪ B = { ξ | ξ ∈ A or ξ ∈ B}
A ∩ B = { ξ | ξ ∈ A and ξ ∈ B}
and
A = { ξ | ξ ∉ A}
A
A B A B A
A∪ B A∩ B A
If A ∩ B = φ , the empty set, then A and B are said to be mutually
exclusive (M.E).
A partition of Ω is a collection of mutually exclusive subsets of Ω such
that their union is Ω.
Ai ∩ A j = φ , and ∪A
i =1
i = Ω.
A1
A2
A B Ai
Aj An
A∩ B =φ
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 31 BG, CH, ÑHBK
2. Review of Probability Theory: Basics (5)
De-Morgan’s Laws:
A∪ B = A∩ B ; A∩ B = A∪ B
A B A B A B A B
A∪ B A∪ B A∩ B
For example, from (ii) we have A ∈ F , B ∈ F , and using (iii) this gives
A ∪ B ∈ F ; applying (ii) again we get A ∪ B = A ∩ B ∈ F , where we
have used De Morgan’s theorem in slide 22.
(Note that (iii) states that if A and B are mutually exclusive (M.E.)
events)
We can also view the event B as new knowledge obtained from a fresh
experiment. We know something about A as P(A). The new information
is available in terms of B. The new information should be used to
improve our knowledge/understanding of A. Bayes’ theorem gives the
exact mechanism for incorporating such new information.
∑i =1
P ( B | Ai ) P ( Ai )
s1
s2
Input Output
s3
Solution:
a. Let Ai = “Switch Si is closed”. Then P ( Ai ) = p , i = 1 → 3 .
Since switches operate independently, we have
P ( Ai A j ) = P ( Ai ) P ( A j ); P ( A1 A2 A3 ) = P ( A1 ) P ( A2 ) P ( A3 ).
P ( R | A1 ) P( A1 ) (2 p − p 2 )(1 − p ) 2 − 2 p + p2
P( A1 | R ) = = = .
P( R ) 3 p − 3 p2 + p3 3 p − 3 p2 + p3
Solution: Let (Ω, F, P) be the probability model for a single trial. The
outcome of n experiments is an n-tuple
ω = { ξ1 , ξ 2 , , ξ n }∈ Ω0 ,
Pn (k )
n = 12, p = 1 / 2.
From Figure, the most probable value of k is that number which maximizes
Pn(k) in Bernoulli formula. To obtain this value, consider the ratio
Pn (k − 1) n! p k −1q n −k +1 (n − k )! k! k q
= k n −k
= .
Pn ( k ) (n − k + 1)! ( k − 1)! n! p q n − k +1 p
Pn (k − 1) n! p k −1q n −k +1 (n − k )! k! k q
= k n −k
= .
Pn ( k ) (n − k + 1)! ( k − 1)! n! p q n − k +1 p
Ω
ξ
A
X (ξ )
x B R
Obviously, if the set A = X-1(B) also belongs to the associated field F, then it
is an event and the probability of A is well defined; in that case we can say
Random Variable (r.v): A finite single valued function X(.) that maps
the set of all experimental outcomes Ω into the set of real numbers R is said
to be a r.v, if the set { ζ | X(ζ) ≤ x } is an event ∈ F for every x in R.
is also an event. All events have well defined probability. Thus the
probability of the event { ζ | X(ζ) ≤ x } must depend on x. Denote
P{ ξ | X (ξ ) ≤ x } = FX ( x) ≥ 0
which is referred to as the Probability Distribution Function (PDF)
associated with the r.v X.
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 53 BG, CH, ÑHBK
2. Review of Probability Theory: R. V (4)
Distribution Function: If g(x) is a distribution function, then
(i) g(+∞) = 1; g(-∞) = 0
(ii) if x1 < x2, then g(x1) ≤ g(x2)
(iii) g(x+) = g(x) for all x.
It is shown that the PDF FX(x) satisfies these properties for any r.v X.
pi = P{X = xi } = FX ( xi ) − FX ( xi− ).
Example: Toss a coin. Ω = {H, T}. Suppose the r.v X is such that X(T) = 0,
X(H) = 1. Find FX(x).
Solution: For x < 0, {X(ζ) ≤ x } = {φ}, so that FX(x) = 0.
0 ≤ x < 1, { X (ξ ) ≤ x } = { T }, so that FX ( x) = P{ T } = 1 − p,
x ≥ 1, { X (ξ ) ≤ x } = { H , T } = Ω, so that FX ( x) = 1.
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 56 BG, CH, ÑHBK
2. Review of Probability Theory: R. V (7)
Example: A fair coin is tossed twice, and let the r.v X represent the number
of heads. Find FX(x).
Solution: In this case Ω = {HH, HT, TH, TT}, and X(HH) = 2, X(HT) =
X(TH) = 1, X(TT) = 0.
x < 0, {X (ξ ) ≤ x} = φ ⇒ FX ( x) = 0,
1
0 ≤ x < 1, {X (ξ ) ≤ x} = { TT } ⇒ FX ( x) = P{ TT } = P (T ) P(T ) = ,
4
3
1 ≤ x < 2, {X (ξ ) ≤ x} = { TT , HT , TH } ⇒ FX ( x) = P{ TT , HT , TH } = ,
4
x ≥ 2, {X (ξ ) ≤ x} = Ω ⇒ FX ( x) = 1.
P{X = 1} = FX (1) − FX (1− ) = 3 / 4 − 1 / 4 = 1 / 2.
FX (x)
1
3/ 4
1/ 4
x
1 2
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 57 BG, CH, ÑHBK
2. Review of Probability Theory: R. V (8)
Probability Density Function (p.d.f): The derivative of the distribution
function FX(x) is called the probability density function fX(x) of the r.v X.
Thus dF ( x )
f X ( x) = X .
dx
Since
dFX ( x ) F ( x + Δx ) − FX ( x )
= lim X ≥ 0,
dx Δx → 0 Δx
from the monotone-nondecreasing nature of FX(x), it follows that fX(x) ≥ 0
for all x. fX(x) will be a continuous function, if X is a continuous type r.v.
However, if X is a discrete type r.v, then its p.d.f has the general form
f X ( x)
f X ( x) = ∑ piδ ( x − xi ), pi
i
xi x
xi x
We also obtain
x
FX ( x) = ∫ f x (u )du.
−∞
which justifies its name as the density function. Further, we also get
P{ x1 < X (ξ ) ≤ x2 } = FX ( x2 ) − FX ( x1 ) = ∫ f X ( x )dx.
x2
x1
x1
FX (x) f X (x)
1
x1 x2 x x
x1 x2
(a) (b)
Often, r.vs are referred by their specific density functions - both in the
continuous and discrete cases - and in what follows we shall list a number
of them in each category.
This is a bell shaped curve, symmetric around the parameter μ and its
distribution function is given by
x 1 ⎛x−μ⎞
FX ( x ) = ∫
2 2
e −( y − μ ) / 2σ dy = G⎜ ⎟,
−∞
2πσ 2
⎝ σ ⎠
x 1 − y2 / 2
where G ( x ) = ∫ e dy is often tabulated. Since fX(x) depends on
−∞ 2π
two parameters μ and σ2, the notation X ∼ N(μ, σ2) will be used to
represent fX(x). f X (x)
x
μ
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 61 BG, CH, ÑHBK
2. Review of Probability Theory: R. V (12)
2. Uniform: X ∼ U(a,b), a < b, if
⎧⎪ 1 f X (x)
, a ≤ x ≤ b, 1
f X ( x) = ⎨ b − a
b−a
⎪⎩ 0, otherwise. x
a b
3. Exponential: X ∼ ε (λ), if
⎧⎪ 1 − x / λ f X (x)
e , x ≥ 0,
f X ( x) = ⎨ λ
⎪⎩ 0, otherwise. x
⎧ 1
⎪ x a −1 (1 − x )b−1 , 0 < x < 1,
f X ( x ) = ⎨ β ( a , b) x
⎪⎩ 0, otherwise. 0 1
6. Chi-Square: X ∼ χ 2(n) if f X ( x)
⎧ 1
⎪ n/2 x n / 2−1e − x / 2 , x ≥ 0,
f X ( x ) = ⎨ 2 Γ( n / 2)
⎪⎩ 0, otherwise. x
8. Nakagami – m distribution:
⎧ 2 ⎛ m ⎞ m 2 m −1 − mx 2 / Ω
⎪ , x≥0
f X ( x ) = ⎨ Γ( m ) ⎜⎝ Ω ⎟⎠
x e
⎪ 0 otherwise
⎩
x
μ
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 64 BG, CH, ÑHBK
2. Review of Probability Theory: R. V (15)
10. Laplace: fX (x)
1 −|x|/ λ
f X ( x) = e , − ∞ < x < +∞.
2λ
x
t
12. Fisher’s F-distribution:
⎧Γ{( m + n ) / 2} m m / 2 n n / 2 z m / 2 −1
⎪ , z≥0
f z ( z) = ⎨ Γ ( m / 2) Γ ( n / 2) ( n + mz ) (m+n) / 2
⎪ 0 otherwise
⎩
⎛n⎞
P( X = k ) = ⎜⎜ ⎟⎟ p k q n −k , k = 0,1,2, , n.
⎝k ⎠ k
12 n
3. Poisson: X ∼ P(λ) if
P(X = k)
−λ λ
k
P( X = k ) = e , k = 0,1,2, , ∞.
k!
4. Discrete-Uniform
1
P( X = k ) = , k = 1,2, , N.
N
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 66 BG, CH, ÑHBK
2. Review of Probability Theory: Function of R. V (1)
Let X be a r.v defined on the model (Ω, F, P) and suppose g(x) is a
function of the variable x. Define
Y = g(X)
Example 1: Y = aX + b
Suppose a > 0,
⎛ y −b⎞ ⎛ y −b⎞
FY ( y ) = P (Y (ξ ) ≤ y ) = P (aX (ξ ) + b ≤ y ) = P⎜ X (ξ ) ≤ ⎟ = FX⎜ ⎟.
⎝ a ⎠ ⎝ a ⎠
and
1 ⎛ y −b⎞
fY ( y ) = fX ⎜ ⎟.
a ⎝ a ⎠
For all a
1 ⎛ y −b⎞
fY ( y ) = fX ⎜ ⎟.
|a | ⎝ a ⎠
FY ( y ) = P ( x1 < X (ξ ) ≤ x2 ) = FX ( x2 ) − FX ( x1 ) Y = X2
= FX ( y ) − FX ( − y ), y > 0. y
⎪ X + c, X ≤ −c.
⎩
In this case
P (Y = 0) = P ( −c < X (ξ ) ≤ c ) = FX (c ) − FX ( −c ).
FX (x )
For y > 0, we have x > c, and Y(ζ) = X(ζ) – c, so that
FY ( y ) = P (Y (ξ ) ≤ y ) = P ( X (ξ ) − c ≤ y ) x
= P ( X (ξ ) ≤ y + c ) = FX ( y + c ), y > 0.
Similarly y < 0, if x < - c, and Y(ζ) = X(ζ) + c, so that FY ( y )
FY ( y ) = P (Y (ξ ) ≤ y ) = P ( X (ξ ) + c ≤ y )
= P ( X (ξ ) ≤ y − c ) = FX ( y − c ), y < 0. y
Thus
⎧ f X ( y + c ), y > 0,
⎪
f Y ( y ) = ⎨[ FX ( c ) − FX ( − c )]δ ( y ),
⎪ f ( y − c ), y < 0.
⎩ X
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 71 BG, CH, ÑHBK
2. Review of Probability Theory: Function of R. V (6)
Example 4: Half-wave rectifier Y
⎧ x, x > 0,
Y = g ( X ); g ( x ) = ⎨
⎩0, x ≤ 0.
X
In this case
P(Y = 0) = P( X (ξ ) ≤ 0) = FX (0).
Thus
⎧ f X ( y ), y > 0,
⎪
fY ( y ) = ⎨ FX (0)δ ( y ) y = 0, = f X ( y )U ( y ) + FX (0)δ ( y ).
⎪ 0, y < 0,
⎩
and determine appropriate events in terms of the r.v X for every y. Finally,
we must have FY(y) for -∞ < y < + ∞ and obtain
dFY ( y )
fY ( y ) = in a < y < b.
dy
The summation index i in this equation depends on y, and for every y the
equation y = g(xi) must be solved to obtain the total number of solutions at
every y, and the actual solutions x1, x2, … all in terms of y.
For example, if Y = X2, then for all y > 0, x1 = -√y and x2 = +√y represent
the two solutions for each y. Notice that the solutions xi are all in terms of
y so that the right side of (♦) is only a function of y. Moreover
Y = X2
dy dy y
= 2 x so that =2 y
dx dx x = xi
X
x1 x2
But this represents the p.d.f of a Cauchy r.v with parameter (1/α). Thus if
X ∼ C(α), then 1/X ∼ C(1/α).
Since X has zero probability of falling outside the interval (0, π), y = sinx has
zero probability of falling outside the interval (0, 1). Clearly fY(y) = 0 outside
this interval.
dx x = xi
x3
x
(a) π
y = sin x
x
x −1 x1 x2 x3
π
(b)
1 1 ⎛ 2 x1 2 x2 ⎞
fY ( y ) = ( f X ( x1 ) + f X ( x2 ) ) = ⎜ 2 + 2 ⎟
1− y 2
1− y ⎝ π
2 π ⎠
⎧ 2 fY ( y )
2( x1 + π − x1 ) ⎪ , 0 < y < 1,
= = ⎨π 1 − y 2
π 1− y
2 2
⎪⎩ 0, otherwise.
2
π
y
1
X takes the values 0, 1, 2,…, k, … so that Y only takes the value 1, 2, 5,…,
k2+1,… and P(Y = k2+1) = P(X = k) so that for j = k2+1
λ j −1
P(Y = j ) = P X = ( )
j −1 = e −λ
( j − 1)!
, j = 1, 2, 5, , k 2 + 1, .
= ∑ xi pi = ∑ xi P( X = xi ) .
i i
Mean represents the average (mean) value of the r.v in a very large number
of trials. For example if X ∼ U(a,b), then
b
b x 1 x2 b2 − a 2 a + b
E( X ) = ∫ dx = = =
a b−a b − a 2 a 2( b − a ) 2
implying that the parameter λ represents the mean value of the exponential
r.v.
i =0
−
Thus the parameter λ also represents the mean of the Poisson r.v.
1 +∞ 1 +∞
∫ ∫
2
/ 2σ 2 2
/ 2σ 2
= ye − y dy + μ ⋅ e− y dy = μ .
2πσ 2 −∞
2πσ 2 −∞
0
1
From the equations, fY(y) is not required to evaluate E(Y) for Y=g(X).
k =0 k =0 k! k =1 k!
∞
λk ∞
λi +1
=e −λ
∑ k ( k − 1)! = e ∑ ( i + 1)
k =1
−λ
i=0 i!
⎛ ∞ λi ∞
λi ⎞ ⎛ ∞ λi λ ⎞
= λe −λ
⎜⎜ ∑ i + ∑ ⎟ = λe −λ
⎜⎜ ∑ i + e ⎟⎟
⎝ i=0 i ! i=0 i! ⎟⎠ ⎝ i =1 i ! ⎠
⎛ ∞ λi λ ⎞ −λ ⎛
∞
λ m +1 ⎞
= λ e ⎜⎜ ∑
−λ
+ e ⎟⎟ = λ e ⎜⎜ ∑ + e λ ⎟⎟
⎝ i =1 ( i − 1)! ⎠ ⎝ m =0 m! ⎠
= λ e − λ (λ e λ + e λ ) = λ 2 + λ .
In general, E(Xk) is known as the kth moment of r.v X. Thus if X ∼ P(λ), its
second moment is given by the above equation.
where σX2 is known as the variance of the r.v X, and its square root
σ X = E ( X − μ ) 2 is known as the standard deviation of X. Note that the
standard deviation represents the root mean square spread of the r.v X
around its mean μ .
Var ( X ) = σ X2 = ∫
+∞
−∞
(x 2
− 2 xμ + μ 2 ) f X ( x )dx
+∞ +∞
=∫ x f X ( x )dx − 2 μ ∫ x f X ( x )dx + μ 2
2
−∞ −∞
= E (X )− μ = E (X ) − [E ( X )] 2
___
2
2 2 2
=X −X .
2
σ = X − X = (λ2 + λ ) − λ2 = λ.
___
2
2 2
X
Thus for a Poisson r.v, mean and variance are both equal to its parameter λ
We have +∞ 1
Var ( X ) = E [( X − μ ) 2 ] = ∫ ( x − μ)
2
/ 2σ 2
e −( x − μ )
2
dx.
−∞
2πσ 2
∫
−( x − μ ) 2 / 2σ 2
e dx = 2π
3 −∞ σ
or
+∞ 1
∫ −∞ ( )
2 2
μ e −( x− μ ) / 2σ
dx = σ 2 ,
2
x −
2 πσ 2
k =0 ⎝ k ⎠ k =0 ⎝ k ⎠
so that
E ( X 2 ) = λ2 + λ ,
and
x1
P ( y1 < Y (ξ ) ≤ y2 ) = FY ( y2 ) − FY ( y1 ) = ∫ fY ( y )dy.
y2
y1
What about the probability that the pair of r.vs (X,Y) belongs to an arbitrary
region D? In other words, how does one estimate, for example,
P[( x1 < X (ξ ) ≤ x2 ) ∩ ( y1 < Y (ξ ) ≤ y2 )] = ?
y2
R0
y1
x1 x2
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 93 BG, CH, ÑHBK
2. Review of Probability Theory: Two r.vs (3)
Joint probability density function (Joint p.d.f): By definition, the
joint p.d.f of X and Y is given by
∂ 2 FXY ( x, y )
f XY ( x, y ) = .
∂x ∂y
and hence we obtain the useful formula
x y
FXY ( x, y ) = ∫ ∫ f XY (u, v ) dudv.
−∞ −∞
P (( X , Y ) ∈ D ) = ∫ ∫ f XY ( x, y )dxdy.
( x , y )∈D
If X and Y are discrete r.vs, then pij = P(X = xi, Y = yi) represents their
joint p.d.f, and their respective marginal p.d.fs are given by
P( X = xi ) = ∑ P( X = xi , Y = y j ) = ∑ pij
j j
P(Y = y j ) = ∑ P( X = xi , Y = y j ) = ∑ pij
i i
The joint P.D.F and/or the joint p.d.f represent complete information about the r.vs,
and their marginal p.d.fs can be evaluated from the joint p.d.f. However, given
marginals, (most often) it will not be possible to compute the joint p.d.f.
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 95 BG, CH, ÑHBK
2. Review of Probability Theory: Two r.vs (5)
Independence of r.vs
Definition: The random variables X and Y are said to be statistically
independent if the events {X(ζ) ∈ A} and {Y(ζ) ∈ B} are independent
events for any two sets A and B in x and y axes respectively. Applying the
above definition to the events {X(ζ) ≤ x} and {Y(ζ) ≤ y}, we conclude that,
if the r.vs X and Y are independent, then
P (( X (ξ ) ≤ x ) ∩ (Y (ξ ) ≤ y ) ) = P ( X (ξ ) ≤ x ) P (Y (ξ ) ≤ y )
i.e.,
FXY ( x, y ) = FX ( x ) FY ( y )
or equivalently, if X and Y are independent, then we must have
f XY ( x, y ) = f X ( x ) fY ( y ).
= x ⎛⎜ − 2 ye − y + 2 ∫ ye − y dy ⎞⎟ = 2 x, 0 < x < 1.
∞ ∞
⎝ 0 0 ⎠
Similarly,
1 y2 −y
fY ( y ) = ∫ f XY ( x, y )dx = e , 0 < y < ∞.
0 2
In this case, f XY ( x, y ) = f X ( x ) fY ( y ), and hence X and Y are independent r.vs.
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 97 BG, CH, ÑHBK
2. Review of Probability Theory: Func. of 2 r.vs (1)
Given two random variables X and Y and a function g(x,y), we form a new
random variable Z = g(X, Y).
Given the joint p.d.f fXY(x, y), how does one obtain fZ(z) the p.d.f of Z ?
Problems of this type are of interest from a practical standpoint. For example,
a receiver output signal usually consists of the desired signal buried in noise,
and the above formulation in that case reduces to Z = X + Y.
X 2 +Y 2 X /Y
tan −1 ( X / Y )
Then
dH ( z ) db( z ) da ( z ) b ( z ) ∂h ( x , z )
= h (b( z ), z ) − h (a ( z ), z ) + ∫ dx.
dz dz dz a ( z ) ∂z
We have
FZ ( z ) = P (X 2 + Y 2 ≤ z ) = ∫ ∫ f XY ( x, y )dxdy.
X 2 +Y 2 ≤ z
∫
z
y =− z
1
2 z−y 2
(f XY ( z − y 2 , y ) + f XY (− z − y 2 , y ) dy. )
− z
(*)
− z / 2σ 2
e π /2 z cos θ 1 − z / 2σ 2
=
πσ 2 ∫ 0
z cos θ
dθ =
2σ 2
e U ( z ),
with y = z sin θ .
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 102 BG, CH, ÑHBK
2. Review of Probability Theory: Func. of 2 r.vs (6)
Thus, if X and Y are independent zero mean Gaussian r.vs with common
variance σ2 then X2 + Y2 is an exponential r.vs with parameter 2σ2.
f Z ( z) = ∫
z
−z
z −y
2
z
2
(f XY ( z 2
− y 2
, y ) + f XY ( − z 2
− y 2
, y ) dy. ) − z
2z π/2 z cosθ z
∫ Rayleigh distribution!
2
/ 2σ 2 2 2
= e −z
dθ = 2 e − z / 2σ U ( z ),
πσ 2 0 z cosθ σ
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 103 BG, CH, ÑHBK
2. Review of Probability Theory: Func. of 2 r.vs (7)
Thus, if W = X + iY, where X and Y are real, independent normal r.vs with
zero mean and equal variance, then the r.v W = X 2 + Y 2 has a Rayleigh
density. W is said to be a complex Gaussian r.v with zero mean, whose real
and imaginary parts are independent r.vs and its magnitude has Rayleigh
distribution.
⎛X ⎞
What about its phase θ = tan −1 ⎜ ⎟?
⎝Y ⎠
Clearly, the principal value of θ lies in the interval (-π/2, +π/2). If we let
U = tan θ = X/Y , then it is shown that U has a Cauchy distribution with
1/ π
fU ( u ) = , − ∞ < u < ∞.
u2 + 1
As a result
1 1 1/π ⎧1 / π , − π / 2 < θ < π / 2,
fθ (θ ) = fU (tan θ ) = =⎨
| dθ / du | (1 / sec 2 θ ) tan 2 θ + 1 ⎩ 0, otherwise .
−( z 2 + μ 2 ) / 2σ 2
ze ⎛⎜ π/2 e zμ cos(θ −φ ) / σ 2 dθ + 3π/2 e zμ cos(θ −φ ) / σ 2 dθ ⎞⎟
⎝ ∫ −π/2 ∫π/2
=
2πσ 2 ⎠
−( z 2 + μ 2 ) / 2σ 2
ze ⎛ zμ ⎞
= I 0 ⎜ 2 ⎟,
2πσ 2 ⎝σ ⎠
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 105 BG, CH, ÑHBK
2. Review of Probability Theory: Func. of 2 r.vs (9)
where x = z cos θ , y = z sin θ , μ = μ X2 + μY2 , μ X = μ cos φ , μY = μ sin φ ,
and 1 2π 1 π
I 0 (η ) =
2π ∫π ∫0
0
eη cos(θ −φ ) dθ = eη cosθ dθ
is the modified Bessel function of the first kind and zeroth order.
In the case of one random variable, we defined the parameters mean and
variance to represent its average behavior. How does one parametrically
represent similar cross-behavior between two random variables? Towards
this, we can generalize the variance definition given as
⎣ ⎦
= a 2 E ( X − μ X ) 2 + 2abE ( ( X − μ X )(Y − μY ) ) + b2 E (Y − μY ) 2
= a 2σ X2 + 2ab ρ XY σ X σ Y + b2σ Y2 .
Following the one random variable case, we can define the joint
characteristic function between two random variables which will turn out to
be useful for moment calculations.
where
μ Z = aμ X + bμY ,
μW = cμ X + dμY ,
σ Z2 = a 2σ X2 + 2abρσ X σ Y + b2σ Y2 ,
σ W2 = c 2σ X2 + 2cdρσ X σ Y + d 2σ Y2 ,
and
acσ X2 + ( ad + bc ) ρσ X σ Y + bdσ Y2
ρ ZW = .
σ Zσ W
Thus, Z and W are also jointly distributed Gaussian r.vs with means,
variances and correlation coefficient as above.
Central Limit Theorem: Suppose X1, X2,…, Xn are a set of zero mean
independent, identically distributed (i.i.d) random variables with some
common distribution. Consider their scaled sum
X1 + X 2 + + Xn
Y= .
n
Then asymptotically (as n →∞)
Y → N (0, σ 2 ).
Mean, Autocorrelation.
Stationarity.
Deterministic Systems.
Stochastic Models.
(iii) RXX(t1, t2) represents a nonnegative definite function, i.e., for any set
of constants {ai}ni=1 n n
∑ ∑ ai a*j R XX
(ti , t j ) ≥ 0.
i =1 j =1
n
this follows by noticing that E{| Y | } ≥ 0 for Y = ∑ ai X (ti ).
2
i =1
The function
C XX (t1 , t 2 ) = RXX (t1 , t 2 ) − μ X (t1 ) μ *X (t 2 )
represents the autocovariance function of the process X(t).
In strict terms, the statistical properties are governed by the joint probability
density function. Hence a process is nth-order Strict-Sense Stationary
(S.S.S) if
f X ( x1 , x2 , xn , t1 , t2 , tn ) ≡ f X ( x1 , x2 , xn , t1 + c, t2 + c , tn + c ) (*)
for any c, where the left side represents the joint density function of the
random variables X1 = X(t1), X2 = X(t2), …, Xn = X(tn), and the right side
corresponds to the joint density function of the random variables X’1=X(t1+c),
X’2 = X(t2+c), …, X’n = X(tn+c). A process X(t) is said to be strict-sense
stationary if (*) is true for all ti, i = 1, 2, …, n; n = 1, 2, … and any c.
However, the basic conditions for the first and second order stationarity are
usually difficult to verify. In that case, we often resort to a looser definition
of stationarity, known as Wide-Sense Stationarity (W.S.S). Thus, a process
X(t) is said to be Wide-Sense Stationary if
(i) E{ X (t )} = μ
and
(ii) E{ X (t1 ) X (t2 )} = R XX (t1 − t2 ),
*
⎯⎯⎯→
X (t )
T [⋅] ⎯Y⎯→
⎯
(t )
t t
Our goal is to study the output process statistics in terms of the input
process statistics and the system function.
Deterministic Systems
Linear-Time Invariant
(LTI) systems
+∞
X (t ) h (t ) Y ( t ) = ∫ − ∞ h ( t − τ ) X (τ ) dτ
+∞
= ∫ − ∞ h (τ ) X ( t − τ ) dτ .
LTI system
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 128 BG, CH, ÑHBK
2. Review of Stochastic Processes: Systems (3)
Memoryless Systems:
The output Y(t) in this case depends only on the present value of the input X(t).
i.e., Y ( t ) = g{ X ( t )}
Impulse Impulse
response
then Y (t )
X (t )
X (t ) Y (t ) t
t LTI
+∞
Y (t ) = ∫ − ∞ h (t − τ ) X (τ )dτ
arbitrary
input +∞
= ∫ − ∞ h (τ ) X (t − τ )dτ
where +∞
X (t ) = ∫ − ∞ X (τ )δ (t − τ )dτ
+∞
= ∫ − ∞ μ X (τ )h(t − τ )dτ = μ X (t ) ∗ h(t ).
Δ
= R XX (τ ) ∗ h * ( −τ ) = R XY (τ ), τ = t1 − t 2 .
Boä moân Vieãn thoâng SSP2008
Khoa Ñieän-Ñieän töû 134 BG, CH, ÑHBK
2. Review of Stochastic Processes: Systems (9)
Thus X(t) and Y(t) are jointly w.s.s. Further, the output autocorrelation
simplifies to
+∞
RYY (t1 , t 2 ) = ∫ −∞ RXY (t1 − β − t 2 )h( β )dβ , τ = t1 − t 2
= RXY (τ ) ∗ h(τ ) = RYY (τ ).
or
RYY (τ ) = RXX (τ ) ∗ h* ( −τ ) ∗ h(τ ).
X (t ) Y (t )
wide-sense LTI system wide-sense
stationary process h(t) stationary process.
(a)
X (t ) Y (t )
LTI system
strict-sense strict-sense
h(t)
stationary process stationary process
(b)
X (t ) Y (t )
Gaussian process Linear system Gaussian process
(also stationary) (also stationary)
(c)
Then, E[xBnxBHn] = RT
⎡ R (0) r H ⎤
R M +1 = ⎢ ⎥
⎣ r R M⎦
or
⎡R M r B∗ ⎤
R M +1 = ⎢ BT ⎥
⎣r R ( 0) ⎦
ρ X (0) = ρ Y (0) = 1
ρ X (k ) > ρY (k )
n
0 k
k =0
RXX (0)
(ARMA) is that
det H n = 0, n≥N (for all sufficiently large n ),
where
⎛ h0 h1 h2 hn ⎞
⎜h h h3 hn +1 ⎟
Δ
Hn = ⎜ 1 2
⎟.
⎜ ⎟
⎜h h h2 n ⎟⎠
⎝ n n +1 hn + 2
i.e., in the case of rational systems for all sufficiently large n, the Hankel
matrices Hn all have the same rank.