You are on page 1of 251

Solutions Manual

to accompany

Probability, Random Variables and Stochastic Processes


Fourth Edition Athanasios Papoulis
Polytechnic University

S. Unnikrishna Pillai
Polytechnic University

Solutions Manual to accompany PROBABILITY, RANDOM VARIABLES AND STOCHASTIC PROCESSES, FOURTH EDITION ATHANASIOS PAPOULIS Published by McGraw-Hill Higher Education, an imprint of The McGraw-Hill Companies, Inc., 1221 Avenue of the Americas, New York, NY 10020. Copyright 2002 by The McGraw-Hill Companies, Inc. All rights reserved. The contents, or parts thereof, may be reproduced in print form solely for classroom use with PROBABILITY, RANDOM VARIABLES AND STOCHASTIC PROCESSES, FOURTH EDITION, provided such reproductions bear copyright notice, but may not be reproduced in any other form or for any other purpose without the prior written consent of The McGraw-Hill Companies, Inc., including, but not limited to, in any network or other electronic storage or transmission, or broadcast for distance learning. www.mhhe.com

Problem Solutions for Chapter 3


3.1 (a) P (A occurs atleast twice in n trials) = 1 P (A never occurs in n trials) P (A occurs once in n trials) = 1 (1 p)n np(1 p)n;1
; ; ; ; ; ;

(b) P (A occurs atleast thrice in n trials) = 1 P (A never occurs in n trials) P (A occurs once in n trials) P (A occurs twice in n trials) = 1 (1 p)n np(1 p)n;1 n(n2 1) p2(1 p)n;2
; ; ; ; ; ; ; ; ; ;

3.2

1= 1 P (doublesix) = 1 6 6 36
j

= 1 50 0 = 0:162
;

P (\double six atleast three times in n trials00)


1 36
0

35 36

50

50 1

1 36

35 36

49

50 2

1 36

35 36

48

3.6 (a) (b) 1 (c) 1


; ;

p1 = 1
5 6 18 1
!

5 6 = 0:665 6
!

12
;

12 1 5 6

1 6
;

5 6 18 2
!

11

= 0:619
2

5 6

18
;

1 6

17

1 6

5 6

16

= 0:597

2 3.7 (a) Let n represent the number of wins required in 50 games so that the net gain or loss does not exceed $1. This gives the net gain to be 1 < n 50 4 n < 1 16 < n < 17:3 n = 17
; ; ;

1 17 3 33 = 0:432 P (net gain does not exceed $1) = 50 17 4 4 P (net gain or loss exceeds $1) = 1 0:432 = 0:568
;

(b) Let n represent the number of wins required so that the net gain or loss does not exceed $5. This gives 5 < n (50 2 n) < 5 13:3 < n < 20
; ; ;

X 50 1 P (net gain does not exceed $5) = 19 n = 14 n 4 P (net gain or loss exceeds $5) = 1 0:349 = 0:651
;

3 4

50;n

= 0:349

3 3.8 De ne the events A=\ r successes in n Bernoulli trials" B =\success at the ith Bernoulli trial" C =\r 1 successes in the remaining n 1 Bernoulli trials excluding the ith trial" P (A) = n r pr qn;r P (B ) = p 1 P (C ) = n r 1 pr;1 qn;r We need ) = P (BC ) = P (B ) P (C ) = r : P (B A) = PP(AB (A) P (A) P (A) n
; ; ; ; j

3.9 There are 52 13 ways of selecting 13 cards out of 52 cards. The number of ways to select 13 cards of any suit (out of 13 cards) equals 13 = 1. Four such (mutually exclusive) suits give the total number 13 of favorable outcomes to be 4. Thus the desired probability is given by 4 ! = 6:3 10;12 52 13
j

4 3.10 Using the hint, we obtain

p (Nk+1 Nk ) = q (Nk Nk;1) 1


; ; ;

Let

Mk+1 = Nk+1 Nk
;

so that the above iteration gives

Mk+1 = q p Mk 1 p 8 n o k q q 1 > k > M 1 ( ) p=q > p 1 p q p < => > M1 k p=q > : p
; ; ; ; 6 ;

This gives

Ni =

i;1 X k=0

Mk+1 M1 + p 1 q
;

8 > > > > > < => > > > > :

i;1 X k=0

q p
;

i p q p=q
; 6

iM1 i(i2p 1)
;

p=q

where we have used No = 0. Similarly Na+b = 0 gives

a + b 1 q=p : M1 + p 1 q = p q 1 (q=p)a+b
; ; ; ;

Thus

8 > > > < Ni = > > > :

i a + b 1 (q=pa)+ i p q 1 (q=p) b p q p = q i(a + b i) p=q


; ; ; ; ; 6 ;

5 which gives for i = a


8 > > > < Na = > > > : 8 > > > < => > > :

)a a a + b 1 (q=pa p q 1 (q=p) +b p q p = q ab p=q )b p = q: a + b 1 (p=qa b 2p 1 2p 1 1 (p=q) +b ab p=q


; ; ; ; ; 6 ; ; ; ; ; 6

6 3.11

Arguing as in (3.43), we get the corresponding iteration equation

Pn = pPn+ + qPn;a Pn = Pn+ + qPn;a

and proceed as in Example 3.15. 3.12 Suppose one best on k = 1 2 Then

6. 5 6 5 6
2

1 p1 = P (k appears on one dice) = 3 1 6 1 2 p2 = P (k appear on two dice) = 3 2 6 5 p0 = P (k appear none) = 6 Thus, we get
3

1 p3 = P (k appear on all the tree dice) = 6

Net gain = 2p1 + 3p2 + 4p3 p0 = 0:343:


;

215 Chapter 15 15.1 The chain represented by

1/2 1/2 0

1/2 P =

1/2 1/2

1/2

is irreducible and aperiodic. The second chain is also irreducible and aperiodic. The third chain has two aperiodic closed sets {e1 , e2 } and {e3 , e4 } and a transient state e5 . 15.2 Note that both the row sums and column sums are unity in this case. Hence P represents a doubly stochastic matrix here, and

Pn

1 = m+1

1 1 1 1

1 1 1 1

1 1 1 1 . . . . . . . . . . . . . . .

1 , k = 0, 1, 2, m. m+1 15.3 This is the success runs problem discussed in Example 15-11 and 15-23. From Example 15-23, we get
n

lim P {xn = ek } =

ui+1 = pi,i+1 ui = so that from (15-206)


1 uo ui = i+1 (i + 1)!

uk = u 0
k=1 k=1

1 = e u0 = 1 k!

216 gives u0 = 1/e and the steady state probabilities are given by uk = 1/e , k! k = 1, 2,

15.4 If the zeroth generation has size m, then the overall process may be considered as the sum of m independent and identically distributed k) branching processes x( n , k = 1, 2, m, each corresponding to unity size at the zeroth generation. Hence if 0 represents the probability of extinction for any one of these individual processes, then the overall probability of extinction is given by
n

lim P [xn = 0|x0 = m] =


(1) (2) m) = 0|x0 {x ( n (m)

(2) = P [{x(1) n = 0|x0 = 1} {xn = 0|x0 = 1}

= 1}]

m k=1

k) P [x ( n = 0|x0 = 1]

(k)

m = 0

15.5 From (15-288)-(15-289), P (z ) = p 0 + p 1 z + p 2 z 2 , since pk = 0, k 3.

Also p0 + p1 + p2 = 1, and from (15-307) the extinction probability is given by sloving the equation P (z ) = z. Notice that P (z ) z = p0 (1 p1 )z + p2 z 2 = (z 1)(p2 z p0 ) and hence the two roots of the equation P (z ) = z are given by z1 = 1, z2 = p0 . p2

= p 0 (p 0 + p 2 )z + p 2 z 2

Thus if p2 < p0 , then z2 > 1 and hence the smallest positive root of P (z ) = z is 1, and it represents the probability of extinction. It follows

217 that such a tribe which does not produce ospring in abundence is bound to extinct. 15.6 Dene the branching process {xn }
xn

xn+1 =
k=1

yk

where yk are i.i.d random variables with common moment generating function P (z ) so that (see (15-287)-(15-289)) P (1) = E {yk } = . Thus E {xn+1 |xn } = E { = E{ = E{ Similarly E {xn+2 |xn } = E {E {xn+2 |xn+1 , xn }} = E {E {xn+2 |xn+1 }|xn } and in general we obtain E {xn+r |xn } = r xn . Also from (15-310)-(15-311) E {x n } = n . Dene wn = This gives Dividing both sider of (i) with n+r we get E{ xn xn xn+r r | x = x } = = = wn n n+r n+r n E {wn } = 1. xn . n (ii) (iii) (i) = E {xn+1 |xn } = 2 xn
xn k=1 m k=1 m k=1

y k |x n = m } y k |x n = m } yk } = mE {yk } = xn

218 or E {wn+r |wn = which gives E {wn+r |wn } = wn , the desired result. 15.7 sn = x 1 + x 2 + + x n where xn are i.i.d. random variables. We have sn+1 = sn + xn+1 so that E {sn+1 |sn } = E {sn + xn+1 |sn } = sn + E {xn+1 } = sn . Hence {sn } represents a Martingale. 15.8 (a) From Bayes theorem P {xn = j |xn+1 = i} = P {xn+1 = i|xn = j } P {xn = j } P {xn+1 = i} qj pji = qi = p ij , x = w } = wn n

(i)

where we have assumed the chain to be in steady state. (b) Notice that time-reversibility is equivalent to p ij = pij and using (i) this gives p ij = qj pji = pij qi (ii)

or, for a time-reversible chain we get qj pji = qi pij . (iii)

219 Thus using (ii) we obtain by direct substitution q pij pjk pki = qj pji i = pik pkj pji , the desired result. 15.9 (a) It is given that A = AT , (aij = aji ) and aij > 0. Dene the ith row sum aik > 0, i = 1, 2, ri =
k

qk p qj kj

qi qk pik

and let pij = Then pji =

aij aij = . ri k aik a a aji = rji = rij j j ajm

ri aij = ri p =r rj ij j ri or ri pij = rj pji . Hence ri pij =


i i

(i)

rj pji = rj
i

pji = rj ,

(ii)

since pji =
i i

aji

rj

rj = 1. rj

Notice that (ii) satises the steady state probability distribution equation (15-167) with qi = c r i , i = 1, 2, where c is given by c
i

ri =
i

qi = 1 = c =

1
i ri

=
i

1
j

aij

220 Thus qi = ri = i ri aij >0 j aij

j i

(iii)

represents the stationary probability distribution of the chain. With (iii) in (i) we get qi pji = pij qj or qj pji pij = = p ij qi and hence the chain is time-reversible. 15.10 (a) M = (mij ) is given by M = (I W )1 or (I W )M = I M = I + WM wik mkj , pik mkj , which gives mij = ij + = ij +
k k

ei , ej T ei , ej T

15.11 If a stochastic matrix A = (aij ), aij > 0 corresponds to the twostep transition matrix of a Markov chain, then there must exist another stochastic matrix P such that A = P 2, P = (pij )

(b) The general case is solved in pages 743-744. From page 744, with N = 6 (2 absorbing states; 5 transcient states), and with r = p/q we obtain (rj 1)(r 6i 1) , ji (p q )(r 6 1) mij = (ri 1)(r 6i rj i ) , j i. (p q )(r 6 1)

221 where pij > 0,


j

pij = 1,

and this may not be always possible. For example in a two state chain, let 1 P = 1 so that

A = P2 =

2 + (1 )(1 ) ( + )(1 )

2 + (1 )(1 )

( + )(1 )

This gives the sum of this its diagonal entries to be a11 + a22 = 2 + 2(1 )(1 ) + 2 = ( + )2 2( + ) + 2
2

(i)

= 1 + ( + 1)

1.

Hence condition (i) necessary. Since 0 < < 1, 0 < < 1, we also get 1 < a11 + a22 2. Futher, the condition (i) is also sucient in the 2 2 case, since a11 + a22 > 1, gives ( + 1)2 = a11 + a22 1 > 0 and hence + =1 a11 + a22 1

and this equation may be solved for all admissible set of values 0 < < 1 and 0 < < 1. 15.12 In this case the chain is irreducible and aperiodic and there are no absorption states. The steady state distribution {uk } satises (15167),and hence we get
N

uk =
j

uj pjk =
j =0

uj

N k N k p j qj . k

222 Then if > 0 and > 0 then xation to pure genes does not occur. 15.13 The transition probabilities in all these cases are given by (page 765) (15A-7) for specic values of A(z ) = B (z ) as shown in Examples 15A-1, 15A-2 and 15A-3. The eigenvalues in general satisfy the equation (k) (k) pij xj = k xi , k = 0, 1, 2, N
j

and trivially j pij = 1 for all i implies 0 = 1 is an eigenvalue in all cases. However to determine the remaining eigenvalues we can exploit the relation in (15A-7). From there the corresponding conditional moment generating function in (15-291) is given by
N

G(s) =
j =0

pij sj

(i)

where from (15A-7) pij = = {Ai (z )}j {B N i (z )}N j {Ai (z ) B N i (z )}N

coecient of sj z N in {Ai (sz ) B N i (z )} {Ai (z ) B N i (z )}N

(ii)

Substituting (ii) in (i) we get the compact expression G(s) = {Ai (sz ) B N i (z )}N . {Ai (z ) B N i (z )}N (iii)

Dierentiating G(s) with respect to s we obtain


N

G (s) =
j =0

Pij j sj 1 {iAi1 (sz ) A (sz )z B N i (z )}N {Ai (z ) B N i (z )}N (iv)

=i

{Ai1 (sz ) A (sz ) B N i (z )}N 1 . {Ai (z ) B N i (z )}N

223 Letting s = 1 in the above expression we get {Ai1 (z ) A (z ) B N i (z )}N 1 G (1) = . pij j = i {Ai (z ) B N i (z )}N j =0 In the special case when A(z ) = B (z ), Eq.(v ) reduces to
N N

(v)

pij j = 1 i
j =0

(vi)

where

{AN 1 (z ) A (z )}N 1 . { A N (z )} N Notice that (vi) can be written as 1 = P x 1 = 1 x1 , x1 = [0, 1, 2, N ]T

(vii)

and by direct computation with A(z ) = B (z ) = (q + pz )2 (Example 15A-1) we obtain 1 = {(q + pz )2(N 1) 2p(q + pz )}N {(q + pz )2N }N 2p{(q + pz ) }N 1 = {(q + pz )2N }N
2N 1

2p

2N q N pN 1 N 1 = 1. 2N N N q p N

Thus N j =0 pij j = i and from (15-224) these chains represent Martingales. (Similarly for Examples 15A-2 and 15A-3 as well). To determine the remaining eigenvalues we dierentiate G (s) once more. This gives
N

G (s) =
j =0

pij j (j 1) sj 2

{i Ai2 (sz ) B N i (z )[(i 1) (A (sz )) + A(sz ) A (sz )]}N 2 . = {Ai (z ) B N i (z )}N

{i(i 1)Ai2 (sz )[A (sz )]2 z B N i (z ) + iAi1 (sz ) A (sz )z B N i (z )}N 1 . {Ai (z ) B N i (z )}N
2

224 With s = 1, and A(z ) = B (z ), the above expression simplies to


N j =0

pij j (j 1) = 2 i(i 1) + i2

(viii)

where 2 = and 2 = {AN 1 (z ) A (z )}N 2 . { A N (z )} N {AN 2 (z ) [A (z )]2 }N 2 { A N (z )} N

Eq. (viii) can be rewritten as


N j =0

pij j 2 = 2 i2 + (polynomial in i of degree 1)

and in general repeating this procedure it follows that (show this)


N j =0

pij j k = k ik + (polynomial in i of degree k 1)

(ix)

where k = {AN k (z ) [A (z )]k }N k , { A N (z )} N k = 1, 2, N. (x)

Equations (viii)(x) motivate to consider the identities P q k = k qk (xi)

where qk are polynomials in i of degree k , and by proper choice of constants they can be chosen in that form. It follows that k , k = 1, 2, N given by (ix) represent the desired eigenvalues. (a) The transition probabilities in this case follow from Example 15A-1 (page 765-766) with A(z ) = B (z ) = (q + pz )2 . Thus using (ix) we

225 obtain the desired eigenvalues to be k = {(q + pz )2(N k) [2p(q + pz )]k }N k {(q + pz )2N }N {(q + pz )2N k }N k {(q + pz )2N }N }

= 2 k pk

= 2k

2N k N k , 2N N

k = 1, 2, N.

(b) The transition probabilities in this case follows from Example 15A-2 (page 766) with A(z ) = B (z ) = e(z1) and hence {e(N k)(z1) k ek(z1) }N k k = {eN (z1) }N = =

k {eN z }N k k (N )N k /(N k )! = N z (N )N /N ! {e } N 1 N! = 1 N (N k )! N k

2 1 k 1 , 1 N N

k = 1, 2, N

(c) The transition probabilities in this case follow from Example 15A-3 (page 766-767) with q A (z ) = B (z ) = . 1 pz Thus {1/(1 pz )N +k }N k k = p k {1/(1 pz )N }N = (1)k (N + k ) N k N N = 2N 1 N k , 2N 1 N r = 2, 3, N

226 15.14 From (15-240), the mean time to absorption vector is given by m = (I W )1 E, where Wik = pjk , j, k = 1, 2, N 1, E = [1, 1, 1]T ,

with pjk as given in (15-30) and (15-31) respectively. 15.15 The mean time to absorption satises (15-240). From there mi = 1 +
kT

pik mk = 1 + pi,i+1 mi+1 + pi,i1 mi1

= 1 + p mi+1 + q mi1 , or mk = 1 + p mk+1 + q mk1 . This gives p (mk+1 mk ) = q (mk mk1 ) 1 Let Mk+1 = mk+1 mk so that the above iteration gives qM 1 Mk+1 = p k p q = p =

k

1 1 + q + q 2 + + q k1 M1 p p p p q q kM 1 k 1 p p q 1 (p) , p = q M1 k p=q p,

227 This gives


i1

mi =
k=0

Mk+1

= =

1 M1 + p q

i1 k=0

q p

i , p=q p q p=q

iM1 1 M1 + p q

i(i 1) 2p ,

1 (q/p)i i , p=q p q 1 q/p i(i 1) 2p , p=q

iM1

where we have used mo = 0. Similarly ma+b = 0 gives M1 + Thus 1 a+b 1 q/p = . pq p q 1 (q/p)a+b

which gives for i = a ma

i a + b 1 (q/p) i , p = q pq 1 (q/p)a+b p q mi = i (a + b i ), p=q a a + b 1 (q/p) a , p = q p q 1 (q/p)a+b pq = ab, p=q b b a + b 1 (p/q ) , p = q a+b

2p 1

2p 1 1 (p/q ) ab,

p=q

by writing

(see also problem 3-10).

(q/p)a (q/p)a+b 1 (p/q )b 1 (q/p)a =1 =1 1 (q/p)a+b 1 (q/p)a+b 1 (p/q )a+b

228 Chapter 16 16.1 Use (16-132) with r = 1. This gives


pn =

n p , n 1 n! 0 n p 0 , 1 < n m 0nm n = p 0
n=0

= n p0 , Thus
m m

pn = p 0
n=0

(1 m+1 ) =1 1

= and hence pn =

p0 =

1 1 m+1

1 n , 1 n+1

0 n m,

=1

and lim 1, we get pn = 1 , m+1 = 1.

16.2 (a) Let n1 (t) = X + Y , where X and Y represent the two queues. Then p n = P { n 1 (t ) = n } = P { X + Y = n }
n

=
k=0 n

P {X = k } P {Y = n k } (i) (1 )k (1 )nk
k=0

= (n + 1)(1 )2 n , where = /.

n = 0, 1, 2,

229 (b) When the two queues are merged, the new input rate = + = 2. Thus from (16-102)

pn = Hence

(2)n ( /)n p 0 = n! n! p 0 , n < 2 22 ( )n p = 2n p , 0 0 2! 2 n 2.

pk = p0 (1 + 2 + 2
k0 k=2

k )

2 ) = p0 (1 + 2 + 12 2 0 = 1p ((1 + 2) (1 ) + 2 ) 0 = 1p (1 + ) = 1

= p0 = Thus pn =

1 , 1+

( = /).

(ii)

2 (1 ) n /(1 + ), n 1 (1 )/(1 + ), n=0

(iii)

(c) For an M/M/1 queue the average number of items waiting is given by (use (16-106) with r = 1)

E {X } = L 1 =
n=2

(n 1) pn

230 where pn is an in (16-88). Thus

L1 =
n=2

(n 1)(1 ) n
2 n=2

= (1 ) = (1 )

(n 1) n2 (iv) k
k=1 k1

1 2 = (1 ) = . (1 )2 (1 )
2

Since n1 (t) = X + Y we have L 1 = E { n 1 (t )} = E { X } + E { Y } = 2L1 = 22 1 (v)

For L2 we can use (16-106)-(16-107) with r = 2. Using (iii), this gives L2 = p r (1 )2 =2 = (1 ) 2 2 3 = 1 + (1 )2 1 2 1+ < L1 (vi)

2 2 1

From (vi), a single queue conguration is more ecient then two separate queues. 16.3 The only non-zero probabilities of this process are 0,0 = 0 = m, i,i+1 = (m i) , 0,1 = i,i1 = i

231 i,i = [(m i) + i], i = 1, 2, , m 1

m,m = m,m1 = m. Substituting these into (16-63) text, we get m p0 = p1 (i)

[(m i) + i] pi = (m i + 1) pi1 + (i + 1) pi+1 , and m pm = pm1 . Solving (i)-(iii) we get m pi = i 16.4 (a) In this case

i = 1, 2, , m 1 (ii) (iii)

mi

i = 0, 1, 2, , m

pn =

= 1 1 1

p0 , n < m

p0 , n m 1 1 1 2 2

n 1 p0 ,

n<m

where

=
n=0

1 nm+1 m 2 p0 , n m, 1

m1

pn = p 0
k=0

k 1

1 m 2 1 n=0 1 2 m 1

n 2 =1

= p0

1 + 1 1 1 2

m 1

232 gives p0 = (b)


1 2 m 1 m 1 1 + 1 1 1 2 1

L =
n=0

n pn
m1

= p0
n=0

n n 1

m1

+
n=m

1 nm+1 n m 2 1

= p 0 1

1 n n + 1 1 n=0 m1

1 2

m2

n=m

d = p 0 1 d1 = p0

d = p 0 1 d1

n 1 + 1
n=0

1 2 1 2

m2

d d2

1 n n 2

n 2

n=m

1 m 1 1

m2

+ 1

d d

m1 1 1 [1 + (m 1)m ] 2 m + [m (m 1)2 ] 1 1 m1 + . 2 (1 1 ) (1 2 )2

m 1

16.5 In this case i =


j<r

p, j r

i =

j, j < r r, j r.

Using (16-73)-(16-74), this gives


pn =

(/)n n<r n! p 0 , (/)r nr r! (p/r) , n r.

233 16.6
m1

P {w > t } =
n=r m1

p n P (w > t | n ) pn (1 Fw (t|n)) =
n=r

pr

nr

(1 Fw (t|n))

fw (t|n) = et and

()nr+1 tnr (n r)!)


nr

(see 16.116)

F w (t | n ) = 1
k=0

(t)k t e k!
nr

(see 4.)

so that 1 F w (t | n ) =
k=0 m1

(t)k t e k! (t)k t k! e nr =i

P {w > t } =
n=r

pr
mr1

nr nr k=0

=
i=0

p r i
k=0 mr1

(t)k t , k! e
k

= pr et
k=0 mr1 k

k
i=0

(t)i i!

mr1 mr1

=
k=0 i=0

=
i=0 k=i

mr1

P {w > t} = pr et
i=0 t r = 1p e

(t)i i!

mr1

k
k=i

mr1 i=0

(t)i i mr ), i ! (

= /.

234 Note that m = M/M/r/m = M/M/r and


t r P (w > t ) = 1 p e i=0

(t)i i! t > 0.

(1)t r = 1p e

and it agrees with (16.119) 16.7 (a) Use the hints (b)

n=1

( + ) p n z n +

pn+1 z n+1 +
n=1 n=1 k=1

pnk ck z n = 0 ck z k
k=1 m=0

( + 1) (P (z ) p0 ) + (P (z ) p0 p1 z ) + z which gives

pm z m = 0

P (z )[1 z z (1 C (z ))] = p0 (1 z ) or p0 (1 z ) . 1 z z (1 C (z )) p 0 p0 1 = P (1) = = 1 + z C (z ) + C (z ) 1 + C (1) P (z ) = = p0 = 1 0 , Let D (z ) = Then P (z ) = (c) This gives P (z ) = (1 c ) (D(z ) + zD (z )) (1 zD(z ))2 0 = C (1).

1 C (z ) . 1z

1 L . 1 zD(z )

235 L = P (1) = (1 c ) (D(1) + D (1)) (1 c )2 1 = (C (1) + D (1)) (1 c ) C (1) = E (x) D (z ) = D (z ) = = By L-Hopitals Rule D (1) = limz1 C (z ) (1)C (z ) (1 z )C (z ) 2(1 z ) C (z ) 2 E (X 2 ) E (X ) k (k 1) Ck = 2 (E (X ) + E (X 2 )) . 2 (1 E (X )) 1 C (z ) 1z

(1 z ) (C (z )) (1 C (z )) (1) (1 z )2 1 C (z ) (1 z )C (z ) (1 z )2

= limz1 = 1/2C (z ) = = 1/2

L= (d)

C (z )z m P (z ) = D (z ) =

E (X ) = m

1 k 1 m k=1 z

1 z m m1 k = z 1z k=0 E (X 2 ) = m 2 (m + m 2 ) 2(1 m)

E (X ) = m, L=

236 (e) C (z ) = P (z ) = qz 1 Pz C (z ) = qz 1 pz

1 0 , 1 zD(z )

1 1qz 1 P z (1 P )z 1z 1 1 C (z ) P (z ) = = = = D (z ) = 1z 1z (1 z )(1 P z ) (1 z )(1 P z ) 1 Pz P (z ) = (1 0 )(1 pz ) (1 0 )(1 pz ) = 1 pz z 1 (p + )z (1 pz )q qz (p) q 1 = = (1 P z )2 q2 q D (z ) =

C (1) =

1 C (z ) 1z D(1) = C (1) 1 c L = P (1) = ( C (1) + D (1)) (1 c )2 D (z ) = (1 z )C (z ) (1 C (z )) ( 1) 1 C (z ) (1 z )C (1) = 2 (1 z ) (1 z )2 C (z ) (1)C (z ) (1 z )C (z ) 2(1 z )

limz1 D (z ) = limz1 =

(1 z )C (z ) (z ) = 2 2(1 z ) D (1) =

C (1) 2 2 1 (E (X ) E (X )) L= E (X ) + (1 c ) 2 16.8 (a) Use the hints. (b)

E (X ) + E (X 2 ) . 2(1 c )

n=1

( + ) p n z n +

zn

pn+m z n+m + z
n=1 n=1

pn1 z n1 = 0

237 or (1 + ) (P (z ) p0 ) + 1 zm
m

P (z )
k=0

pk z k + zP (z ) = 0

which gives
m

P (z ) z m+1 ( + 1) z m + 1 =
k=0

pk z k p0 (1 + ) z m

or P (z ) =

pk z k p0 (1 + ) z m
k=0 z m+1

( +

1) z m

+1

N (z ) . M (z )

(i)

(c) Consider the denominator polynomial M (z ) in (i) given by M (z ) = z m+1 (1 + ) z m + 1 = f (z ) + g (z ) where f (z ) = (1 + ) z m , g (z ) = 1 + z m+1 . Notice that |f (z )| > |g (z )| in a circle dened by |z | = 1 + , > 0. Hence by Rouches Theorem f (z ) and f (z )+ g (z ) have the same number of zeros inside the unit circle (|z | = 1 + ). But f (z ) has m zeros inside the unit circle. Hence f (z ) + g (z ) = M (z ) also has m zeros inside the unit circle. Hence M (z ) = M 1 (z ) (z z 0 ) (ii) where |z0 | > 1 and M1 (z ) is a polynomial of degree m whose zeros are all inside or on the unit circle. But the moment generating function P (z ) is analytic inside and on the unit circle. Hence all the m zeros of M (z ) that are inside or on the unit circle must cancel out with the zeros of the numerator polynomial of P (z ). Hence N (z ) = M1 (z ) a. (iii)

238 Using (ii) and (iii) in (i) we get P (z ) = a N (z ) = . M (z ) z z0

But P (1) = 1 gives a = 1 z0 or z0 1 P (z ) = z0 z = 1 = pn = 1 where r = 1/z0 . (d) Average system size

1 z0
n

(z/z0 )n
n=0

1 z0

1 z0

= (1 r) r n ,

n0

(iv)

L=
n=0

n pn =

r . 1r

16.9 (a) Use the hints in the previous problem. (b)


n=m

( + ) p n z n +
n=m m1

pn+m z n +
n=m

pn1 z n

(1 + ) P (z )
k=0

pk z k +
m2

2m1 1 P ( z ) pk z k zm k=0

+ z P (z )
k=0

pk z k = 0.

After some simplications we get


m1

P (z ) z m+1 ( + 1) z m + 1 = (1 z m )
k=0

pk z k
m1

or (1 z m ) P (z ) =

m1

pk z k
k=0

(z0 1) =
k=0

zk

z m+1 ( + 1) z m + 1

m (z 0 z )

239 where we have made use of Rouches theerem and P (z ) 1 as in problem 16-8. (c)
m1

zk 1 r k=0 pn z n = P (z ) = m 1 rz n=0

gives pn = where

(1 + r + + r k ) p0 ,

k m1

rnm+1 (1 + r + + r m1 ) p0 , k m 1r , m

p0 = Finally L=

r=

1 . z0

n pn = Pn (1).
n=0 m1

But 1r P (z ) = m so that L = P (1) =

m1

k z k1 (1 rz )
k=1 k=0

z k ( r )

(1 rz )2

1r m1+r m (1 r) = 2 m (1 r) m (1 r) 1 1 = . 1r m

16.10 Proceeding as in (16-212),

A (u ) = =

eu dA( )
m

m u + mz

240 This gives B (z ) = A ( (1 z )) =

m (1 z ) + m
m
m

1 = 1 1 + (1 z ) = (1 + ) z

(i) . m

Thus the equation B (z ) = z for 0 reduce to (1 + ) z or = z 1/m , (1 + ) z which is the same as z 1/m = (1 + ) z Let x = z 1/m . Sustituting this into (ii) we get x = (1 + ) xm or xm+1 (1 + ) xm + 1 = 0 (iii) (ii)
m

=z

16.11 From Example 16.7, Eq.(16-214), the characteristic equation for Q(z ) is given by ( = /m ) 1 z [1 + (1 z )]m = 0

241 which is equivalent to 1 + (1 z ) = z 1/m . Let x = z 1/m in this case, so that (i) reduces to [(1 + ) xm ] x = 1 or the characteristic equation satises xm+1 (1 + ) x + 1 = 0. 16.12 Here the service time distribution is given by dB (t) = dt
k

(i)

(ii)

d i (t T i )
i=1

and this Laplace transform equals


k

s (s) =
i=1

di es Ti

(i)

substituting (i) into (15.219), we get A(z ) = s ( (1 z ))


k

=
i=1 k

di e Ti (1z) di e Ti e Ti z
i=1 k ( Ti )j z j j!

= =
i=1

di e

Ti j =0

=
j =0

aj z j .

Hence
k

aj =
i=1

di e Ti

( T i ) j , j!

j = 0, 1, 2, .

(i)

242 To get an explicit formula for the steady state probabilities {qn }, we can make use of the analysis in (16.194)-(16.204) for an M/G/1 queue. From (16.203)-(16.204), let
n

c0 = 1 a 0 ,
(m)

cn = 1
k=0

ak ,

n1

and let {ck } represent the mfold convolution of the sequence {ck } with itself. Then the steady-state probabilities are given by (16.203) as
n

qn = (1 )
m=0 k=0

ak cnk .

(m)

(b) State-Dependent Service Distribution Let Bi (t) represent the service-time distribution for those customers entering the system, where the most recent departure left i customers in the queue. In that case, (15.218) modies to ak,i = P {Ak |Bi } where Ak = k customers arrive during a service time and Bi = i customers in the system at the most recent departure. This gives

ak,i =

et

(t)k dBi (t) k! (t)k 1 k et 1 e1 t dt = , i=0 k! ( + 1 )k+1 0

(i)

et

2 k (t)k 2 e2 t dt = , i1 k! ( + 2 )k+1

243 This gives


A i (z ) =
k=0

ak,i z k =

1 , i=0 1 + 1 (1 z ) 1 , i1 1 + 2 (1 z )

(ii)

where 1 = /1 , 2 = /2 . Proceeding as in Example 15.24, the steady state probabilities satisfy [(15.210) gets modied]
j +1

qj = q0 aj,0 +
i=1

qi aj i+1,i

(iii)

and (see(15.212))

Q (z ) =
j =0

qj z j

= q0
j =0

aj,0 z j +
j =0

qi aj i+1,i (iv)

= q 0 A 0 (z ) +
i=1

qi z i
m=0

am,i z m z 1

= q0 A0 (z ) + (Q(z ) q0 ) A1 (z )/z where (see (ii)) A 0 (z ) = and A 1 (z ) = From (iv) Q (z ) = q0 (z A0 (z ) A1 (z )) . z A 1 (z ) (vii) 1 . 1 + 2 (1 z ) (vi) 1 1 + 1 (1 z ) (v)

244 Since Q(1) = 1 = = we obtain q0 = 1 2 . 1 + 1 2 (viii) q0 A0 (1) + A0 (1) A1 (1) 1 A1 (1)

q0 (1 + 1 2 ) 1 2

Substituting (viii) into (vii) we can rewrite Q(z ) as Q(z ) = (1 2 ) = (1 z ) A1 (z ) 1 z A0 (z )/A1 (z ) 1 A 1 (z ) z 1 + 1 2 1z 1 1 1 + 1 2 1


2 1+1 1 1+1

1 2 1 2 z

z z

= Q 1 (z ) Q 2 (z ) (ix) where Q 1 (z ) = and 1 Q 2 (z ) = 1 + 1 2 2 1 z 1 + 1


i=0 1 2 k k = (1 2 ) 2z 1 2 z k=0

1 1 + 1

zi.

Finally substituting. Q1 (z ) and Q2 (z ) into (ix) we obtain qn = q 0

n i=0

1 1 + 1

ni

n1

i 2
i=0

+1 i 2

i1 n 1 . n = 1, 2, (1 + 1 )ni

with q0 as in (viii).

245 16.13 From (16-209), the Laplace transform of the waiting time distribution is given by w (s) = = 1 1 1
1s (s) s

1
1s (s) s

(i) .

Let F r (t ) =
t 0

[1 B ( )]d
t 0

(ii) B ( )d .

= t

represent the residual service time distribution. Then its Laplace transform is given by F (s) = L {Fr (t)} = = 1 s (s) . s 1 s (s) s s (iii)

Substituting (iii) into (i) we get w (s) =


1 = (1 ) [ F (s)]n , 1 F (s) n=0

|F (s)| < 1.

(iv)

Taking inverse transform of (iv) we get

Fw (t) = (1 )
n=0

n Fr(n) (t),

where Fr(n) (t) is the nth convolution of Fr (t) with itself. 16.14 Let in (16.198) that represents the average number of customers that arrive during any service period be greater than one. Notice that

246

= A (1) > 1 where

A (z ) =
k=0

ak z k

From Theorem 15.9 on Extinction probability (pages 759-760) it follows that if = A (1) > 1, the eqution A (z ) = z (i)

has a unique positive root 0 < 1. On the other hand, the transient state probabilities {i } satisfy the equation (15.236). By direct substii tution with xi = 0 we get

pij xj =
j =1 j =1

j aj i+1 0

(ii)

where we have made use of pij = aj i+1 , i 1 in (15.33) for an M/G/1 queue. Using k = j i + 1 in (ii), it reduces to
k+i1 i1 ak 0 = 0 k=2i k=0 k ak 0

i i1 = xi 0 = 0 = 0

(iii)

since 0 satises (i). Thus if > 1, the M/G/1 system is transient i with probabilities i = 0 . 16.15 (a) The transition probability matrix here is the truncated version of (15.34) given by

247

P =

m2

a0 a1 a2 a0 a1 a2 0 . . . 0 0 a 0 a1 . . . 0 0 . . . 0 0

. . .

am2 am2 am3 . . . a1 a0

a0 0

m2 ak 1 k=0 m3 1 ak k=0 . . . 1 (a 0 + a 1 )
k=0

ak

(i)

1 a0

and it corresponds to the upper left hand block matrix in (15.34) followed by an mth column that makes each row sum equal to unity. (b) By direct sybstitution of (i) into (15-167), the steady state prob m1 abilities {qj }j =0 satisfy
j +1 qj

q0

aj +
i=1

qi aj i+1 ,

j = 0, 1, 2, , m 2

(ii)

and the normalization condition gives


m2 qm 1

=1
i=0

qi .

(iii)

Notice that (ii) in the same as the rst m 1 equations in (15-210) m1 for an M/G/1 queue. Hence the desired solution {qj }j =0 must satisfy the rst m 1 equations in (15-210) as well. Since the unique solution set to (15.210) is given by {qj } j =0 in (16.203), it follows that the desired probabilities satisfy
= c qj , qj

j = 0, 1, 2, , m 1

(iv)

1 where {qj }m j =0 are as in (16.203) for an M/G/1 queue. From (iii) we also get the normalization constant c to be

248

c=

1
m1

. qi

(v)

i=0

16.16 (a) The event {X (t) = k } can occur in several mutually exclusive ways, viz., in the interval (0, t), n customers arrive and k of them continue their service beyond t. Let An = n arrivals in (0, t), and Bk,n =exactly k services among the n arrivals continue beyond t, then by the theorem of total probability

P { X (t ) = k } =
n=k

P {An Bk,n } =
n=k

P {Bk,n |An }P (An ).

But P (An ) = et (t)n /n!, and to evaluate P {Bk,n |An }, we argue as follows: From (9.28), under the condition that there are n arrivals in (0, t), the joint distribution of the arrival instants agrees with the joint distribution of n independent random variables arranged in increasing order and distributed uniformly in (0, t). Hence the probability that a service time S does not terminate by t, given that its starting time x has a uniform distribution in (0, t) is given by pt =
t 0

P (S > t x|x = x)fx (x)dx

t 0

1 1 [1 B (t x)] dx = t t

t 0

(1 B ( )) d =

(t) t

Thus Bk,n given An has a Binomial distribution, so that n k p (1 pt )nk , k t

P {Bk,n |An } =

k = 0, 1, 2, n,

249 and

P { X (t ) = k } =
n=k

et

(t)n n n! k
k n=k

(t ) t t 1 t
t 0

1 t

t 0

nk

B ( )d
nk

= et =

[(t)] k!

B ( )d

(n k )!

[(t)]k t 0t B ( )d e k! [(t)]k t [1B ( )]d e 0 = k! [(t)]k (t) = e , k = 0, 1, 2, k! (i) (b)


t

lim (t) =

[1 B ( )]d (ii)

= E {s }

where we have made use of (5-52)-(5-53). Using (ii) in (i), we obtain lim P {x(t) = k } = e k k! (iii)

where = E {s}.

You might also like