0
0
0 0 . . .
1
(
1
+
1
)
1
0 . . .
0
2
(
2
+
2
)
2
. . .
0 0
3
(
3
+
3
) . . .
.
.
.
.
.
.
.
.
.
.
.
.
_
_
(4.2.1)
is called the innitesimal generator of the process. The parameters
i
and
i
are
called, respectively, the innitesimal birth and death rates. The process x(t) is
known as a birth and death process. In Postulates 1 and 2 we are assuming that
if the process starts in state i, then in a small interval of time the probabilities of the
population increasing or decreasing by 1 are essentially proportional to the length
of the interval. Sometimes a transition from zero to some ignored state is allowed.
Since the p
i j
(t) are probabilities, we have p
i j
(t) 0 and
j=0
p
i j
(t) = 1.
4.2. BIRTH AND DEATH PROCESSES 113
In order to obtain the probability that x(t) = n, we must specify where the
process starts or more generally the probability distribution for the initial state. We
then have
P{x(t) = n} =
i=0
P{x(0) = i}p
in
(t).
4.2.1. Waiting times
With the aid of the above assumptions we may calculate the distribution of the
random variable t
i
which is the waiting time of x(t) in state i; that is, given the
process in state i, we are interested in calculating the distribution of the time t
i
until
it rst leaves state i. Letting
P{t
i
t} = G
i
(t)
it follows easily by the Markov property that as h 0
G
i
(t + h) = G
i
(t)G
i
(h) = G
i
(t)(p
ii
(h) + o(h)) = G
i
(t)[1 (
i
+
i
)h] + o(h)
or
G
i
(t + h) G
i
(t)
h
= (
i
+
i
)G
i
(t) + o(1),
so that
G
i
(t) = (
i
+
i
)G
i
(t).
If we use the condition G
i
(0) = 1, the solution of this equation is
G
i
(t) = e
(
i
+
i
)t
.
That is, t
i
follows an exponential distribution with mean (
i
+
i
)
1
.
According to Postulates 1 and 2, during a time duration of length h a transition
occurs from state i to i + 1 with probability
i
h + o(h) and from state i to i 1 with
probability
i
h + o(h). It follows intuitively that, given that a transition occurs at
time t, the probability this transition is to state i + 1 is
i
(
i
+
i
)
1
and to state i 1
is
i
(
i
+
i
)
1
.
The description of the motion of x(t) is as follows: The process sojourns in a
given state i for a random length of time following an exponential distribution with
parameter
i
+
i
. When leaving state i the process enters either state i + 1 or state
i 1 with probabilities
i
i
+
i
, or,
i
i
+
i
, respectively. The motion is analogous to
114 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
that of a random walk except that transitions occur at random times rather than at
xed time periods.
4.2.2. The Kolmogorov equations
The transition probabilities p
i j
(t) satisfy a systemof dierential equations known
as the backward Kolmogorov dierential equations. These are given by
p
0j
(t) =
0
p
0j
(t) +
0
p
1j
(t), (4.2.2)
p
i j
(t) =
1
p
i1, j
(t) (
i
+
i
)p
i j
(t) +
i+1, j
(t), i 1,
and the boundary condition p
i j
(0) =
i j
.
To derive these we have from (4.2.1)
p
i j
(t + h) =
k=0
p
ik
(h)p
k j
(t) = p
i,i1
(h)p
i1, j
(t) +
p
ii
(h)p
i j
(t) + p
i,i+1
p
i+1, j
(t) +
k
p
ik
(h)p
k j
(t), (4.2.3)
where the last summation is over all k i 1, i, i + 1. Using Postulates 1, 2, and 3
we obtain
k
p
ik
(h)p
k j
(t)
k
p
ik
(h)
= 1 [p
ii
(h) + p
i,i1
(h) + p
i,i+1
(h)]
= 1 [1 (
i
+
i
)h + o(h) +
i
h + o(h) +
i
h + o(h)]
= o(h),
so that
p
i j
(t + h) =
i
hp
i1, j
(t) + (1 (
i
+
i
)h)p
i j
(t) +
i
hp
i+1, j
(t) + o(h).
Transposing the terms p
i j
(t) to the lefthand side and dividing the equation by h, we
obtain the equations (4.2.2), after letting h 0.
The backward equations are deduced by decomposing the time interval (0, t +h),
where h is positive and small, into the two periods
(0, h), (h, t + h),
4.2. BIRTH AND DEATH PROCESSES 115
and examining the transitions in each period separately. The equations (4.2.2) fea
ture the initial state as the variable. A dierent result arises from splitting the time
interval (0, t + h) into the two periods
(0, t), (t, t + h)
and adapting the preceding analysis. In this viewpoint, under more stringent condi
tions, we can derive a further system of dierential equations
p
i0
(t) =
0
p
i0
(t) +
1
p
i1
(t), (4.2.4)
p
i j
(t) =
j1
p
i, j1
(t) (
j
+
j
)p
i j
(t) +
j+1
p
i, j+1
(t), j 1,
with the same initial condition p
i j
(0) =
i j
. These are known as the forward Kol
mogorov dierential equations. To do this we interchange t and h in (4.2.3) and
under stronger assumptions in addition to Postulates 1, 2, and 3 it can be shown that
the last term is again o(h). The remainder of the argument is the same as before.
In general the innitesimal parameters {
n
,
n
} may not determine a unique sto
chastic process obeying
j=0
p
mn
(t) = 1 and p
mn
(t + s) =
k=0
p
mn
(t)p
kn
(s).
A sucient condition that there exists a unique Markov process with p
mn
(t)
satisfying the above is that
n=0
n
n
k=0
1
n
= ,
where
0
= 1 and
n
=
n1
n1
n
=
0
. . .
n1
1
. . .
n
, n S \ {0}. (4.2.5)
The quantities
n
are called the potential coecients.
We discuss nowbriey the behavior of p
i j
(t) as t becomes large. It can be proved
that the limits
lim
t
p
i j
(t) = p
j
(4.2.6)
116 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
exist and are independent of the initial state i and also that they satisfy the equations
0 =
0
p
0
+
1
p
1
(4.2.7)
0 =
n1
p
n1
(
n
+
n
)p
n
+
n+1
p
n+1
, n 1. (4.2.8)
which is obtained by setting the lefthand side of (4.2.4) equal to zero. The con
vergence of
_
j
p
j
follows since
_
j
p
i j
(t) = 1. If
_
j
p
j
= 1, then the sequence {p
j
}
is called a stationary distribution (steady state probabilities). The reason for this is
that p
j
also satisfy
p
j
=
i=0
p
i
p
i j
(t), (4.2.9)
which tells us that if the process starts in state i with probability p
i
, then at any given
time t it will be in state i with same probability p
i
. The proof of (4.2.9) follows from
(4.2.1) and (4.2.6) and if we let t and use the fact that
_
i=0
p
i
< . The solution
to (4.2.7) and (4.2.8) is obtained by rewriting them as
n+1
p
n+1
n
p
n
=
n
p
n
n1
p
n1
= =
1
p
1
0
p
0
= 0.
Therefore,
p
j
=
j
p
0
, j = 1, 2, . . . . (4.2.10)
In order that the sequence {p
j
} dene a distribution we must have
_
j=0
p
j
= 1.
If
_
k
< , we see in this case that
p
j
=
j
_
k=0
k
, j = 0, 1, 2, . . . .
If
_
k
= , then necessarily p
0
= 0 and the p
j
are all zero. Hence, we do have a
limiting stationary distribution.
Example 4.2.1. Poisson process: For this process, the rates are given by
n
=
(constant) and
n
= 0. This process plays an important role as an arrival process in many
contexts. In this case, the value of j in the function f
j
of the random variable technique is 1
and
f
1
(x(t)) = , independent of x(t).
The probability generating function P(s, t) is given by
p
t
= (s 1)P, P(s, 0) = s
4.2. BIRTH AND DEATH PROCESSES 117
which leads to P(s, t) = e
(s1)t
. Therefore,
P
n
(t) = e
t
(t)
n
n!
, n = 0, 1, . . .
E[x(t)] = t and Var[x(t)] = t, t 0.
For a Poisson process, the time between two subsequent occurrences of events is expo
nentially distributed with parameter . That is, if t
1
denotes the time of the rst event and t
n
denotes the time between the (n 1)
th
and n
th
events, the random variables {t
i
, i = 1, 2, . . . }
are independent exponentially distributed with the parameter of the Poisson process.
Example 4.2.2. Simple birth process (Yule process): For this process, the rates are
given by
n
= n and
n
= 0. This process models a population in which each member acts
independently and gives birth at an exponential rate . The simplest example of a pure birth
process is the Poisson process, which has a constant birth rate
n
= , n 0. Here
f
1
(x(t)) = x(t).
Starting with a single individual, the probability generating function using random vari
able technique P(s, t) can be given by
P
t
= s(s 1)
P
s
, P(s, 0) = s.
Then,
P(s, t) =
_
1 e
t
_
1
1
s
__
1
.
Therefore, {x(t)} follows a geometric distribution. Hence,
E[x(t)] = e
t
and Var[x(t)] = e
t
(e
t
1).
Example 4.2.3. The rates are given by
n
= (N n),
n+1
= (n + 1), n = 0, 1, . . . , N 1.
Thus,
P
n
(t) =
m
( + )
N
_
1 e
(+)t
_
m
_
+ e
(+)t
_
Nm
i=0
_
m
i
_ _
N m
n i
_ _
+ e
(+)t
(1 e
(+)t
)
_
i
_
(1 e
(+)t
)
+ e
(+)t
_
ni
,
118 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
where x(0) = m. The mean and variance of x(t) at time t are given by
E[x(t)] =
1
+
_
N e
(+)t
(N m( + ))
_
,
Var[x(t)] =
1
( + )
2
_
N e
(+)
(N( ) m(
2
2
)
2
)
e
2(+)t
(
2
N + m(
2
2
))
_
.
Exercises 4.2.
4.2.1. If the birth and death rates are given by
n
= ,
n
= n. Obtain the prob
ability generating function of the number of customers at time t.( M/M/, Innite
server queue )
4.2.2. For
n
= ,
n
= n, obtain P
00
(t). (M/M/1 queueing system)
4.2.3. For a BDP on {0,1} obtain the transition probabilities.
4.2.4. For a BDP with rates
0
= 1 and other
n
= 1/2 and
n
= 1/2, show that
P
00
(t) = e
t
I
0
(t).
4.2.5. Arrivals to a supermarket is a Poisson process with arrival rate 10 per hour.
Given that 100 persons arrived up to 1 PM, what is the probability that 2 more per
sons will arrive in 10 minutes? What is the average number of arrivals between 11
AM and 1 PM?
4.2.6. Consider a Poisson process. Given that an event has occurred before T nd
the pdf of this arrival.
4.2.7. Arrivals at a telephone booth are considered to be Poisson, with an average
time of 10 minutes between one arrival and the next. The length of a phone call is
assumed to be distributed exponentially with mean 3 minutes. Find the probability
that a person arriving at the booth will have to wait and his average time spent in
the booth.
4.3. CONTINUED FRACTIONS 119
4.3. Continued Fractions
A CF is denoted by
a
1
b
1
+
a
2
b
2
+
a
3
b
3
+
or economically by
a
1
b
1
+
a
2
b
2
+
a
3
b
3
+ , (4.3.1)
where a
n
and b
n
are real or complex numbers. This fraction can be terminated by
retaining the terms a
1
, b
1
, a
2
, b
2
, . . . , a
n
, b
n
and dropping all the remaining terms
a
n+1
, b
n+1
, . . . . The number obtained by this operation is called the n
th
convergent
or n
th
approximant and is denoted by A
n
/B
n
. Both A
n
and B
n
satisfy the recurrence
relation
U
n
= a
n
U
n2
+ b
n
U
n1
(4.3.2)
with initial values A
0
= 0, A
1
= a
1
and B
0
= 1, B
1
= b
1
.
A
2n
B
2n
(
A
2n+1
B
2n+1
) is called the
even (odd) part of the CF.
A CF is denoted by
a
1
b
1
+
a
2
b
2
+
a
3
b
3
+
or economically by
a
1
b
1
+
a
2
b
2
+
a
3
b
3
+ , (4.3.3)
where a
n
and b
n
are real or complex numbers. This fraction can be terminated by
retaining the terms a
1
, b
1
, a
2
, b
2
, . . . , a
n
, b
n
and dropping all the remaining terms
a
n+1
, b
n+1
, . . . . The number obtained by this operation is called the n
th
convergent
or n
th
approximant and is denoted by A
n
/B
n
. Both A
n
and B
n
satisfy the recurrence
relation
U
n
= a
n
U
n2
+ b
n
U
n1
(4.3.4)
120 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
with initial values A
0
= 0, A
1
= a
1
and B
0
= 1, B
1
= b
1
.
A
2n
B
2n
(
A
2n+1
B
2n+1
) is called the
even (odd) part of the CF.
A CF is said to be convergent if at most a nite number of its denominators B
n
vanish and the limit of its sequence of approximants
lim
n
A
n
B
n
, (4.3.5)
exists and is nite. Otherwise, the CF is said to be divergent. If
a
n
> 0 , b
n
> 0 ,
n=1
_
b
n
b
n1
a
n+1
= ,
then the CF is convergent. The value of a CF is dened to be the limit (4.3.5). No
value is assigned to a divergent CF.
For any convergent CF, the exact value of the fraction is between any two neigh
boring convergents. All even numbered convergents lie to the left of the exact value,
that is, they give an approximation to the exact value by defect. All odd numbered
convergents lie to the right of the exact value, that is, they give an approximation
to the exact value by excess. For a terminating CF, that is, a fraction with nite
number of terms, the last convergent coincides with the exact value. If a CF is
nonterminating, the sequence of convergents is innite. For example,
2 = 1 + (
2 1) = 1 +
1
1 +
2
= 1 +
1
2
+
1
2
+ .
The successive convergents are
1
1
,
3
2
,
7
5
,
17
12
,
41
29
,
99
70
,
239
169
, . . . .
That is,
1.0, 1.5, 1.4, 1.417, 1.4138, 1.41429, 1.41420, . . . .
Observe that
7
5
<
2 <
3
2
,
41
29
<
2 <
99
70
, and so on.
Using the power series expansion of
8 tan
_
h
2
_
tan(h)
3h
=
5
480
h
4
5
7
16 7!
h
6
,
4.3. CONTINUED FRACTIONS 121
we get a CF representation of as follows:
= 3 +
1
7
+
1
15
+
1
1
+
1
292
+
1
1
+ .
Its successive convergents are
3
1
,
22
7
,
333
106
,
355
113
, We observe that
355
113
< 3 10
7
.
This shows how fast the CF is converging to the exact value. This result is spectac
ular. However,
e
x
2
2
_
x
e
t
2
2
dt =
1
x
+
1
x
+
2
x
+
3
x
+
requires about 70 terms to pin the error down to 2.4 10
7
when x = 1.
It is wellknown that a sequence of OPs {P
n
(x)} satisfy a threeterm recurrence
relation
P
n+1
(x) = (x + b
n
)P
n
(x) a
n
P
n1
(x) , n 0 (4.3.6)
with P
0
= 1, P
1
= 0 and a
n
> 0. This recurrence relation when compared with the
recurrence relation (4.3.4) suggests consideration of the Jacobi CF
1
x + b
1
a
1
x + b
2
a
2
x + b
3
.
In this case the n
th
convergent is a rational function A
n1
(x)/B
n
(x) whose de
nominator B
n
(x) is a polynomial of degree n and numerator A
n1
(x) is a poly
nomial of degree n 1. The sequences {B
n
(x)} and {A
n
(x)} satisfy (4.3.6) with
B
0
(x) = 1 , B
1
(x) = x + b
1
and A
0
(x) = 0 , A
1
(x) = 1. Further, A
n1
(x) form an
associated OP system and the zeros of A
n1
(x) and B
n
(x) are distinct and interlace.
4.3.1. CFs and BDPs
In this section, approximate transient system size probability values for an in
nite BDP {x(t), t 0} is obtained using the CF technique discussed below. Dene
L
r1
=
r1
_
k=0
k
; M
r
=
r
_
k=1
k
, r = 0, 1, 2, . . .
122 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
with L
1
= 1 = M
0
. Dene
f
r
(s) = (1)
r
M
r
_
0
e
st
P
r
(t)dt, r = 0, 1, 2, . . . .
Taking the Laplace transform of the system of equations given by (4.2.4) and as
suming m = 0, f
0
(s) gives the expression
f
0
(s) =
1
s +
0
+
f
1
(s)
f
0
(s)
.
Similarly,
f
r
(s)
f
r1
(s)
=
r1
r
s +
r
+
r
+
f
r+1
(s)
f
r
(s)
, r = 1, 2, 3, . . . .
This leads to
f
0
(s) =
1
s +
0
1
s +
1
+
1
2
s +
2
+
2
. (4.3.7)
Let the n
th
approximant of f
0
(s) be given by
A
n1
(s)
B
n
(s)
where
A
0
(s) = 1
A
1
(s) = s +
1
+
1
A
n
(s) = (s +
n1
+
n1
)A
n1
(s)
n2
n1
A
n2
(s), n = 2, 4, . . . ,
and
B
1
(s) = s +
0
B
2
(s) = (s +
1
+
1
)B
1
(s)
0
1
B
n
(s) = (s +
n1
+
n1
)B
n1
(s)
n2
n1
B
n2
(s), n = 3, 4, . . . .
Then
f
0
(s)
A
n1
(s)
B
n
(s)
, (4.3.8)
4.3. CONTINUED FRACTIONS 123
where signies approximately and B
n
(s) can be written in tridiagonal determi
nant form as follows:
B
n
(s) =
s +
0
1
1
s +
1
+
1
1
2
s +
2
+
2
1
.
.
.
n2
n1
s +
n1
+
n1
nn
. (4.3.9)
A
n1
(s) is obtained from B
n
(s) by deleting the rst row and rst column.
We observe that B
n
(s) given by (4.3.9) is clearly zero when s is an eigenvalue
of the matrix
F
n
=
_
0
1
0
1
+
1
2
1
2
+
2
3
.
.
.
n2
n1
+
n1
_
_
nn
.
This matrix is quasisymmetric and can be transformed into a real symmetric matrix
E
n
by a similarity transformation E
n
:= D
1
n
F
n
D
n
, where
D
n
:= Diag
_
1,
_
L
0
M
1
,
_
L
1
M
2
, . . . ,
_
L
n2
M
n1
_
.
The matrix thus obtained is given by
E
n
=
_
1
1
+
1
2
2
+
2
3
.
.
.
n2
n1
n1
+
n1
_
_
nn
.
This is a real symmetric diagonal dominant positive denite tridiagonal matrix with
nonzero subdiagonal elements and therefore the eigenvalues are real and distinct.
We denote these eigenvalues by s
(n)
1
, s
(n)
2
, s
(n)
3
, . . . , s
(n)
n
, that is, s
(n)
1
, s
(n)
2
, s
(n)
3
, . . . , s
(n)
n
are the roots of B
n
(s). Similarly, the roots of A
n1
(s) are negative, real and distinct
and let z
(n)
1
, z
(n)
2
, z
(n)
3
, . . . , z
(n)
n1
be its roots.
124 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
Using partial fractions, (4.3.8) can be expressed as
f
0
(s)
n
j=1
n1
_
r=1
(z
(n)
r
s
(n)
j
)
(s + s
(n)
j
)
n
_
r=1,rj
(s
(n)
r
s
(n)
j
)
.
On inverting we get,
P
0
(t)
n
j=1
n1
_
r=1
(z
(n)
r
s
(n)
j
)
n
_
r=1,rj
(s
(n)
r
s
(n)
j
)
e
s
(n)
j
t
.
Example 4.3.1. BDP with constant birth and death rates: The rates are given by
n
= ;
n+1
= , n = 0, 1, 2, . . . , N 1. (4.3.10)
This process corresponds to an M/M/1/N queuing model with arrival rate and service rate
. We get the roots of s
k
of B
N+1
(s) as follows:
s
0
= 0, s
k
= + 2
_
cos
k
N + 1
, k = 1, 2, ..., N.
By induction, for i = 0, 1, 2, . . . , N,
B
i
(s
k
) =
_
i
, if k = 0
4
i
()
i
2
y
i,k
4
i1
()
i1
2
y
i1,k
, if k = 1, 2, 3, . . . , N.
,
where, for k = 1, 2, 3, . . . , N and i = 2, 3, 4, . . . , N + 1,
y
i1,k
=
i
_
j=1
sin
__
(i + 1)k + (N + 1) j
_
2(N + 1)(i + 1)
_
sin
__
(i + 1)k (N + 1) j
_
2(N + 1)(i + 1)
_
.
4.3. CONTINUED FRACTIONS 125
We get for n = 0, 1, 2, . . . , N,
P
n
(t) =
Nn
n
N
_
j=1
_
+ 2
_
cos
j
N + 1
_
+4
n+m2
k=1
_
4
Y
n,k
Y
n1,k
_ _
4
Y
m,k
y
m1,k
_
y
N1,k
e
t(+2
cos
k
N+1
)
_
1 + 2
cos
k
N + 1
_
N
_
j=1, jk
sin
_
( j + k)
2(N + 1)
_
sin
_
( j k)
2(N + 1)
_
,
where =
n
=
n( + n 1 )
( + 2n 1)( + 2n)
, n = 1, 2, 3, . . .
CF : z
2
F
1
( + 1, 1; + 1, z) =
z
1
+
0
z
1
+
1
z
1
+
1
z
1
+
2
z
1
+
P
00
(t) =
k=0
(1)
k
( + 1)
k
( + 1)
k
t
k
k!
=
1
F
1
( + 1, + 1, t).
In particular, if = 0 and = 1, then
P
00
(t) =
1 e
t
t
.
Example 4.3.3.
Rates :
n1
= (N n + 1)p ,
n
= nq , n = 1, 2, , N
CF :
N
j=0
b(N, j; p)
z + j
=
1
z
+
Np
1
+
q
z
+
(N 1)p
1
+
2q
z
+ +
Nq
z
, (z) > 0
where b(N, j; p) :=
_
N
j
_
p
j
q
Nj
, j = 0, 1, 2, , N , q = 1 p
P
00
(t) =
N
j=0
b(N, j; p)e
jt
.
126 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
Example 4.3.4.
Rates :
n1
=
1
2n
,
n
=
1
2n + 1
, n = 1, 2, 3,
CF :
1
F
1
(1; z + 1; z) = 1 +
2z
2
+
3z
3
+
4z
4
+
5z
5
+
for any complex number z,
P
00
(t) =
m=1
e
m
m
m2
(m 1)!
e
t/m
.
P
00
(0) = 1
Exercises 4.3.
4.3.1. If the rates are
n1
=
U
n
(1/)
2U
n1
(1/)
,
n
=
U
n1
(1/)
2U
n
(1/)
, < 1, n = 1, 2, 3, , show
that P
00
(t) = 2e
t
I
1
(t)
t
.
4.3.2. For the rates
n1
=
nr
1 r
,
n
=
n
1 r
, r < 1, n = 1, 2, 3, ,
prove that P
00
(t) =
1r
1re
t
.
4.3.3. For a BDP with rates
n
= 1/2 and
n
= n/(2n + 1), show that P
00
(t) =
1e
t
t
.
4.3.4. Obtain a CF expansion of
5
4.3.5. Express
3551
1132
as a CF and calculate the rst four convergents.
4.3.6. Use recurrence relation to evaluate the tridiagonal determinant with 2cos in
the diagonal and 1 in the subdiagonal and superdiagonal.
4.3.7. Give one CF expansion of .
4.4. ORTHOGONAL POLYNOMIALS 127
4.4. Orthogonal Polynomials
Denition 4.4.1. A sequence of polynomials {Q
n
(x)}, where Q
n
(x) is of exact
degree n in x, is said to be orthogonal with respect to a LebesgueStieltjes measure
d(x) if
_
Q
m
(x)Q
n
(x)d(x) = 0, m n. (4.4.1)
Implicit in this denition is the assumption that the moments
m
n
=
_
x
n
d(x), n = 0, 1, 2, . . . (4.4.2)
exist and are nite. If the nondecreasing, realvalued, bounded function (x) also
happens to be absolutely continuous with d(x) = w(x)dx, w(x) 0, then (4.4.1)
reduces to
_
Q
m
(x)Q
n
(x)w(x)dx = 0, m n, (4.4.3)
and the sequence {Q
n
(x)} is said to be orthogonal with respect to the weight function
w(x). If, on the other hand, (x) is a stepfunction with jumps w
j
at x = x
j
, j =
0, 1, 2, . . . , then (4.4.1) takes the form of a sum:
j=0
Q
m
(x
j
)Q
n
(x
j
)w
j
= 0, m n. (4.4.4)
In this case we refer to the sequence {Q
n
(x)} as OPs of a discrete variable.
General OPs are characterized by the fact that they satisfy a threeterm recur
rence relation of the form
xQ
n
(x) = a
n
Q
n+1
(x) + b
n
Q
n
(x) + c
n
Q
n1
(x), (4.4.5)
where Q
1
(x) = 0, Q
0
(x) = 1, a
n
, b
n
, c
n
are real and a
n
c
n+1
> 0 for n = 0, 1, 2, . . . .
This is an important characterization which has been the starting point for much of
the modern work on general OPs. Monic OPs for a positive Borel measure on the
real line are polynomials P
n
, n = 0, 1, 2, . . . , of degree n and leading coecient one
such that
_
P
n
(x)x
k
d(x) = 0, k = 0, 1, 2, . . . , n 1. (4.4.6)
128 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
These n orthogonality conditions give n linear equations for the n unknown coe
cients a
k,n
, k = 0, 1, 2, . . . , n1, of the monic polynomial P
n
(x) =
_
n
k=0
a
kn
x
k
, where
a
nn
= 1. This system of n equations for n unknowns always has a unique solution
since the matrix with entries
_
x
i+j
d(x), 0 i, j n 1, known as Gram matrix,
is positive denite and hence nonsingular. It is well known that such polynomials
satisfy a threeterm recurrence relation
P
n+1
(x) = (x b
n
)P
n
(x) a
2
n
P
n1
(x), n 0 (4.4.7)
with P
0
= 1 and P
1
= 0. Often it is more convenient to consider the orthonormal
polynomials p
n
(x) =
n
P
n
(x) for which
_
p
n
(x)p
m
(x)d(x) =
mn
, m, m 0 (4.4.8)
so that
n
=
__
P
2
n
(x)d(x)
_
1/2
> 0. The recurrence relation for these orthonormal
polynomials is given by
xp
n
(x) = a
n+1
p
n+1
(x) + b
n
p
n
(x) + a
n
p
n1
(x), n 0, (4.4.9)
where
a
n
=
_
xp
n1
(x)p
n
(x)d(x) =
n1
n
, b
n
=
_
xp
2
n
(x)d(x). (4.4.10)
Let us mention at this point the best known examples: the classical OPs of
Jacobi, Laguerre and Hermite. These three families of OPs satisfy a second order
linear dierential equation, their derivatives are again OPs of the same family but
with dierent parameters.
Apart from the threeterm recurrence relations, these OPs satisfy a number of
relationships of the same general form. They satisfy the second order dierential
equation
g
2
(x)p
n
+ g
1
(x)p
1
+ d
n
p
n
= 0.
Between any two zeros of p
n
(x) there is a zero of p
n+1
(x). Many interesting proper
ties of these polynomials depend on their connection with problems of interpolation
and mechanical quadrature.
Remarkably for many specic BDPs of practical interest, the weight function
d(x) has been explicitly determined and general formulas are derived for the OPs.
The spectral representation then gives considerable insight into the timedependent
behavior of the BDPs.
4.4. ORTHOGONAL POLYNOMIALS 129
For each system of OPs R
n
satisfying the recurrence relation
xR
n
(x) = A
n
R
n+1
(x) (A
n
+ C
n
)R
n
(x) + C
n
R
n1
(x),
we can introduce an extra scaling parameter 0 and study the polynomials
S
n
(x) = R
n
(x), which satisfy the recurrence relation
xS
n
(x) =
A
n
S
n+1
(x)
(A
n
+ C
n
)
S
n
(x) +
C
n
S
n1
(x).
If the polynomials R
n
are orthogonal on the nite set {x
0
, x
1
, . . . , x
N
}, then the poly
nomials S
n
are orthogonal on the set {x
0
/, x
1
/, . . . , x
N
/}
Example 4.4.1. Chebyshev polynomials of the second kind: Chebyshev polynomials
of the second kind U
n
satisfy the recurrence relation
2xU
n
(x) = U
n+1
(x) + U
n1
(x), n 1, (4.4.11)
with U
0
(x) = 1 and U
1
(x) = 2x. They are orthogonal on the interval [1, 1] with respect to
the weight
1 x
2
.
If we consider only the nite set {U
0
, U
1
, . . . , U
N
}, then this is also orthogonal on the
set {cos j/(N + 2), j = 1, 2, . . . , N + 1}, i.e., the roots of U
N+1
. Consider orthonormal
polynomials p
n
satisfying the relation
xp
n
(x) = a
n+1
p
n+1
(x) + b
n
p
n
(x) + a
n
p
n1
(x),
with p
0
(x) = 1 and p
1
(x) = 0. Then the Chebyshev polynomials of the second kind are the
orthonormal polynomials with constant recurrence coecients a
n
= 1/2 (n = 1, 2, . . .) and
b
n
= 0 (n = 0, 1, 2, . . .). If we change the rst recurrence coecient to b
0
= b/2, then we
get polynomials P
n
which are a special case of corecursive OPs. Obviously U
n
and U
n1
are both solutions of the recurrence relation
2xP
n
(x) = P
n+1
(x) + P
n1
(x),
hence the general solution of this second order linear recurrence is P
n
(x) = AU
n
(x) +
BU
n1
(x), where A and B are determined from the initial conditions P
0
(x) = 1 and P
1
(x) =
2x b. This gives
P
n
(x) = U
n
(x) bU
n1
(x).
If we require a nite system of OPs, then we take a
N+1
= 0. Finally taking b
N
= 1/2b
gives
P
N+1
(x) =
_
x
1
2b
_
P
N
(x)
1
2
P
N1
(x)
= xU
N
(x) bxU
N1
(x)
1
2b
U
N
(x) +
b
2
U
N2
(x).
130 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
Now use U
N2
(x) = 2xU
N1
(x) U
N
(x) to nd
P
N+1
(x) =
_
x
b + 1/b
2
_
U
N
(x). (4.4.12)
Hence this nite system of modi ed Chebyshev polynomials, with b
0
= b/2 and b
N
= 1/2b,
is orthogonal on the zeros of P
N+1
, i.e., on
_
b + 1/b
2
, cos
j
N + 1
, j = 1, 2, . . . , N
_
.
We note that the Chebyshev polynomials of the second kind can be written as
U
n
(cos()) =
sin((n + 1))
sin()
, n = 0, 1, . . . ; 0 . (4.4.13)
The modi ed Chebyshev polynomials of second kind V
n
(x) and W
n
(x) are de ned as
follows for n = 0, 1, . . . :
V
n
(cos()) =
cos
_
(n +
1
2
)
_
cos
_
2
_ (4.4.14)
and
W
n
(cos()) =
sin
_
(n +
1
2
)
_
sin(
2
)
. (4.4.15)
The signi cance of these observations is that the properties of sines and cosines can be used
to establish many of the properties of Chebyshev polynomials. Chebyshev polynomials
have acquired great practical importance in polynomial approximation methods. Specif
ically, it has been shown that a series of Chebyshev polynomials converges much more
rapidly than a power series.
Chebyshev polynomials satisfy the generating function
(1 2xt + t
2
)
1
=
n=0
U
n
(x)t
n
.
We can write (4.4.13) as
U
n
(x) =
[n/2]
k=0
_
n k
k
_
(1)
k
(2x)
n2k
.
De ne
V
n
(x) = U
n
(x) U
n1
(x) and W
n
(x) = U
n
(x) + U
n1
(x), n 1,
where V
0
(x) = W
0
(x) = 1. The polynomials V
n
(x) and W
n
(x) are orthogonal in (1, 1) with
weight functions
_
1+x
1x
and
_
1x
1+x
, respectively (Szeg o (1959)).
4.4. ORTHOGONAL POLYNOMIALS 131
Example 4.4.2. Krawtchouk Polynomials: These polynomials K
n
(x; p, N), with 0 <
p < 1, satisfy the recurrence relation
xK
n
(x) = p(N n)K
n+1
(x) [p(N n) + n(1 p)]K
n
(x) + n(1 p)K
n1
(x). (4.4.16)
They are orthogonal on the set {0, 1, 2, . . . , N} with respect to the binomial distribution with
parameters N and p.
Example 4.4.3. Dual Hahn Polynomials H
n
(x; , , N): These polynomials satisfy the
recurrence relation
xH
n
(x) = A
n
H
n+1
(x) (A
n
+ C
n
)H
n
(x) + C
n
H
n1
(x), (4.4.17)
where A
n
= (N n)(n + d + 1), C
n
= n(N n + + 1).
They are orthogonal on the set {n(n + d + + 1), n = 0, 1, 2, . . . , N}.
Example 4.4.4. Racah Polynomials R
n
(x; , , , ): These polynomials with + 1 =
N or + + 1 = N or + 1 = N, satisfy the recurrence relation
xR
n
(x) = A
n
R
n+1
(x) (A
n
+ C
n
)R
n
(x) + C
n
R
n1
(x), (4.4.18)
where
A
n
=
(n + + + 1)(n + + 1)(n + + + 1)(n + d + 1)
(2n + + + 1)(2n + + + 2)
,
C
n
=
n(n + )(n + + d)(n + )
(2n + + )(2n + + + 1)
.
They are orthogonal on the set {n(n + d + + 1), n = 0, 1, 2, . . . , N}.
4.4.1. OPs and BDPs
For every BDP {x(t), t 0} with state space S, one has KarlinMcGregor poly
nomials Q
n
(x) (n = 0, 1, . . . ), satisfying the threeterm recurrence relation
Q
1
(x) := 0, Q
0
(x) := 1
n
Q
n+1
(x) + (x
n
n
)Q
n
(x) +
n
Q
n1
(x) = 0, n 0
(4.4.19)
which may be written in matrix notation as
Q
0
(x) = 1, xQ(x) = Q
Q(x), (4.4.20)
132 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
where Q(x) is the column vector with elements Q
i
(s), i S and
Q =
_
0
1
0
0
1
1
2
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
_
_
. (4.4.21)
We observe that the monic polynomials P
n
(x) := (1)
n
(
0
1
. . .
n
)Q
n
(x), n > 0,
satisfy the recurrence formula
xP
n
(x) = P
n+1
(x) + (
n
+
n
)P
n
(x) +
n1
n
P
n1
(x), n > 0,
P
0
(x) = 1, P
1
(x) = x
0
(4.4.22)
and that P
n
(x) can be interpreted as the characteristic polynomials of the matrix
Q
T
truncated at the n
th
row and column. Since
n1
n
> 0 for n > 0, the recur
rence relation (4.4.22) and hence (4.4.19) denes a sequence of polynomials which
is orthogonal with respect to a positive measure. Moreover, the parameters in the
recurrence relation (4.4.22) are such that {P
n
(x)} is orthogonal with respect to a mea
sure on the nonnegative real axis. Let q
n
(x) denote the n
th
orthonormal polynomial
corresponding to Q
n
(x) or P
n
(x). It can easily be veried that
q
n
(x) =
1/2
n
Q
n
(x), n 0, (4.4.23)
where
n
are given by (4.2.5).
We now obtain the transition probability functions P
n
(t), which satisfy (4.2.4),
in terms of the polynomials Q
n
(x) for the given rates
n
and
n
. Suppose that the
polynomials {q
n
(x)} are orthonormal with respect to the measure d, with support
in [0, ), that is,
_
0
q
n
(x)q
m
(x)d(x) =
m,n
and dene the functions
f
i
(x, t) =
j=0
P
j
(t)Q
j
(x) =
j=0
1/2
j
P
j
(t)q
j
(x), i S (4.4.24)
or equivalently
f (x, t) = P(t)Q(x), (4.4.25)
where P(t) = (P
mn
(t))
m,n=0
. We observe that this matrix satises
P
(t) = P(t)Q
(4.4.26)
4.4. ORTHOGONAL POLYNOMIALS 133
with
P(0) = I, (4.4.27)
where I = (
i j
)
i, j=0
is the innite identity matrix. Using (4.4.20) and (4.4.26), we
note that (4.4.25) satises the equation
t
f (x, t) = P
(t)Q(x) = P(t)Q
Q(x) = x f (x, t)
and by (4.4.27), the initial condition is
f (x, 0) = Q(x).
Hence,
f (x, t) = e
xt
Q(x)
or
f
i
(x, t) = e
xt
Q
i
(x), i S.
By (4.4.24),
1/2
j
P
j
(t) is the j
th
Fourier coecient of f
i
(x, t) with respect to the
orthonormal system {q
n
(x)} and hence
1/2
j
P
j
(t) =
_
0
e
xt
Q
i
(x)q
j
(x)d(x), i, j S
or
P
j
(t) =
j
_
0
e
xt
Q
i
(x)Q
j
(x)d(x), i, j S. (4.4.28)
For every nite BDP {x(t), t 0}
f
with nite state space S
f
one has KarlinMcGregor
polynomials Q
n
, n = 0, 1, 2, . . . , N, dened by the recurrence relation
xQ
n
(x) =
n
Q
n+1
(x) (
n
+
n
)Q
n
(x) +
n
Q
n1
(x), n = 1, 2, . . . , N 1(4.4.29)
with initial values
Q
0
(x) = 1, Q
1
(x) =
0
x
0
and the polynomial Q
N+1
of degree N + 1, dened by
xQ
N
(x) = Q
N+1
(x)
N
Q
N
(x) +
N
Q
N1
(x). (4.4.30)
134 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
Karlin and McGregor have shown that Q
N+1
has N + 1 distinct, real zeros s
0
<
s
1
< s
2
< < s
N
. Furthermore, the polynomials {Q
n
, n = 0, 1, . . . , N} are
orthogonal on the set {s
0
, s
1
, . . . , s
N
}:
N
k=0
Q
n
(s
k
)Q
m
(s
k
)
k
= 0, 0 n m N, (4.4.31)
where
k
are positive weights given by
k
:=
1
N
i=0
Q
2
i
(s
k
)
i
> 0. (4.4.32)
They gave the spectral representation of the P
n
(t) in terms of these polynomials, as
P
n
(t) =
n
N
k=0
e
s
k
t
Q
m
(s
k
)Q
n
(s
k
)
k
, n S
f
. (4.4.33)
The monic orthogonal KarlinMcGregor polynomials, corresponding to the poly
nomials Q
n
(s), are given by P
n
(x) := (1)
n
(
0
1
. . .
n
)Q
n
(x), n = 0, 1, . . . , N and
they satisfy the recurrence formula
xP
n
(x) = P
n+1
(x) + (
n
+
n
)P
n
(x) +
n1
n
P
n1
(x), n = 1, 2, . . . , N (4.4.34)
with P
0
(x) = 1 and P
1
(x) = 0. So we can also nd the monic KarlinMcGregor
polynomial P
N+1
(x), and its zeros are important because the KarlinMcGregor poly
nomials are orthogonal on the set consisting of these zeros.
Example 4.4.5. Consider the BDP discussed in example 4.3.1. For this BDP, the Karlin
McGregor polynomials Q
n
(x) satisfy the recurrence relation
xQ
n
(x) = Q
n+1
(x) ( + )Q
n
(x) + Q
n1
(x), n = 1, 2, . . . , N 1, (4.4.35)
with Q
0
(x) = 1 and Q
1
(x) = ( x)/. Take the modi ed Chebyshev polynomials P
n
(x) =
U
n
(x) bU
n1
(x) with b =
_
/, then the recurrence relation (4.4.11) and the change of
variable
Q
n
(x) =
_
_ n
2
P
n
_
+ x
2
_
give the recurrence (4.4.35) with the initial conditions Q
0
(x) = 1 and Q
1
(x) = ( x)/. So
the KarlinMcGregor polynomials for this nite BDP turn out to be the modi ed Chebyshev
4.4. ORTHOGONAL POLYNOMIALS 135
polynomials
U
n
(y)
_
/ U
n1
(y), y =
+ x
2
,
which are orthogonal on the set {0, + 2
cos
k
N+1
, k = 1, 2, . . . , N} (see Example
4.4.1 for details).
Example 4.4.6. Consider the BDP discussed in example 4.4.2.
The KarlinMcGregor polynomials for this process are the Krawtchouk polyno
mials
K
n
_
x
+
;
+
, N
_
,
which are orthogonal on the set
{0, + , 2( + ), . . . , N( + )}.
Example 4.4.7. Consider a nite BDP with quadratic rates
n
= (N n)(b (n 1)c), n = 0, 1, . . . , N 1,
n
= (N + n + 1)(b + (n + 2)c), n = 1, 2, . . . , N,
(4.4.36)
with b > 0, b/(N + 2) < c < 0. Conditions on the parameters are to ensure the positivity
of the rates.
The KarlinMcGregor polynomials for this process are the Racah polynomials
R
n
_
x
4c
;
1
2
,
1
2
, N 1,
2b + 5c
2c
_
,
which are orthogonal on the set
_
4cn
_
n
2b + 5c
2c
N
_
, n = 0, 1, 2, . . . , N
_
.
Example 4.4.8. Consider a nite BDP with quadratic rates
n
= (N n)(b + (n 1)d), n = 0, 1, . . . , N 1,
n
= n(c (n 1)d), n = 1, 2, . . . , N,
(4.4.37)
where b, c > 0 and
_
d 0, if N = 1 or 2,
b
N2
< d 0, if N > 2.
The conditions on the parameters are to ensure the positivity of the rates.
136 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
The KarlinMcGregor polynomials for this BDP are the dual Hahn polynomials
H
n
_
x
d
;
b 2d
d
,
c Nd
d
, N
_
,
which are orthogonal on the set
_
dn
_
n +
b + c
d
N 1
_
, n = 0, 1, 2, . . . , N
_
.
Exercises 4.4.
4.4.1. Identify the birth and death rates associated with the continued fraction
1
N
_
N1
j=0
1
z+j
=
1
z
+
0
1
+
1
z
+
N
z
, z > 0, and show that P
00
(t) =
1
N
_
N1
j=0
e
jt
.
4.4.2. Identify the birth and death rates associated with the continued fraction
_
N
j=0
b(N, j;p)
z+j
=
1
z
+
Np
1
+
q
z
+
(N1)p
1
+
2q
z
+ +
Nq
z
, (z) > 0
where b(N, j; p) :=
_
N
j
_
p
j
q
Nj
, j = 0, 1, 2, , N , q = 1 p and show that
P
00
(t) =
_
N
j=0
b(N, j; p)e
jt
.
Reference
Parthasarathy, P.R. and Lenin, R.B. (2004). Birth and Death Process (BDP) Models
with Applications Queueing, Communication Systems, Chemical Models, Bio
logical Models: The StateoftheArt with a TimeDependent Perspective, Ameri
can Series in Mathematical and Management Sciences 51, Syracuse, NY, American
Sciences Press, Inc.
4.5. ESTIMATION OF A PARAMETER 137
4.5. Estimation of a Parameter by Rankedset
Sampling
[This section is based on the lectures of P. Yageen Thomas, Department of Statistics, University
of Kerala, Trivandrum 695 581, India.]
4.5.1. Introduction
The concept of rankedset sampling (RSS) was rst introduced by McIntyre
(1952) as a process of improving the precision of the sample mean as an estima
tor of the population mean. rankedset sampling as described in McIntyre (1952)
is applicable whenever ranking of a set of sampling units can be done easily by
a judgement method (for a detailed discussion on the theory and applications of
rankedset sampling see, Chen et al (2004). Ranking by judgement method is not
recommendable if the judgement method is too crude and is not powerful for rank
ing by discriminating the units of a moderately large sample. In certain situations,
one may prefer exact measurement of some easily measurable variable associated
with the study variable rather than ranking the units by a crude judgement method.
Suppose the variable of interest say Y, is dicult or much expensive to measure, but
an auxiliary variable X correlated with Y is readily measurable and can be ordered
exactly. In this case as an alternative to McIntyre (1952) method of rankedset sam
pling, Stokes (1977) used an auxiliary variable for the ranking of the sampling units.
The procedure of rankedset sampling described by Stokes (1977) using auxiliary
variate is as follows: Choose n
2
independent units, arrange them randomly into n
sets each with n units and observe the value of the auxiliary variable X on each of
these units. In the rst set, that unit for which the measurement on the auxiliary
variable is the smallest is chosen. In the second set, that unit for which the mea
surement on the auxiliary variable is the second smallest is chosen. The procedure
is repeated until in the last set, that unit for which the measurement on the auxiliary
variable is the largest is chosen. The resulting new set of n units chosen by one from
each set as described above is called the RSS dened by Stokes (1977). If X
(r)r
is
the observation measured on the auxiliary variable X from the unit chosen from the
rth set then we write Y
[r]r
to denote the corresponding measurement made on the
study variable Y on this unit, then Y
[r]r
, r = 1, 2, . . . , n form the rankedset sample.
Clearly Y
[r]r
is the concomitant of the rth order statistic arising from the rth sample.
138 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
A striking example for the application of the rankedset sampling as proposed
by Stokes (1977) is given in Bain (1978, p.99), where the study variate Y represents
the oil pollution of sea water and the auxiliary variable X represents the tar deposit
in the nearby sea shore. Clearly collecting sea water sample and measuring the oil
pollution in it is strenuous and expensive. However the prevalence of pollution in
the sea water is much reected by the tar deposit in the surrounding terminal sea
shore. In this example ranking the pollution level of sea water based on the tar de
posit in the sea shore is more natural and scientic than ranking it visually or by
judgement method.
Stokes (1995) has considered the estimation of parameters of locationscale
family of distributions using RSS. Lam et al (1994, 1996) have obtained the BLUEs
of location and scale parameters of exponential distribution and logistic distribu
tion. The Fisher information contained in RSS have been discussed by Chen (2000)
and Chen and Bai (2000). Stokes (1980) has considered the method of estimation
of correlation coecient of bivariate normal distribution by using RSS. Modarres
and Zheng (2004) have considered the problem of estimation of dependence pa
rameter using RSS. Robust estimate of correlation coecient for bivariate normal
distribution have been developed by Zheng and Modarres (2006). Stokes (1977)
has suggested the rankedset sample mean as an estimator for the mean of the study
variate Y, when an auxiliary variable X is used for ranking the sample units, un
der the assumption that (X, Y) follows a bivariate normal distribution. Barnett and
Moore (1997) have improved the estimator of Stokes (1977) by deriving the Best
Linear Unbiased Estimator (BLUE) of the mean of the study variate Y, based on
ranked set sample obtained on the study variate Y.
In this paper we are trying to estimate the mean of the population, under a
situation where in measurement of observations are strenuous and expensive. Bain
(1978, P.99) has proposed an exponential distribution for the study variate Y, the
oil pollution of the sea samples. Thus in this paper we assume a Morgenstern type
bivariate exponential distribution (MTBED) corresponding to a bivariate random
variable (X, Y), where X denotes the auxiliary variable (such as tar deposit in the sea
shore) and Y denotes the study variable (such as the oil pollution in the sea water).
A random variable (X, Y) follows MTBED if its probability density function (pdf)
is given by (see, Kotz et al, 2000, P.353)
4.5. ESTIMATION OF A PARAMETER 139
f (x, y) =
_
_
1
2
exp
_
2
_ _
1 +
_
1 2 exp
_
1
__ _
1 2 exp
_
2
___
,
x > 0, y > 0; 1 1;
1
> 0,
2
> 0
0, otherwise.
(4.5.1)
In Section 4.5 of this note we have derived an estimator
2
of the parameter
2
involved in (4.5.1) using the rankedset sample mean. It may be noted that if (X, Y)
has a MTBED as dened in (4.5.1) then the marginal distributions of both X and Y
have exponential distributions and the pdf of Y is given by
f
Y
(y) =
1
2
e
y/
2
, y > 0,
2
> 0. (4.5.2)
The CramerRao Lower Bound (CRLB) of any unbiased estimator of
2
based
on a random sample of size n drawn from (4.5.2) is
2
2
n
. In Section 4.5, we have also
shown that the variance of the proposed estimator
2
is strictly less than the CRLB
2
2
n
(associated with the marginal pdf (4.5.2)) for all A where A = [1, 1] {0}.
In this section, we have further made an eciency comparison between
2
and the
maximumlikelihood estimator (MLE)
2
based on a random sample of size n arising
from(4.5.1). In Section 4.5, we have derived the BLUE
2
of
2
involved in MTBED
based on the rankedset sample and obtained the eciency of
2
relative to
2
. In
Section 4.5, we have considered the situation where we apply censoring and ranking
on each sample and ultimately used a rankedset sample arising out of this procedure
to estimate
2
. In Section 4.5, we have derived the BLUE of
2
involved in (4.5.1)
using unbalanced multistage rankedset sampling method. In this section, we have
further analysed the eciency of the estimator of
2
based on unbalanced MSRSS
when compared with the
2
the MLE of
2
.
4.5.2. rankedset sample mean as an estimator
Let (X, Y) be a bivariate random variable which follows a MTBED with pdf
dened by (4.5.1). Suppose RSS in the sense of stokes (1977) as explained in
Section 4.5 is carried out. Let X
(r)r
be the observation measured on the auxiliary
variate X in the rth unit of the RSS and let Y
[r]r
be the measurement made on the
Y variate of the same unit, r = 1, 2, , n. Then clearly Y
[r]r
is distributed as the
concomitant of rth order statistic of a random sample of n arising from (4.5.1). By
using the expressions for means and variances of concomitants of order statistics
140 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
arising from MTBED obtained by Scaria and Nair (1999), the mean and variance
of Y
[r]r
for 1 r n are given below
E[Y
[r]r
] =
2
_
1
n 2r + 1
2(n + 1)
_
, (4.5.3)
Var[Y
[r]r
] =
2
2
_
1
n 2r + 1
2(n + 1)
2
(n 2r + 1)
2
4(n + 1)
2
_
. (4.5.4)
Since Y
[r]r
and Y
[s]s
for r s are measurements on Y made from two units involved
in two independent samples we have
Cov[Y
[r]r
, Y
[s]s
] = 0, r s. (4.5.5)
In the following theorem we propose an estimator
2
of
2
involved in (4.5.1) and
prove that it is an unbiased estimator of
2
.
Theorem 4.5.1. Let Y
[r]r
, r = 1, 2, , n be the ranked set sample observations
on a study variate Y obtained out of ranking made on an auxiliary variate X, when
(X, Y) follows MTBED as dened in (4.5.1). Then the rankedset sample mean given
by
2
=
1
n
n
r=1
Y
[r]r
,
is an unbiased estimator of
2
and its variance is given by,
Var[
2
] =
2
2
n
_
_
1
2
4n
n
r=1
_
n 2r + 1
n + 1
_
2
_
_
.
Proof 4.5.1.
E[
2
] =
1
n
n
r=1
E[Y
[r]r
].
On using (4.5.3) for E[Y
[r]r
] in the above equation we get,
E[
2
] =
1
n
n
r=1
[1
n 2r + 1
2(n + 1)
]
2
(4.5.6)
It is clear to note that
n
r=1
(n 2r + 1) = 0. (4.5.7)
4.5. ESTIMATION OF A PARAMETER 141
Applying (4.5.7) in (4.5.6) we get,
E[
2
] =
2
.
Thus
2
is an unbiased estimtor of
2
. The variance of
2
is given by
Var[
2
] =
1
n
2
n
r=1
Var(Y
[r]r
).
Now using (4.5.4) and (4.5.7) in the above sum we get,
Var[
2
] =
2
2
n
_
_
1
2
4n
n
r=1
_
n 2r + 1
n + 1
_
2
_
_
.
Thus the theorem is proved.
Now we compare the variance of
2
with the CRLB
2
2
/n of any unbiased esti
mator of
2
involved in (4.5.2) which is the marginal distribution of Y in (4.5.1). If
we write e
1
(
2
) to denote the ratio of
2
2
/n with Var(
2
) then we have,
e
1
(
2
) =
1
_
1
2
4n
_
n
r=1
_
n2r+1
n+1
_
2
_. (4.5.8)
It is very trivial to note that
e
1
(
2
) 1.
Thus we conclude that there is some gain in eciency on the estimator
2
due to
rankedset sampling. The reason for the above conclusion is that a rankedset sam
ple always provides more information than simple random sample even if ranking is
imperfect (see, Chen et al, 2004, P. 58). It is to be noted that Var(
2
) is a decreasing
function of
2
and hence the gain in eciency of the estimator
2
increases as 
increases.
Again on simplifying (4.5.8) we get
e
1
(
2
) =
1
1
2
4
[
2
3
(
2+1/n
1+1/n
) 1]
.
142 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
Then
lim
n
e
1
(
2
) = lim
n
1
1
2
4
[
2
3
(
2+1/n
1+1/n
) 1]
=
1
1
2
12
.
From the above relation it is clear that the maximum value for e
1
(
2
) is attained
when  = 1 and in this case e
1
(
2
) tends to 12/11.
Next we obtain the eciency of
2
by comparing the variance of
2
with the
asymptotic variance of MLE of
2
involved in MTBED. If (X, Y) follows a MTBED
with pdf dened by (4.5.1), then,
log f (x, y)
1
=
1
1
_
_
1 +
x
1
+
2x
1
e
1
(2e
2
1)
1 + (2e
1
1)(2e
2
1)
_
_
and
log f (x, y)
2
=
1
2
_
_
1 +
y
2
+
2y
2
e
2
(2e
1
1)
1 + (2e
1
1)(2e
2
1)
_
_
.
Then we have,
I
1
() = E
_
log f (x, y)
1
_
2
,
=
1
2
1
_
1 + 4
2
_
0
_
0
u
2
e
3u
(2e
v
1)e
v
{1 + (2e
u
1)(2e
v
1)}
dvdu
_
.
I
2
() = E
_
log f (x, y)
2
_
2
,
=
1
2
2
_
1 + 4
2
_
0
_
0
v
2
e
3v
(2e
u
1)e
u
{1 + (2e
u
1)(2e
v
1)}
dvdu
_
and
I
2
() = E
_
2
log f (x, y)
2
_
,
=
1
2
_
4
_
0
_
0
uve
2u
e
2v
{1 + (2e
u
1)(2e
v
1)}
dvdu
_
.
4.5. ESTIMATION OF A PARAMETER 143
Thus the Fisher information matrix associated with the random variable (X, Y) is
given by
I() =
_
I
1
() I
2
()
I
2
() I
2
()
_
. (4.5.9)
We have evaluated the values of
2
1
I
1
() and
1
1
1
2
I
2
() numerically for
= 0.25, 0.50, 0.75, 1 (clearly
2
1
I
1
() =
2
2
I
2
()) and are given below.
Table 4.5.1
2
1
I
1
()
1
1
1
2
I
2
()
2
1
I
1
()
1
1
1
2
I
2
()
0.25 1.0062 0.0625 0.25 1.0062 0.0628
0.50 1.0254 0.1258 0.50 1.0254 0.1274
0.75 1.0596 0.1914 0.75 1.0596 0.1955
1.00 1.1148 0.2624 1.00 1.1148 0.2712
Thus from (4.5.9) the asymptotic variance of the MLE
2
of
2
involved in
MTBED based on a bivariate sample of size n is obtained as
Var(
2
) =
1
n
I
(1)
2
(), (4.5.10)
where I
(1)
2
() is the (2, 2)th element of the inverse of I() given by (4.5.9).
We have obtained the eciency e(
2

2
) =
Var(
2
)
Var(
2
)
of
2
relative to
2
for n =
2(2)10(5)20; = 0.25, 0.50, 0.75, 1 and are presented in Table 4.5.1. From
the table, one can easily see that
2
is more ecient than
2
and eciency increases
with n and  for n 4.
4.5.3. Best linear unbiased estimator
In this section we provide a better estimator of
2
than that of
2
by deriving
the BLUE
2
of
2
provided the parameter is known. Let X
(r)r
be the observation
measured on the auxiliary variate X in the rth unit of the RSS and let Y
[r]r
be the
measurement made on the Y variate of the same unit, r = 1, 2, , n. Let
r
= 1
n 2r + 1
2(n + 1)
(4.5.11)
144 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
and
r
= 1
n 2r + 1
2(n + 1)
2
(n 2r + 1)
2
4(n + 1)
2
. (4.5.12)
Using (4.5.11) in (4.5.3) and (4.5.12) in (4.5.4) we get E[Y
[r]r
] =
2
r
, 1 r n,
and Var[Y
[r]r
] =
2
2
r
, 1 r n. Clearly from (4.5.5), we have Cov[Y
[r]r
, Y
[s]s
] =
0, r, s = 1, 2, . . . , n and r s. Let Y
[n]
= (Y
[1]1
, Y
[2]2
, , Y
[n]n
)
2
of
2
is obtained as
2
= (
G
1
)
1
G
1
Y
[n]
(4.5.13)
and
Var(
2
) = (
G
1
)
1
2
2
, (4.5.14)
where = (
1
,
2
, ,
n
)
and G = diag(
1
,
2
, ,
n
). On substituting the values
of and G in (4.5.13) and (4.5.14) and simplifying we get
2
=
_
n
r=1
(
r
/
r
)Y
[r]r
_
n
r=1
2
r
/
r
(4.5.15)
and
Var(
2
) =
1
_
n
r=1
2
r
/
r
2
2
.
Thus
2
as given in (4.5.15) can be explicitly written as
2
=
_
n
r=1
a
r
Y
[r]r
, where
a
r
=
r
/
r
_
n
r=1
2
r
/
r
, r = 1, 2, . . . , n.
Remark 4.5.1. As the association parameter in (4.5.1) is involved in the BLUE
2
of
2
and its variance, an assumption that is known may sometimes viewed
as unrealistic. Hence when is unknown, our recommendation is to compute the
sample correlation coecient q of the observations (X
(r)r
, Y
[r]r
), r = 1, 2, . . . and
consider the model (4.5.1) for a value of equal to
0
given by
0
=
_
_
1, if q <
1
4
4q, if
1
4
q
1
4
1, if q >
1
4
,
4.5. ESTIMATION OF A PARAMETER 145
as we know that the correlation coecient between X and Y involved in the Mor
genstern type bivariate exponential random vector is equal to
4
.
We have computed the the ratio e(
2

2
)=
Var(
2
)
Var(
2
)
as the eciency of
2
relative to
2
for = 0.25, 0.5, 0.75, 1.0 and n = 2(2)10(5)20 and are also given in Table
4.5.1. From the table, one can easily see that
2
is relatively more ecient than
2
.
Further we observe from the table that e(
2

2
) increases as n and  increase.
4.5.4. Estimation based on censored rankedset sample
In the case of the example of pollution study on sea samples (see, Bain, 1978,
P.99), sometimes if there is no tar deposit at the seashore then the correspondingly
located sea sample will be censored and hence on these units the observations on
Y is not measured. For ranking on X observations in a sample, the censored units
are assumed to have distinct and consecutive lower ranks and the remaining units
are ranked with the next higher ranks in a natural order. If in this censored scheme
of ranked set sampling, k units are censored, then we may represent the rankedset
sample observations on the study variate Y as
1
Y
[1]1
,
2
Y
[2]2
, . . . ,
n
Y
[n]n
where,
i
=
_
_
0, if the i
th
unit is censored
1, otherwise
and hence
_
n
i=1
i
= n k. In this case the usual ranked set sample mean is equal
to
n
i=1
i
Y
[i]i
nk
. It may be noted that
i
= 0 need not occur in a natural order for
i = 1, 2, . . . n. Hence if we write m
i
, i = 1, 2, . . . n k as the integers such that
1 m
1
< m
2
< < m
nk
n and for which
m
i
= 1, then,
E
__
n
i=1
i
Y
[i]i
n k
_
=
2
_
_
1
2(n + 1)(n k)
nk
i=1
(n 2m
i
+ 1)
_
_
.
Thus it is clear that the rankedset sample mean in the censored case is not
an unbiased estimator of the population mean
2
. However we can construct an
unbiased estimator of
2
based on this mean. In the following theorem we have
given the constructed unbiased estimator
2
(k) of
2
based on the rankedset sample
mean under censored situation and its variance.
146 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
Theorem 4.5.2. Suppose that the random variable (X, Y) has a MTBED as de
ned in (4.5.1). Let Y
[m
i
]m
i
, i = 1, 2, . . . n k be the rankedset sample observations
on the study variate Y resulting out of censoring and ranking applied on the auxil
iary variable X. Then an unbiased estimator of
2
based on the rankedset sample
mean
1
nk
_
nk
i=1
Y
[m
i
]m
i
is given by,
2
(k) =
2(n + 1)
_
2(n + 1)(n k)
_
nk
i=1
(n 2m
i
+ 1)
_
nk
i=1
Y
[m
i
]m
i
and its variance is given by,
Var[
2
(k)] =
4(n + 1)
2
2
2
_
2(n + 1)(n k)
_
nk
i=1
(n 2m
i
+ 1)
_
2
nk
i=1
m
i
,
where
m
i
is as dened in (4.5.12).
Proof 4.5.2.
E[
2
(k)] =
2(n + 1)
_
2(n + 1)(n k)
_
nk
i=1
(n 2m
i
+ 1)
_
nk
i=1
E[Y
[m
i
]m
i
],
=
2(n + 1)
_
2(n + 1)(n k)
_
nk
i=1
(n 2m
i
+ 1)
_
nk
i=1
_
1
n 2m
i
+ 1
2(n + 1)
_
2
,
=
2(n + 1)
_
2(n + 1)(n k)
_
nk
i=1
(n 2m
i
+ 1)
_
_
_
n k
2(n + 1)
nk
i=1
(n 2m
i
+ 1)
_
2
,
=
2
.
Thus
2
(k) is an unbiased estimator of
2
. The variance of
2
(k) is given by,
Var[
2
(k)] =
4(n + 1)
2
_
2(n + 1)(n k)
_
nk
i=1
(n 2m
i
+ 1)
_
2
nk
i=1
Var(Y
[m
i
]m
i
),
=
4(n + 1)
2
2
2
_
2(n + 1)(n k)
_
nk
i=1
(n 2m
i
+ 1)
_
2
nk
i=1
m
i
,
where
m
i
is as dened in (4.5.12). Hence the theorem.
4.5. ESTIMATION OF A PARAMETER 147
As a competitor of the estimator
2
(k), we nowpropose the BLUE of
2
based on
the censored rankedset sample, resulting out of ranking of observations on X. Let
Y
[n]
(k) = (Y
[m
1
]m
1
, Y
[m
2
]m
2
, . . . , Y
[m
nk
]m
nk
)
, G(k) = diag(
m
1
,
m
2
, . . . ,
m
nk
).
If the parameter involved in (k) and G(k) is known then (4.5.16) and (4.5.17)
together denes a generalized GuassMarkov set up and hence the BLUE
2
(k) of
2
is obtained as,
2
(k) = [((k))
(G(k))
1
(k)]
1
((k))
(G(k))
1
Y
[n]
(k) (4.5.18)
and
Var(
2
(k)) = [((k))
(G(k))
1
(k)]
1
2
2
. (4.5.19)
On substituting the values of (k) and G(k) in (4.5.18) and (4.5.19) and simplifying
we get
2
(k) =
_
nk
i=1
(
m
i
/
m
i
)Y
[m
i
]m
i
_
nk
i=1
2
m
i
/
m
i
(4.5.20)
and
Var(
2
(k)) =
1
_
nk
i=1
2
m
i
/
m
i
2
2
.
Remark 4.5.2. In the expression for the BLUEs
2
given in (4.5.15) and
2
(k)
given in (4.5.20) the quantities
r
and
r
for 1 r n are all nonnegative, and con
sequently the coecients of rankedset sample observations are also nonnegative.
Thus the BLUEs
2
and
2
(k) of
2
are always nonnegative. Thus unlike in cer
tain situation of BLUEs where one encounters with inadmissible estimators, the
estimators given by
2
and
2
(k) using ranked set sample are admissible estimators.
Remark 4.5.3. Since both the BLUE
2
(k) and the unbiased estimator
2
(k) based
on the censored rankedset sample utilize the distributional property of the parent
distribution they lose the usual robustness property. Hence in this case the BLUE
2
(k) shall be considered as a more preferable estimators than
2
(k).
148 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
4.5.5. Unbalanced multistage rankedset sampling
AlSaleh and AlKadiri (2000) have extended rst the usual concept of RSS
to double stage rankedset sampling (DSRSS) with an objective of increasing the
precision of certain estimators of the population when compared with those ob
tained based on usual RSS or using random sampling. AlSaleh and AlOmari
(2002) have further extended DSRSS to multistage rankedset sampling (MSRSS)
and shown that there is increase in the precision of estimators obtained based on
MSRSS when compared with those based on usual RSS and DSRSS. The MSRSS
(in r stages)procedure is described as follows:
(1) Randomly select n
r+1
sample units from the target population, where r is the
number of stages of MSRSS.
(2) Allocate the n
r+1
selected units randomly into n
r1
sets, each of size n
2
.
(3) For each set in step (2), apply the procedure of rankedset sampling method
to obtain a (judgment) rankedset, of size n; this step yields n
r1
(judgment)
rankedsets, of size n each.
(4) Arrange n
r1
rankedsets of size n each, into n
r2
sets of n
2
units each and with
out doing any actual quantication, apply rankedset sampling method on each
set to yield n
r2
second stage rankedsets of size n each.
(5) This process is continued, without any actual quantication, until we end up
with the rth stage (judgment) rankedset of size n.
(6) Finally, the n identied elements in step (5) are now quantied for the variable
of interest.
Instead of judgment method of ranking at each stage if there exists an auxiliary
variate on which one can make measurement very easily and exactly and if the aux
iliary variate is highly correlated with the variable of interest, then we can apply
ranking based on these measurements to obtain the rankedset units at each stage
of MSRSS. Then on the nally selected units one can make measurement on the
variable of primary interest. In this section we deal with the MSRSS by assuming
that the random variable (X, Y) has a MTBED as dened in (4.5.1), where Y is the
variable of primary interest and X is an auxiliary variable. In Section 4.5, we have
4.5. ESTIMATION OF A PARAMETER 149
considered a method for estimating
2
using the Y
[r]r
measured on the study variate
Y on the the unit having rth smallest value observed on the auxiliary variable X, of
the rth sample r = 1, 2, . . . , n and hence the RSS considered there was balanced.
AboEleneen and Nagaraja (2002) have shown that in a bivariate sample of size n
arising from MTBED the concomitant of largest order statistic possesses the max
imum Fisher information on
2
whenever > 0 and the concomitant of smallest
order statistic possesses the maximum Fisher information on
2
whenever < 0.
Hence in this section, rst we consider > 0 and carry out an unbalanced MSRSS
with the help of measurements made on an auxiliary variate to choose the rankedset
and then estimate
2
involved in MTBED based on the measurement made on the
variable of primary interest. At each stage and from each set we choose an unit of
a sample with the largest value on the auxiliary variable as the units of rankedsets
with an objective of exploiting the maximum Fisher information on the ultimately
chosen rankedset sample.
Let U
(r)
i
, i = 1, 2, . . . , n be the units chosen by the (r stage) MSRSS. Since
the measurement of auxiliary variable on each unit U
(r)
i
has the largest value, we
may write Y
(r)
[n]i
to denote the value measured on the variable of primary interest on
U
(r)
i
, i = 1, 2, . . . , n. Then it is easy to see that each Y
(r)
[n]i
is the concomitant of the
largest order statistic of n
r
independently and identically distributed bivariate ran
dom variables with MTBED. Moreover Y
(r)
[n]i
, i = 1, 2, . . . , n are also independently
distributed with pdf given by (see, Scaria and Nair, 1999)
f
(r)
[n]i
(y; ) =
1
2
e
2
_
1 +
_
n
r
1
n
r
+ 1
_
_
1 2e
2
_
_
. (4.5.21)
Thus the mean and variance of Y
(r)
[n]i
for i = 1, 2, . . . , n are given below.
E[Y
(r)
[n]i
] =
2
_
1 +
2
_
n
r
1
n
r
+ 1
__
, (4.5.22)
Var[Y
(r)
[n]i
] =
2
2
_
1 +
2
_
n
r
1
n
r
+ 1
_
2
4
_
n
r
1
n
r
+ 1
__
. (4.5.23)
If we denote
n
r = 1 +
2
_
n
r
1
n
r
+ 1
_
(4.5.24)
and
n
r = 1 +
2
_
n
r
1
n
r
+ 1
_
2
4
_
n
r
1
n
r
+ 1
_
(4.5.25)
150 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
then (4.5.22) and (4.5.23) can be written as
E[Y
(r)
[n]i
] =
2
n
r (4.5.26)
and
Var[Y
(r)
[n]i
] =
2
2
n
r . (4.5.27)
Let Y
(r)
[n]
= (Y
(r)
[n]1
, Y
(r)
[n]2
, , Y
(r)
[n]n
)
n
r 1 (4.5.28)
and
D[Y
(r)
[n]
] =
2
2
n
r I, (4.5.29)
where 1 is the column vector of n ones and I is an identity matrix of order n. If
> 0 involved in
n
r and
n
r is known then (4.5.28) and (4.5.29) together dene a
generalized GaussMarkov setup and hence the BLUE of
2
is obtained as
n(r)
2
=
1
n
n
r
n
i=1
Y
(r)
[n]i
, (4.5.30)
with variance given by
Var(
n(r)
2
) =
n
r
n(
n
r )
2
2
2
. (4.5.31)
If we take r = 1 in the MSRSS method described above, then we get the usual single
stage unbalanced RSS. Then the BLUE
n(1)
2
of
2
is given by
n(1)
2
=
1
n
n
n
i=1
Y
[n]i
,
with variance
Var(
n(1)
2
) =
n
n(
n
)
2
2
2
,
where we write Y
[n]r
instead of Y
(1)
[n]r
and it represent the measurement on the variable
of primary interest of the unit selected in the RSS. Also
n
and
n
are obtained by
putting r = 1 in (4.5.24) and (4.5.25) respectively.
We have evaluated the ratio e(
n(1)
2

2
) =
Var(
2
)
Var(
n(1)
2
)
for = 0.25(0.25)1; n =
2(2)10(5)20 as a measure of eciency of our estimator
n(1)
2
relative to the MLE
2
of
2
based on n observations and are also provided in Table 4.5.2. From the ta
ble, one can see that the eciency increases with increase in and n. Moreover the
4.5. ESTIMATION OF A PARAMETER 151
eciency of the estimator
n(1)
2
is larger than the estimator
2
based on RSS mean
and the BLUE
2
based on usual RSS.
AlSaleh (2004) has considered the steadystate RSS by letting r to +. If
we apply the steadystate RSS to the problem considered in this paper then the
asymptotic distribution of Y
(r)
[n]i
is given by the pdf
f
()
[n]i
(y; ) =
1
2
e
2
_
1 +
_
1 2e
2
__
. (4.5.32)
Fromthe denition of our unbalanced MSRSS it follows that Y
()
[n]i
, i = 1, 2, . . . , n are
independently and identically distributed random variables each with pdf as dened
in (4.5.32). Then Y
()
[n]i
, i = 1, 2, . . . , n may be regarded as unbalanced steadystate
rankedset sample of size n. Then the mean and variance of Y
()
[n]i
for i = 1, 2, . . . , n
are given below.
E[Y
()
[n]i
] =
2
_
1 +
2
_
and
Var[Y
()
[n]i
] =
2
2
_
1 +
2
2
4
_
.
Let Y
()
[n]
= (Y
()
[n]1
, Y
()
[n]2
, , Y
()
[n]n
)
n()
2
based on Y
()
[n]
and the
variance of
n()
2
is obtained by taking the limit as r in (4.5.30) and (4.5.31)
respectively and are given by
n()
2
=
1
n
_
1 +
2
_
n
i=1
Y
()
[n]i
,
and
Var(
n()
2
) =
_
1 +
2
2
4
_
n
_
1 +
2
_
2
2
2
. (4.5.33)
152 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
From (4.5.10) and (4.5.33) we get the eciency of
n()
2
relative to
2
by taking the
ratio of Var(
2
) with Var(
n()
2
) and is given by
e(
n()
2

2
) =
Var(
2
)
Var(
n()
2
)
,
=
I
(1)
2
()
_
1 +
2
_
2
_
1 +
2
2
4
_ .
Thus the eciency e(
n()
2

2
) is free of the sample size n. That is, for a xed
, e(
n()
2

2
) is a constant for all n. We have evaluated the value e(
n()
2

2
) for
= 0.25(0.25)1 and are presented in Table 4.5.3. From the table one can see
that the eciency of
n()
2
increases as increases. Moreover, the estimator
n()
2
possesses the highest eciency amoung the other estimators of
2
proposed in this
paper and the value of the eciencies ranges from 1.1382 to 1.7093.
As mentioned earlier, for MTBED the concomitant of smallest order statistic
possesses the maximum Fisher information on
2
whenever < 0. Therefore when
< 0 we consider an unbalanced MSRSS in which at each stage and from each set
we choose an unit of a sample with the smallest value on the auxiliary variable as
the units of rankedsets with an objective of exploiting the maximum Fisher infor
mation on the ultimately chosen rankedset sample.
Let Y
(r)
[1]i
, i = 1, 2, . . . , n be the value measured on the variable of primary interest
on the units selected at the rth stage of the unbalanced MSRSS. Then it is easy to see
that each Y
(r)
[1]i
is the concomitant of the smallest order statistic of n
r
, independently
and identically distributed bivariate random variables with MTBED. Moreover Y
(r)
[1]i
,
i = 1, 2, . . . , n are also independently distributed with pdf given by
f
(r)
[1]i
(y; ) =
1
2
e
2
_
1
_
n
r
1
n
r
+ 1
_
_
1 2e
2
_
_
(4.5.34)
Clearly from (4.5.21) and (4.5.34) we have
f
(r)
[1]i
(y; ) = f
(r)
[n]i
(y; ). (4.5.35)
4.5. ESTIMATION OF A PARAMETER 153
and hence E(Y
(r)
[n]i
) for > 0 and E(Y
(r)
[1]i
) for < 0 are identically equal. Similarly
Var(Y
(r)
[n]i
) for > 0 and Var(Y
(r)
[1]i
) for < 0 are identically equal. Consequently if
1(r)
2
is the BLUE of
2
, involved in MTBED for < 0, based on the unbalanced
MSRSS observations Y
(r)
[1]i
, i = 1, 2, . . . , n, then the coecients of Y
(r)
[1]i
, i = 1, 2, . . . , n
in the BLUE
1(r)
2
for < 0 are same as the coecients of Y
(r)
[n]i
, i = 1, 2, . . . , n in the
BLUE
n(r)
2
for > 0. Further, we have Var(
1(r)
2
) = Var(
n(r)
2
) and hence Var(
1(1)
2
) =
Var(
n(1)
2
) and Var(
1()
2
) = Var(
n()
2
), where
1(1)
2
is the BLUE of
2
, for < 0
based on the usual unbalanced single stage RSS observations Y
[1]i
, i = 1, 2, . . . , n
and
1()
2
is the BLUE of
2
, for < 0 based on the unbalanced steadystate RSS
observations Y
()
[1]i
, i = 1, 2, . . . , n. We have obtained the eciency e(
1(1)
2

2
) of
the BLUE
1(r)
2
relative to
2
, the MLE of
2
for = 0.25, 0.5, 0.75, 1; n =
2(2)10(5)20 and are incorporated in Table 4.5.2. Similarly as in the case of
n()
2
for
a xed the eciency e(
1()
2

2
) is same for all n. We have evaluated e(
1()
2

2
) for
= 0.25, 0.5, 0.75, 1 and are incorporated in Table 4.5.3. From the table, one
can see that eciency increases as  increases and the value of eciency ranges
from 1.1382 to 1.7161.
Remark 4.5.4. If (X, Y) follows an MTBED with pdf dened in (4.5.1), then the
correlation coecient between X and Y is given by,
Corr(X, Y) =
4
, 1 1.
Clearly when  goes to 1, correlation coecient between X and Y is high. Thus,
using the ranks of Y to induce the ranks of X becomes more accurate. Thus when
 is large (that is tends to 1) we see that the rankedset sample obtained based
on the ranking made on X becomes more informative for making inference on
2
than the case with small values of . From the table we notice that for a given
sample size the eciencies of all estimators increase as  increases. Consequently
we note that more information on
2
can be extracted from the rankedset sample
when  is large subject to  1. Thus we conclude that concomitant ranking is
more eective in estimating
2
when the absolute value of the association parameter
is large (that is when tends to 1).
154 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
Table 4.5.2
Eciencies of the estimators
2
,
2
,
n(1)
2
and
1(1)
2
relative to
2
of
2
involved in Morgenstern type bivariate exponential distribution
n e(
2

2
) e(
2

2
) e(
n(1)
2

2
) e(
2

2
) e(
2

2
) e(
1(1)
2

2
)
0.25 0.9994 0.9994 1.0410 0.25 0.9994 0.9994 1.0410
2 0.50 0.9970 0.9970 1.0795 0.50 0.9974 0.9974 1.0800
0.75 0.9911 0.9911 1.1130 0.75 0.9926 0.9926 1.1147
1.00 0.9767 0.9768 1.1349 1.00 0.9806 0.9807 1.1394
0.25 1.0008 1.0008 1.0782 0.25 1.0008 1.0008 1.0782
4 0.50 1.0026 1.0027 1.1613 0.50 1.0030 1.0031 1.1618
0.75 1.0038 1.0044 1.2466 0.75 1.0054 1.0060 1.2485
1.00 0.9996 1.0019 1.3263 1.00 1.0036 1.0059 1.3316
0.25 1.0014 1.0014 1.0948 0.25 1.0014 1.0014 1.0948
6 0.50 1.0051 1.0052 1.1994 0.50 1.0055 1.0056 1.1998
0.75 1.0094 1.0105 1.3111 0.75 1.0109 1.0120 1.3131
1.00 1.0097 1.0140 1.4224 1.00 1.0137 1.0180 1.4281
0.25 1.0018 1.0018 1.1042 0.25 1.0018 1.0018 1.1042
8 0.50 1.0064 1.0066 1.2213 0.50 1.0068 1.0070 1.2218
0.75 1.0125 1.0139 1.3490 0.75 1.0141 1.0154 1.3511
1.00 1.0154 1.0211 1.4800 1.00 1.0195 1.0252 1.4860
0.25 1.0020 1.0020 1.1103 0.25 1.0020 1.0020 1.1103
10 0.50 1.0073 1.0075 1.2355 0.50 1.0077 1.0079 1.2360
0.75 1.0145 1.0161 1.3739 0.75 1.0161 1.0176 1.3760
1.00 1.0191 1.0259 1.5184 1.00 1.0232 1.0300 1.5245
0.25 1.0023 1.0023 1.1189 0.25 1.0023 1.0023 1.1189
15 0.50 1.0085 1.0088 1.2560 0.50 1.0089 1.0092 1.2565
0.75 1.0173 1.0192 1.4100 0.75 1.0189 1.0208 1.4122
1.00 1.0243 1.0328 1.5747 1.00 1.0284 1.0370 1.5810
0.25 1.0024 1.0024 1.1234 0.25 1.0024 1.0024 1.1234
20 0.50 1.0091 1.0095 1.2669 0.50 1.0095 1.0099 1.2674
0.75 1.0188 1.0209 1.4295 0.75 1.0204 1.0225 1.4317
1.00 1.0270 1.0366 1.6054 1.00 1.0311 1.0408 1.6118
4.5. ESTIMATION OF A PARAMETER 155
Table 4.5.3
Eciencies of the estimators
n()
2
and
1()
2
relative to
2
e(
n()
2

2
) e(
1()
2

2
)
0.25 1.1382 0.25 1.1382
0.50 1.3028 0.50 1.3033
0.75 1.4943 0.75 1.4960
1.00 1.7093 1.00 1.7161
References
AboEleneen, Z.A. and Nagaraja, H.N. (2002). Fisher information in an order statis
tic and its concomitant. Annals of Institute of Statistical Mathematics, 54, 667680.
AlSaleh, M.F. (2004). Steadystate ranked set sampling and parametric inference.
Journal of Statistical Planning and Inference, 123, 8395.
AlSaleh, M.F. and AlKadiri, M. (2000). Double ranked set sampling. Statistics
and Probability Letters, 48, 205212.
AlSaleh, M.F. and AlOmari, A. (2002). Multistage ranked set sampling. Journal
of Statistical Planning and Inference, 102, 273286.
Bain, L.J. (1978). Statistical Analysis of Reliability and Life Testing Models: The
ory and Methods. Marcel Dekker, New York.
Barnett, V. and Moore, K. (1997). Best linear unbiased estimates in rankedset sam
pling with particular reference to imperfect ordering. Journal of Applied Statistics,
24, 697710.
Chen, Z. (2000). The eciency of rankedset sampling relative to simple random
sampling under multiparameter families. Statistica Sinica, 10, 247263.
Chen, Z. and Bai, Z. (2000). The optimal rankedset sampling scheme for paramet
ric families. Sankhya Series A, 46, 178192.
Chen, Z., Bai, Z. and Sinha, B. K. (2004). Lecture Notes in Statistics, RankedSet
Sampling, Theory and Applications. SpringerVerlag, New York.
156 4. BIRTH AND DEATH PROCESSES AND ORDER STATISTICS
David, H.A. and Nagaraja, H.N (2003). Order Statistics, Wiley, New York.
Kotz, S., Balakrishnan, N. and Johnson, N.L. (2000). Distributions in Statistics:
Continuous Multivariate Distributions, Wiley, New York.
Lam, K., Sinha, B.K. and Wu, Z. (1994). Estimation of a twoparameter expo
nential distribution using rankedset sample. Annals of the Institute of Statistical
Mathematics, 46, 723736.
Lam, K., Sinha, B.K. and Wu, Z. (1996). Estimation of location and scale parame
ters of a logistic distribution using rankedset sample. In H.N. Nagaraja, P. K. Sen
and D.F. Morrison(Ed.) Statistical Theory and Applications: Papers in Honor of
Herbert A. David. NewYork: Springer.
McIntyre, G.A. (1952). A method for unbiased selective sampling, using ranked
sets. Australian Journal of Agricultural Research, 3, 385390.
Modarres, R. and Zheng, G. (2004). Maximumlikelihood estimation of dependence
parameter using rankedset sampling. Statistics and Probability Letters, 68, 315
323.
Scaria, J. and Nair, N.U. (1999). On concomitants of order statistics from Morgen
stern family. Biometrical Journal, 41, 483489.
Stokes, S. L. (1977). rankedset sampling with concomitant variables. Communi
cations in Statistics; Theory and Methods, 6, 12071211.
Stokes, S. L. (1980). Inference on the correlation coecient in bivariate normal
populations from rankedset Samples. Journal of the American Statistical Associa
tion, 75, 989995.
Stokes, S. L. (1995). Parametric rankedset sampling. Annals of the Institute of
Statistical Mathematics, 47, 465482.
Zheng, G. and Modarres, R. (2006). A robust estimate of correlation coecient
for bivariate normal distribution using rankedset sampling. Journal of Statistical
Planning and Inference, 136, 298309.