Professional Documents
Culture Documents
DISTRIBUTIONS
Structure
11.1 Introduction
Objectives
11.2 Normal Distribution
11.3 ~ x ~ o n e n t iand
a l Gamma Distributions
11.4 Beta Distribution
11.5 Summary
11.6 Solutions and Answers
Appendix : Tables of the Normal Distribution
1 1 . INTRODUCTION
In the previous unit, we have introduced the notions of distribution function, density
function, expectation, variance and moments for a univariate continuous random
variable. You have seen several examples dealing with these concepts. In this unit,
we study the properties of some standard (absolutely) continuous distributions.
These distributions are widely used in statistical inference as you will see in Block 4.
Our main emphasis here is on normal, exponential, gamma and beta distribution. We
have chosen these distributions because of their wide spread applicability. You have
already met some of these in Unit 10. Here we shall take them one by one and
discuss them in detail.
Objectives
After reading this unit, you should be able to :
compute the mean and variance of the normal, exponential, gamma and beta
distributions ;
investigate properties of other distributions which you come across.
Let us start with the normal distribution.
Non-negativity off is obvious from its definition. Let us check the second condition :
= I 1
2
,(by the transformation y = -x-CL
)
- m G e - Y "!dy a
Now, to evaluate this integral, we start by writing
Then
Apply the transformation y = r cos 8 and z = r sine. Then, from the change of
variable formula for double integrals (see Unit 11, Block 4, MTE -07), if follows that
Hence I = 1.
m
You know that if p = 0 and a = 1, then the distribution is called standard normal
distribution. We usually denote the standard normal distribution by @ (x) and the
Distriiutim standard normal density by rp (x). In Fig. 1 you can see graphs of the p.d.f. of the
standard normal distribution and that of an arbitrary normal distribution.
Fig. 1 : The grnph of the p.d.f. of a (a) norn~aldistribution (b) standard normal distribution.
From Fig. l(a) you can see that the density function f (x; p, 2)is symmetric about
p. Further p is the median as well a s the mode of the distribution. The graph of f is
bell shaped. The shaded area in Fig. l(b) gives the standard normal distribution
function. Now let us calculate the expectation, variance and moments for a normal
distribution.
Example 1: Let us calculate the expected value and the variance of the r.v. X,
whose density function, f, is given by
-
Haven't you seen a similar density function before? In Unit 10, Example 9 you have
seen a particular case of this with p = 0 and a 1.
Now,
Standard Contlnuous
Dlstrlhutlons
Solve this integral by using the method of integration by parts, and check that
var (x)'= d.
Next we shall find the moments.
Example 2 :Suppose X is a random variable with density function
Then
pk = E[ (x - P)~]
=
2 2 "
- J (25)"' 2
e-' (2~)-lnd z , by putting z = 2 In Unit 10.
dzz 0
I
Distribution 'l%e~'Y Now we make use of some more properties of the gamma function.
Hence,
-
yl 0 implies that the normal density is symmetric. And = 3 implies that it is
mesa- kurtic (see Unit 3, Block 1).
An in the previous unit you must have realized that the calculation of moments is a
somewhat tedious process. Can you guess what an easier way is? This is what we
shall do now.
Example 3 : Suppose X has the normal density function with mean p and variance
d.Then
MX (t) -E (eiX)
00
- $ e t ( a ~ + P-
-
-m
)
1
e-iy dy
6
1 2
(by the transformation
u
- y)
The integral on the right hand side is equal to 1 since the integrand can be Standard Continuous
considered as the density function of a normal distribution with mean ta and Distributions
variance 1.
Hence,
Thus in the case of normal distribution, the m.g.f. exists for all real t.
We now state and prove an important theorem. You will realise the usefulness of
this theorem from the examples that follow.
Theorem 1: If a random variable x has normal distribution with mean p and
variance c? ,then the random variable Y = aX + b has normal distribution with mean
a p + b and variance a2 c?.
Proof: Let us compute the m.g.f. or Y. Then
My (I) = E [e'
=E
= e'
e'(aX+b)
[ I
E[et a
= e' Mx (at)
For simplicity, we denote the normal distribution with mean p and variance c? by
N (p, dl.If follows, from Theorem 1 that if Y is N (p, d ), then X = is
a
N (0, 1).
We now give an example to show how Theorem 1 can be used to find the
probabilities relating to an r.v.with normal distribution.
Example 4 : Suppose Y is N(l, 4). Let us compute P[3 < Y c 51.
Y-1
Here p = 1and a = 2. We apply the transformation X = -
2 .
Then we obtain that X is N(O, 1) and
Now
Distribution Theory where @ denotes the standard normal distribution function.
Thus, we have found the required probability relating to the variable Y in terms of
.the probabilities relating to X, which has a standard normal distribution. So, if we
know the probabilities relating to a standard normal distribution, we will be able to
calculate those for any normal distribution. But in general, the exact evaluation of @
(x) is not possible since there is no explicit formula for computing definite integral.
However, extensive tables giving probabilities calculated from the standard normal
distribution are available (see Table 1 in Appendix). We can use these hbles for
computing probabilities connected with a general normal distribution. You may find
the computation easy if you not that
a (- -
x) = 1 cD (x) v x.
This is because by the symmetry of the standard normal density function about zero,
the area to the left of x must be equal to (1 - area of the right of x).
Let us now see how we use Table 1 in the Appendix. From Table 1, check that
(3.00) - 0.9987, and
cD (2.00) = 0.9773
Similarly,
Now, how do we interpret this? We can say that more than 99% of the total
probability is carried by the interval [ - 3 , 3 ] and about 95% of the total probability
is supported by.the interval [ - 2 , 2 ] for a standard normal distribution.
P[p-3a<Y<p+3o]=P[-3<X<3]=0.9974
and P [ p - 20 < Y < p + ~ C T ]= .9546.
Now if you go back to Example 1, themusing Table 1 we get
Now, we'll show how the normal distribution is used in some practical situations.
Example 5 : It has been observed that the marks obtained by the students in an
examination generally follows a normal distribution. In other words the
~--- ~
approximating curve of the histogram obtained from the data should be bellshaped
(see Fig. 2).
This method is referred to as 'grading on the curve'. Standard Continuous
Dlstrlbutlon
The teacher uses the percentage marks to estimate the mean p and the variance 2.
Suppose the letter grade A is assigned to a student if the percentage of marks is
greater than p + a, B to those whose percentage is between p and p + u, C to those
between p - a and p, D to those between p - 2 0 and p - 0 and E to those getting a
percentage below p - 20. Suppose we want to calculate the percentage of students
having grades A, B, C, D and E.
Let X denote the r.v. corresponding to the data and p and udenote the mean and
standard deviation respectively..Then the percentages of students having grades A,
B,C,DandEaregivenbyP[x~p+u],P[pcXcp+a],P[p-u<X<p],
P [p - 2 0 c X < p - a] and P [X < p - 201 respectively. Applying the
transformation Y - -, u
we find that Y has normal distribution, N(0, 1). Then
Similarly we get
-
P [p - u < X < p ] = @ (0) - @ (- 1) 0.3413.
P [p-2a<X<p-0]=@(-2)-@(- 1)=0.1359.
and
P < p - 2 a] a (- 2) = 0.0228.
In other words, approximately 16% will receive grade A, 34% grade B, 34% grade
C, 14% grade D and 2% will fail, since they get grade E.
Distribution Theory Now, try to solve these exercises.
E2) If the m.g.f. of X is Mx (t) = expt (- 6t + 32t2), find P [-4 s X < 161.
E3) Show that if X has a normal distribution with zero mean, so does -X.
E(X) = expjp +
2 and
WL:now consider two more distributions which are l'ound quite useful in
applications. You have ;~lreadymet one of these before in Unit 1 0 (see Example 8).
Let us quickly recall the facts you have already studied in Unit 10 ahout the
exponential distribution.
where h > 0. You know that the distribution function F, corresponding to the density
f is
for all s, t r 0. This follows from the following observation (A I B) denotes the
conditional probability of an
event A given the event B.
P[X>s+tl 1-P[Xss+t]
P[X> s+tlX>t]= -
P [X > t] 1-Ppcst]
-
This means that the only continuous r.v. X assuming non-negative values for which
P [ X > s + t I x > s ] P[X>t], for all s, t r 0, is an exponentially distributed r.v.
Here is how we go about it.
Suppose X is a non-negative random variable such that
1 = 0, x < 0.
I
Distribution Tbeory The probability that the first week is accident free is
-
ES) Calls arrive at a switchboard following an exponential distribution with
parameter A 5 per hour. If we are at the switchboard, what is the probability
that the waiting time for a call is
i) at least 15 minutes,
ii) not more than 10 minutes,
iii) exactly 5 minutes.
We have seen in Unit 8 that the Poisson distribution can be chosen as a model for the
distribution of the number of occurrences of an event during a fixed time interval.
Let us denote by N(t) the number of occurrences of the event (such as a telephone
call, or an accident etc.) in the interval [0, t] and assume that, for t > 0, N(t) has
Poisson distribution with mean At, where A > 0. A denotes the average rate of
occurrence in unit interval. Assume that N(0) = 0.
Let us denote by X the waiting time involved before the first occurrence. It is clear
that
P [X > t] = P[N (t) = 0]
= e-'I, for t > 0.
Therefore,
Thus X has an exponential distribution. These relations show the link between the
exponential and the Poisson distributions. So you know that the exponential
distribution describes the probability distribution of the waiting time for the first
occurrence. Now, suppose we are interested in,the waiting time for r occurrences.
Let us denote by Y the waiting time that elapses before r occurrences of the event.
Then, Standard Continuous
Distributions
P > t] = P[N (t) < r]
r ( a ) -$xa - edXdx
0
(The integral on the right hand side exists for any a > 0.) You have to already come
across some properties of the gamma function in Sec. 11.2 and used these properties
for evaluating certain integrals.
Now with the help of the gamma function we shall define a gamma distribution.
10
, xso
where a > 0 and h > 0 and r is the gamma function. Then X is said to have gamma
distribution with parameters a and L
Check that f(x) satisfies the properties of a density function. Fig. 3(a) and @) show
graphs of some typical gamma density functions.
Distribution Theory
Fig. 3 :Graphs of Gamma Density Functions for different values of (a) a (b) L
Let us compute the m.g.f. of the gamma density function now. By definition,
The last integral is finite if and only if a > 0 and ?L - t > 0. This can be proved by
using some properties of r. But for the time being you will have to take our word for
it. Hence the m.g.f. Mx(t) exists if and only if t < ?L a > 0. Then, fort < h,
Check that
and
In particular
and
Hence Standard C o ~ ~ t i n u o u s
Distributions
The gamma distribution is a good model for waiting times. It has also been used as a
model for the distribution of incomes as the parameters a and A provide a flexibility
in fitting the model to the data. The following exercises will lead to a better
understanding of this distribution.
E9) Let X be a gamma random variable with parameters a and A. Compute E(Xm)
for all m 2 1 directly, and hence derive E(X) and Var(X).
1
E10) Suppose X is a gamma random variable with E(X) = 2 and Var(X) = 7. Find a
and A.
E l l ) Suppose X is N(0, 1). Show that Y = x2has gamma distribution with a = 1/2
and A = 1/2.
In the next section we shall take up one last distribution the beta distribution.
While studying Statistics and many other sciences, scientists found that they came
across some particular functions quite often. They have identified a few such
functions. These functions are called special functions. You have already seen one
special function so far -the gamma function. You have also seen how the gamma
distribution is defined with the help of this function. Now we are going to introduce
another special function, the beta function, B. Then, with the help of B ,we shall
define the beta distribution.
If a > 0 , p > 0, then
1
B(a, p) =$xa-' (I-X)~-' dx
0
,
I
I Let's prove this.
From the definition of the gamma function, we get
Now to evaluate the integral we apply the transformation You will find discussion of
Jacobians and double integals
in Blocks 3 and 4 of MTE-07.
Nore that as and take Hence, from the change of variable formula for double integrals, it follows that
values in [O, m], u varies 1 OD
over [lo, I.] and v varies
over [o,m] r ( a ) r ( $ ) =JJua-' (I-u)P-l va+P-l e-" dvdu
10 , otherwise
where a > 0, $ > 0.
An alternative representation for a beta density function with parameters a and f3 is
r ( a + P) o < x < ~
f(x) = r ( a ) r (p) xa-I (1-x)P-l
I O , otherwise
Let us now compute the moments of a random variable X with the beta distribution
having parameters a and p. For any integer k 2 1.
1
E [ x ~ ]=Jxk f(x) dx
0
i Therefore,
In Fig. 4 you can see some typical graphs of beta density function.
This brings us to the end of this unit. Let us now summarise its main points.
11.5 SUMMARY
In this unit, we have derived the properties of the following distributions with
emphasis on calculation of their distribution functions, density functions, their
means, variances and moment generating functions :
Distribution Theory 1) Normal distribution :
I t X is N(p, *
d),then Y = u is N(0, 1)
2) Exponential distributiom :
The exponential and gamma distributions are also called waiting time distribution.
4) Beta distribution:
1 xa-'(l-x)fi-', O<x<l
otherwise
iii) 0..2711
iv) PI/ X 1 > 1.391 = 1 -PI 1 X 1 s 1.391
E2) From the m.g.f. of X, we get that X is a normal random variable with mean
- 6 and a = 8 i.e. N ( 4 , b4).
To calculate the probability P[-4 5 X 5 161, we apply the transformation
y = -X+6
8
Standard Continuous
DlsMbution
= Mx(t)
-
for w < t < w. :. by Theorem 1 in Unit 10, the distributions of X and -X
are the same. Hence - X is N(0, d).
E4) Let Y --
X- 60
2
Hence
E7) We prove a general result here. The only right continuous solution of the
functional equation
g(s+t) = g(s)g(t), s > 0, t > 0
which is not identically zero for x > 0, is
-
g(s) e-IS, h > 0.
This can be seen in the following way
Since g(s+t) = g(s) g(t), it follows that
But
Therefore
and
or equivalently
In particular
and
Therefore
I Ell) F O ~ ~ > O , P P [ ~ S ~ J - P [ - ~ ~ X ~ ~ ]
6
1 - 4 2 du.
- 20 J z e
Hence the density function of Y is
= 0, y so.
This can be written in the form
and
fy(y) = 0 for y s 0
since l?(1/2) = G.This shows that the density function of Y is gamma
densitv with a = 1/2 and A = 1/2.
Distribution 'Iheory
= 0, otherwise.
which is the beta density with the parameters fl and a .
E14) Here a = 4 and f! = 7. Further
Standard Continuous
Distributions
Appendix
Table 1. Values of the Standard Normal Distribution Function
Table 1. (Contd.)