Professional Documents
Culture Documents
Remark: If E is a sure event then all the N simple events connected with the random
experiment are favorable to E. Thus P (E) = 1. If E is an impossible event then none of the
simple events connected with the random experiment are favorable to E, thus P (E) = 0. If
E is any event connected with the random experiment, then 0 ≤ m ≤ N , thus
0 ≤ P (E) ≤ 1. Also (N − m) simple events are favorable to the occurrence of E. Thus
P (Ē) = 1 − P (E).
12. Theorems on probability:
(i) The probability of a sure event is 1.
(ii) The probability of an impossible event is 0.
(iii) If E is any event connected with a random experiment then 0 ≤ P (E) ≤ 1.
(iv) If E is any event connected with a random experiment then P (E) = 1 − P (E).
13. Counting techniques
(a) Multiplication Principle
Suppose one operation can be performed in m ways and for each way of performing
this operation a second operation can be performed in n ways. Then the total number
of ways of performing these two operations together is mn.
(b) The Addition principle:
Suppose an opearation can be performed either in one of m ways or in one of n ways,
where none of the set of m ways is the same as any of the set of n ways then there are
(m + n) ways of performing this operation.
(c) Factorial
Definition: If n is a positive integer then factorial n is denoted by n! and is defined by
n! = 1.2.3 . . . n. By definition 0! = 1.
(d) Permutation
Any arrangement that can be made out of a given number of objects by taking some
or all of the objects at a time is called a permutation of the objects. If n objects are
given and we form an arrangement of r objects chosen from these n objects, then it is
called a permutation of these n objects taken r at a time. The number of
permutations of n distinct objects taken r at a time is denoted by n Pr .
Theorems on permutation:
n!
i. n Pr = = n.(n − 1).(n − 2) . . . (n − r + 1)
(n − r)!
Note: n Pn = n!, n P1 = n.
ii. The number of permutations of n objects taken r at a time, where p of the objects
are alike of one kind, q of the objects are alike of a second kind, r of the objects
n!
are alike of a third kind and the rest are all different is .
p! q! r!
iii. The number of permutations of n different objects taken r at a time, when each
object may occur any number of times is nr .
(e) Combination
Any group that can be made out of a given number of objects by taking some or all of
the objects at a time is called a combination of the objects. If n objects are given and
we form a group of r objects chosen from these n objects, then it is called a
combination of these n objects taken r at a time. The number of combinations of n
distinct objects taken r at a time is denoted by n Cr .
Theorems on combination:
n
Pr n!
i. n Cr = = .
r! r! (n − r)!
ii. n Cr =n Cn−r .
Note: n Cn = 1, n C1 = n, n C0 = 1.
iii. The number of combinations of n different objects taken some or all at a time is
(2n − 1).
iv. The number of combinations of n objects taken some or all at a time, where p of
the objects are alike of one kind, q of the objects are alike of a second kind, r of
the objects are alike of a third kind and the remaining s objects are all different is
2s (p + 1)(q + 1)(r + 1).
v. The number of ways in which (m + n) different objects can be divided into two
(m + n)!
groups one containing m objects, and the other containing n objects is .
m! n!
(f) Choosing balls from an urn:
The total number of ways of choosing m white balls and n black balls from an urn
containing M white and N black balls is M Cm .N Cn .
(g) Ordered Partitions (distinct objects):
The total number of ways of distributing n distinct objects into r boxes marked
1, 2, 3, . . . , r is rn .
The number of ways in which the n objects can be distributed so that the boxes
n!
contain respectively n1 , n2 , . . . , nr objects is .
n1 ! n2 ! . . . n r !
(h) Ordered Partitions (identical objects):
The total number of ways of distributing n identical objects into r boxes marked
1, 2, 3, . . . , r is n+r−1 Cr−1 .
If none of the boxes should remain empty, the total number of ways of distributing the
n objects is n−1 Cr−1 .
(i) Sum of points on dice:
When n dice are thrown, the number of ways of getting a total of r points is given by
the coefficient of xr in the expansion of (x + x2 + x3 + x4 + x5 + x6 )n
(j) Derangements and Matches:
If n objects numbered 1, 2, 3, . . . , n are distributed at random in n places also
numbered 1, 2, 3, . . . , n, a match is said to occur if an object occupies the place
corresponding to its number.
The number of permutations
in which no match occurs (this is also known as
1 1 1 n
1
derangement) is tn = n! 1 − + − + · · · + (−1)
1! 2! 3! n!
The number of permutations of n objects in which exactlyr matches occur is
n n! 1 1 1 1
Cr tn−r = 1 − + − + · · · + (−1)n−r
r! 1! 2! 3! (n − r)!
14. Odds in favor/ against an event
If E is an event connected with a random experiment, then
P (E)
Odds in favor of the event E = ,
P (E)
P (E)
Odds against the event E =
P (E)
m
Remark: If the odds in favor of event E are m:n, then P (E) = . If the odds against
m+n
n
event E are m:n, then P (E) = .
m+n
15. Union and intersection of events
The union of two events A and B connected with a random experiment is denoted by
(A + B) and is defined to be the event that occurs if at least one of the events A or B
occur, i.e. if A or B or both the events occur simultaneously. More generally, the union of
the events E1 , E2 , . . . , En connected with a random experiment is denoted by
(E1 + E2 + · · · + En ) and is defined to be the event that occurs if at least one of the events
E1 , E2 , . . . , En occurs.
The intersection of two events A and B connected with a random experiment is denoted
by AB and is defined to be the event that occurs if both the events occur simultaneously.
More generally, the intersection of the events E1 , E2 , . . . , En connected with a random
experiment is denoted by (E1 E2 · · · En ) and is defined to be the event that occurs if all the
events E1 , E2 , . . . , En occur simultaneously.
16. Events connected with a random experiment are said to be mutually exclusive if occurrence
of any one of the events prevents the occurrence of any of the others.
17. Events connected with a random experiment are said to be exhaustive if at least one of
them must necessarily occur if the random experiment is performed.
18. Addition theorems
(i) If A and B are any two events connected with a random experiment then
P (A + B) = P (A) + P (B) − P (AB).
(ii) If A, B, C are any three events connected with a random experiment then
P (A + B + C) = P (A) + P (B) + P (C) − P (AB) − P (BC) − P (CA) + P (ABC).
Corollary:
(a) If the events A and B connected with a random experiment are mutually exclusive,
then P (A + B) = P (A) + P (B).
(b) If the events E1 , E2 , · · · , En connected with a random experiment are mutually
exclusive then P (E1 + E2 + · · · + En ) = P (E1 ) + P (E1 ) + · · · + P (En ).
(c) If the events E1 , E2 , · · · , En connected with a random experiment are mutually
exclusive and exhaustive then P (E1 ) + P (E1 ) + · · · + P (En ) = 1.
(d) If A and B are any two events connected with a random experiment then
P (A B) = P (A) − P (AB) and P (A B) = P (B) − P (AB).
19. De Morgan’s Laws:
If A and B are events connected with a random experiment then (i) A + B = A B and
AB = A + B
20. Conditional probability
Suppose A and B are two events connected with a random experiment and P (B) 6= 0. The
conditional probability of A given B is denoted by P (A/B) and is defined to be the
probability of occurrence of the event A given that the event B has already occurred.
21. Multiplication theorem
If A and B are two events connected with a random experiment, then
P (AB) = P (A) P (B/A) if P (A) 6= 0
.
= P (B) P (A/B) if P (B) 6= 0
P (AB) P (AB)
Remark: P (A/B) = and P (B/A) = .
P (B) P (A)
22. Independent events
Let A and B be events connected with a random experiment and P (B) 6= 0. The event A is
said to be independent of the event B if the probability of occurrence event A is not
affected by the occurrence or non occurrence of B i.e. if P (A/B) = P (A).
Corollary:
(a) Let A and B be events connected with a random experiment and P (B) 6= 0, P (A) 6= 0.
If A is independent of B then B is also independent of A. Thus, we say that the events
A and B are independent.
(b) Let A and B be events connected with a random experiment and P (B) 6= 0, P (A) 6= 0.
Then A and B are independent if and only if P (AB) = P (A)P (B).
(c) Suppose A and B be events connected with a random experiment. If A and B are
independent events then
(i) A and B are independent events. Thus, P (AB) = P (A)P (B).
(ii) A and B are independent events. Thus, P (AB) = P (A)P (B).
(iii) A and B are independent events. Thus, P (A B) = P (A)P (B).
(d) Suppose A and B be events connected with a random experiment. If A and B are
independent events then P (A + B) = P (A) + P (B) − P (A)P (B).
23. Three events A, B, C connected with a random experiment are said to be mutually
independent (or simply independent) if (i) P(AB) = P(A) P(B) (ii) P(BC) = P(B) P(C)
(iii)P(CA)=P(C) P(A) (iv) P(ABC)=P(A) P(B) P(C)
24. Bayes’ Theorem
Suppose an event A can occur only if one of the mutually exclusive and exhaustive set of
events B1 , B2 , · · · , Bn occurs. Suppose the unconditional probabilities
P (B1 ), P (B2 ), · · · , P (Bn ) are known and the conditional probabilities
P (A/B1 ), P (A/B2 ), · · · , P (A/B1 ) are also known. Then the conditional probability
P (Bi /A) of the occurrence of the event Bi given that the event A has already occurred is
given by:
P (Bi )P (A/B1 )
P (Bi /A) =
P (B1 )P (A/B1 ) + P (B2 )P (A/B2 ) + · · · + P (Bn )P (A/Bn )
Random Variable
1. A random variable is a rule that associates with each simple event of a random
experiment a definite real number. Random variables will be denoted by capital letters and
their values by small letters.
Theorem
(a) If X1 , X2 are random variables and c1 , c2 are constants then c1 X1 + c2 X2 and X1 X2 are
also random variables.
(b) If f (x) is a continuous function and X is a random variable, then f (X) is a random
variable.
2. Discrete and Continuous Random Variables:
Random variables can be of two types- discrete and continuous. A random variable that
can assume a finite number of values or a countably infinite number of values is called a
discrete random variable whereas a random variable that can assume all possible values in
a certain interval is called a continuous random variable.
3. Probability Distribution of a Discrete Random Variable:
Let X be a discrete random variable which can assume a finite number of values
x1 , x2 , . . . , xn with probabilities p1 , p2 , . . . , pn respectively. Then the set of values of X
together with their probabilities is called the probability distribution of the discrete
random variable X. The probability distribution of a discrete random variable can be
represented by a formula, a table or a graph.
4. Probability Mass Function of a Discrete Random Variable:
Let X be a discrete random variable which can assume a finite number of values
x1 , x2 , . . . , xn . Then the function f (x) defined by
(
P (X = xi ), if x = xi
f (x) =
0, otherwise
is called the probability mass function or p.m.f. of the discrete random variable X.
X
The p.m.f. f (x) always satisfies the conditions (i) 0 ≤ f (x) ≤ 1 (ii) f (x) = 1
all x
Thus P (α ≤ X ≤ β) is the area bounded by the curve y = f (x), the x-axis, and the vertical
lines x = α and x = β.
Note: In case of a continuous random variable X, P (X = c) = 0 for any real number c.
Thus, P (α ≤ X ≤ β) = P (α < X ≤ β) = P (α < X < β) = P (α < X ≤ β).
15. Properties of the Probability Density Function:
The p.d.f. of a continuous random variable X has the following properties:
Z ∞
(i) f (x) ≥ 0 (ii) f (x) dx = 1
−∞
1 (x−µ)2
f (x) = √ e− 2σ2 where ∞ < x < ∞, −∞ < µ < ∞, σ ≥ 0 (4)
σ 2π
X −µ
22. Let X v N (µ, σ). Let Z = . Then the random variable Z has normal distribution
σ
with mean 0 and S.D. 1. Z is called a standard normal variate. The p.d.f. of Z is given
by
1 (z)2
φ(z) = √ e− 2 (5)
2π
The cumulative distribution
Z z function
Z zof Z is denoted by Φ(z) and is given by
(u)2
− 2
Φ(z) = P (Z ≤ z) = φ(z) dz = e du.
−∞ −∞
Remark: Any normal distribution can be converted into a standard normal distribution
X −µ
using the transformation Let Z = . Area under a standard normal curve between the
σ
ordinates 0 to z, for z > 0, as also the values of Φ(z) which gives the area under a standard
normal curve to the left of the ordinate at z are available in statistical tables, and are used
in the calculation of probabilities.
23. Given P (0 ≤ Z ≤ z) = Area under a standard normal curve between the ordinates 0 to z,
for z ≥ 0, the following probabilities can be obtained:
(a) P (z > 0) = P (Z < 0) = 0.5
(b) For b > a > 0, P (a ≤ Z ≤ b) = P (0 ≤ Z ≤ b) − P (0 ≤ Z ≤ a).
(c) For a > 0, P (−a ≤ Z ≤ 0) = P (0 ≤ Z ≤ a).
(d) For a > 0, b > 0, P (−a ≤ Z ≤ b) = P (0 ≤ Z ≤ a) + P (0 ≤ Z ≤ b).
(e) For b > a > 0, P (−b ≤ Z ≤ −a) = P (a ≤ Z ≤ b) = P (0 ≤ Z ≤ b) − P (0 ≤ Z ≤ a).
(f) For a > 0, P (−a ≤ Z ≤ a) = 2 P (0 ≤ Z ≤ a).
(g) For a > 0, P (Z ≥ a) = P (Z ≤ −a) = 0.5 − P (0 ≤ Z ≤ a).
(h) For a > 0, P (Z ≥ −a) = P (Z ≤ a) = 0.5 + P (0 ≤ Z ≤ a).
Joint Distribution of Two Discrete Random Variables:
1. Let X and Y be random variables associated with the same random experiment. Then the
pair (X,Y) is called a bivariate random variable. If both the random variables are
discrete, then (X,Y) is called a discrete bivariate random variable. The set of possible
values of (X,Y) will be called the range of (X,Y). If both the random variables are
continuous, then (X,Y) is called a continuous bivariate random variable.
X
pY (yi ) = P (Y = yi ) = pXY (xi , yj ) (8)
xi
pX (xi ) and pY (yi ) are called the marginal pmf ’s of X and Y respectively.
Suppose a random variable X takes three possible values x1 , x2 , x3 and a random variable Y
takes three values y1 , y2 , y3 . Let pij = P (X = xi , Y = yj ). The joint probability distribution
of the variables (X,Y) are described by the following table called the joint probability
table:
y1 y2 y3 Row sum
x1 p11 p12 p13 pX (x1 )
x2 p21 p22 p23 pX (x2 )
x3 p31 p32 p33 pX (x3 )
Column Sum pY (y1 ) pY (y2 ) pY (y3 ) 1
Thus pX (xi ) = Sum of entries in row i, and pY (yj ) = Sum of entries in column j.
4. Independent Random Variables:
Let (X,Y) be a discrete bivariate random variable. If for all (xi , yj ) in the range of (X,Y)
the events (X = xi ) and (Y = yj ) are independent, then we say X and Y are independent
random variables. Thus, a pair of discrete random variables X and Y are said to be
independent if and only if
pXY (xi , yj ) = pX (xi ) pY (yj ) (9)
for all (xi , yj ) in the range of (X,Y).
5. Expectation, Variance, Standard Deviation, Covariance:
Let (X,Y) be a discrete bivariate random variable with joint pmf pXY (xi , yj ), then
" #
X X X XX
(a) E(X) = µX = xi pX (xi ) = xi pXY (xi , yj ) = xi pXY (xi , yj )
xi xi j xi yj
XX X
Also, E(X 2 ) = x2i pXY (xi , yj ) = x2i pX (xi ),
xi yj xi
2
Var(X) = σX = E(X 2 ) − [E(X)]2 .
" #
X X X XX
(b) E(Y ) = µY = yj pY (yj ) = yj pXY (xi , yj ) = yj pXY (xi , yj ),
yj yj xi xi yj
XX X
Also, E(Y 2 ) = yj2 pXY (xi , yj ) = yj2 pY (yj ),
xi yj yj
(c) The standard deviation of X, denoted by σX , is the positive square root of the
variance.
(d) The covariance of X, Y denoted by Cov(X,Y) is defined by
Cov(X, Y )X X − µX )(Y − µY ) = E(XY ) − E(X)E(Y ), where
= E(X
E(XY ) = xi yj pXY (xi , yj )
xi yj
(e) If Cov(X, Y ) = 0 then X and Y are said to be uncorrelated. Thus if X and Y are
uncorrelated then E(XY ) = E(X)E(Y ).
Theorem: (i) If X and Y are independent they are uncorrelated, but the converse is
not true in general.
(ii) If (X,Y) is a discrete bivariate random variable, then
(a) E(X + Y ) = E(X) + E(Y )
(b) V ar(X + Y ) = V ar(X) + V ar(Y ) + 2 Cov(X, Y ).
(c) If X and Y are independent then V ar(X + Y ) = V ar(X) + V ar(Y ).
(f) The correlation coefficient of X and Y is denoted by ρXY or by ρ and is defined by
Cov(X, Y )
ρ= . (10)
σX σY
Theorem: −1 ≤ ρ ≤ 1. The equality holds when there is an exact linear relation
between X and Y.
(g) Conditional Distribution:
Let X and Y be discrete random variables with joint pmf pXY (xi , yj ) and marginal
pmf’s pX (xi ) and pY (yj ) respectively, then the conditional pmf of X given Y = yj
is denoted by pX/yj (xi ) and is defined by
P (X = xi , Y = yj ) pXY (xi , yj )
pX/yj (xi ) = P (X = xi /Y = yj ) = = (11)
pY (yj ) pY (yj )
pXY (xi , yj )
pY /xi (yj ) = (12)
pX (xi )
1. Let X and Y be continuous random variables associated with the same random experiment.
If there exists a function fXY (x, y) such that :
(i) fXY (x, y) ≥ 0,
(ii) fXY (x, y) is continuous for all (x,y) except possibly a finite set,
ZZ
(iii) For any region A of the two dimensional space, P ((X, Y ) ∈ A) = fXY (x, y) dx dy,
A
Z ∞ Z ∞
(iv) fXY (x, y) dx dy = 1,
x=−∞ y=−∞
Z b Z d
Then X and Y are called jointly continuous random variables and fXY (x, y) is called
the joint (or bivariate) probability density function of the continuous bivariate
random variable (X,Y).
Remark: (i) Graphically z = fXY (x, y) represents a surface called the probability surface.
The total volume bounded by this surface and the x-y plane is 1.
(ii) It is possible to have continuous bivariate random variables X, Y that are not jointly
continuous i.e. X has probability density function fX (x), Y has probability density function
fY (y), but there is no joint density function fXY (x, y).
(iii) fXY (a, b) does not represent probability of anything.
Z ∞
Z b Z b Z ∞
xfXY (x, y) dy dx
x=−∞ y=−∞
Z ∞ Z ∞ Z ∞
2 2
Also, E(X ) = x fX (x) dx = x2 fXY (x, y) dy dx
x=−∞ x=−∞ y=−∞
2
Var(X) = σX = E(X 2 ) − [E(X)]2 .
(b) The standard deviation of X, denoted by σX , is the positive square root of the
variance.
(c) The covariance of X, Y denoted by Cov(X,Y) is defined by
Cov(X, Y )Z= E(XZ − µX )(Y − µY ) = E(XY ) − E(X)E(Y ), where
∞ ∞
(d) If Cov(X, Y ) = 0 then X and Y are said to be uncorrelated. Thus if X and Y are
uncorrelated then E(XY ) = E(X)E(Y ).
Theorem: (i) If X and Y are independent they are uncorrelated, but the converse is
not true in general.
(ii) If (X,Y) is a discrete bivariate random variable, then
(a) E(X + Y ) = E(X) + E(Y )
(b) V ar(X + Y ) = V ar(X) + V ar(Y ) + 2 Cov(X, Y ).
(c) If X and Y are independent then V ar(X + Y ) = V ar(X) + V ar(Y ).
(e) The correlation coefficient of X and Y is denoted by ρXY or by ρ and is defined by
Cov(X, Y )
ρ= . (19)
σX σY
Theorem: −1 ≤ ρ ≤ 1. The equality holds when there is an exact linear relation
between X and Y.
(f) Conditional Distribution:
Let X, Y be jointly continuous random variables with joint probability density
function fXY (x, y) and marginal probability density functions fX (x) and fY (y), then
for any real number y such that fY (y) > 0, the conditional probability density
function of X given Y = y is denoted by fX/y (x) and is defined by
fXY (x, y)
fX/y (x) = . (20)
fY (y)
Similarly, for any real number x such that fX (x) > 0, the conditional probability
density function of Y given X = x is denoted by fY /x (y) and is defined by
fXY (x, y)
fY /x (y) = (21)
fX (x)
.
Thus,
fXY (x, y) = fX (x)fY /x (yj ) = fY (y)fX/y (x). (22)
Theorem: If X and Y are independent then
fX/y (x) = fX (x) for all y, x such that fX (x) > 0 (24)
5. Conditional Mean and Variance
Let X, Y be jointly continuous random variables, the conditional expection (mean) of
Y given X = x, is denoted by E(Y /x) or µY /x and is defined by
Z ∞
E(Y /x) = yfY /x (y) dy. (25)
y=−∞