Professional Documents
Culture Documents
Rules of Probability and Combinatorial Principles
..
Basic Set Operations
A ∪ B = {x ∣ x ∈ A or x ∈ B or both} ; A ∩ B = {x ∣ x ∈ A and x ∈ B}
..
c c c c c c
′
A or A
c
= complement of A = {x ∣ x ∉ A} ; (A ∪ B) = A ∩ B ; (A ∩ B) = A ∪ B
A ⊂ B = B contains all the sample points in event A. It calls A is a subevent of B.
If A ⊂ B , then A ∩ B = A and A ∪ B = B
Basic Probability
. For a sample space S , ; ∅] where ∅ is the null or empty set
..
Pr[S ] = 1 Pr[ = 0
For an event A in the sample space S, 1 ≥ Pr[A] ≥ 0
.
c
For any event A, Pr[A] = 1 − Pr[A ]
∅),
General Formula :
..
Pr[A ∪ B] = Pr[A] + Pr[B] − Pr[A ∩ B]
c
]
Conditional Probability
. For two events A and B, the conditional probability of A given B has occurred is:
Pr[A ∩ B]
Pr[A ∣ B ] =
Pr[B]
. If A ⊂ B , then Pr[A ∣ B] =
Pr[A]
and Pr [B ∣A ] = 1
.
Pr[B]
If
c
Pr[A ∣ B ] = 1 − Pr[A ∣ B ] and A1∩ A2 = ∅ then Pr[ A 1 ∪ A 2 ∣ B ] = Pr[ A 1 ∣ B ] + Pr[ A 1 ∣ B ]
Baye's Theorem
Let A1 , A2 , … , An be a collection of n mutually exclusive and exhaustive events with Pr[ Ai] > 0 for i = 1 , …., n .
Then, for any other event B ,
n
Pr[B] = ∑ Pr [B ∣ A i ] Pr[A i ]
i=1
Independence
Two events A and B are independent if and only if Pr[ B ∣ A ] = Pr[B ]
Fundamental Principle of Counting: (also known as the multiplication rule for counting) If a task
can be performed in n1 ways, and for each of these a second task can be performed in n2 ways, and
for each of the latter a third task can be performed in n3 ways, ... , and for each of the latter a kth task
can be performed in nk ways, then the entire sequence of k tasks can be performed in
n1 • n2 • n3 • ... • nk ways.
Permutations
An ordered sequence of k objects taken from a set of n distinct objects is called a permutation of the
n
objects of size k. P k denotes the number of size k permutations that can be constructed from the
n objects.
n n!
Pk =
(n − k)!
Combinations
An unordered subset of k objects taken from a set of n distinct objects is called a combination .
n
( )
k
denotes the number of combinations of size k that can be formed from n objects.
n n!
( ) =
k (n − k)! k!
n n n n n n
( ) = 1 ( ) = 1 ( ) = ( ) = n ( )= ( )
n 0 1 n−1 k n−k
Random Variables and Probability Distributions, Expectation and
Other Distributional Parameters
Random Variables
Probability Distributions and Independence
Given an experiment with sample space S , a random variable X is any rule that associates a number with each
outcome in S .
The cumulative distribution function (cdf) of a random variable written F (x) is defined for every number x by: F (x) = Pr[X ≤ x]
The cumulative distribution function (cdf) of a discrete random variable with pmf p(x) written F(x) is defined by:
F (x) = ∑ Pr[X = k ] for every number x
k ≤x
Two random discrete variables X and Y are independent if Pr [ X=x, Y=y ] = Pr [ X=x ] Pr [ Y=y ] for every x and y.
Continuous Random Variables
A random variable X is continuous if its set of possible values is an entire interval of numbers.
The probability distribution or probability density function (pdf) of a continuous random variable written f (x) is :
Pr[a ≤ X ≤ b] = ∫ f (x) dx
a
∫
∞ ( ) 1
The probability density function f (x ) must satisfy: f (x ) ≥ 0 and ∫ −∞
d
f (x) x = 1 .
x
The cumulative distribution function of a continuous random variable written F (x) is: F (x) = Pr[X ≤ x] = ∫
−∞
f (y) dy
′
We have Pr[a ≤ x ≤ b] = F (b) − F (a) and F (x) = f (x)
Expectation, Variance, Standard deviation, Covariance
The expected value or mean value of a discrete random variable written E(X) or μ is:
X
2 2
V ar(X) = ∑(x − μX ) Pr[X = x] if X is dicrete and V ar(X) = ∫ (x − μ
X
) f (x) dx if X is continuous.
x −∞
√
σ = standard deviation =
X V ar(X)
Rules of expected value:
For any constant a, E(aX) = aE(X)
Rules of variance:
For any constants a and b, V ar(aX + b) 2
= a [V ar(X)]
p = F [η(p)] = ∫ f (y) y
−∞
~ ~
The median of a continuous distribution written μ satisfies 0.5 = F (μ) percentile of the distribution.
th
= the 50
The mode of a distribution is any point m at which the probability mass function p(x) or the density function f (x) is maximized.
3
E[(X− μ) ]
The skewness of a distribution is:
3
σ
Discrete Distributions
. possible outcomes,
2. Binomial An experiment consisting of a fixed number of n trials. Each trial is identical and results in one of two
denoted
success (S)
or failure (F) The trials are independent and the probability of a success is denoted p. The binomial random
variable X is equal
to the
number
of successes in n trials.
n n
p(x)= ( x )px (1 − p) n−x, x = 0, 1, 2, ....
,n
, E(X) = np V ar(X) = np(1 − p) M (t) = ( 1 − p + p e t )
X
4. Pascal Related
to binomial
. An experiment continues
until a total of r successes have been observed (
r is a positive integer).
p(x)
x −1 r x− r
, x = r , r+ 1, r+ 2, ....
r r(1−p) p et
[r
= ( ) p (1 − p) (X) =
E V ar(X) =
r−1 p p
2
MX (t) = [ 1−(1−p) e t
6. Poisson Not based p
of event that occurs in a
a certain type
on any simple experiment . M ost often used to count the number of
period of time. x
−λ t
e λ
p(x) = ,
x = 0, 1, 2, … E(X) = λ V ar(X) = λ M (t) = e
λ ( e −1)
x! X
1. Uniform X is distributed uniformly over (a, b) .
2
f (x) =
1
,
a < x < b E(X) =
b−a
V ar(X) =
(b−a)
M (t) =
e
bt
−e
at
b−a 2 12 X (b−a) t
2. Normal Distribution given that is the mean and is the variance of the distribution.
2
(x−μ) 2 2
1 − 2 σ t
f (x) = e 2 σ2 E (X) = μ V ar(X) = σ M (t) = exp [μt + ]
√2πσ
X 2
Z √2 π
X−μ
If X has a normal distribution, it can be standardized by : Z =
σ
.
∞
3 Gamma For α > 0 , the gamma function Γ (α) is defined
by: Γ(α) = ∫
0
x
α−1 −x
e dx .
Γ(α + 1) = α Γ( α ) and Γ (n ) = (n −1 )! for n positive integer.
A continuous random variable X is said to have a gamma distribution with parameters α , β < 0 if:
α
β x β
, x < 0
α−1 − β α α α
f (x) = x e E (X) = V ar(X) = M (t) = ( ) f or t < β
Γ(α) β
2 X β−t
β
< 1
f (x) = λe
−λx
,
x 0 E(X) =
1
V ar(X) =
2
M (t) =
λ
f or t < λ
X
λ λ λ−t
Joint, Marginal and Conditional Distributions
Let X and Y be two discrete random variables
defined the
on sample
space o f an experiment. The
S
joint probability mass function p (x, y), is defined for each (x,y) pair by:
X, Y
Let X and Y be two continuous random variables. Then, the joint probability density function f (x, y) = f (x, y)
X, Y
must satisfy: ∫
∞
∞
∫∞
∞
( , ) = 1
f (x, y) ≥ 0 and ∫−∞ ∫−∞ f (x, y) dx dy = 1 .
Marginal Distribution
The marginal proability mass function for the discrete random variables X and Y , denoted p (x) and
X
p (y) are given by:
Y
In the continuous case, they are given by:
∞ ∞
f (x) = ∫
X −∞
f (x, y) d y and f (y) = ∫
Y −∞
f (x, y) dx
Expected Value
Let X and Y be two jointly distributed
random variables. The expected value of a function h of (X, Y )
denoted E [h(X, Y ) ] is:
E[ h(X, Y ) ] = ∑ ∑ h(x, y) p (x, y) If X and Y are discrete
x y
∞ ∞
E[ h(X, Y ) ] = ∫ ∫ h(x, y) f (x, y) dx d y If X and Y are continuous
−∞ −∞
Moment Generating Function
The moment generating function for two jointly distributed random variables X and Y , denoted
MX,Y (t1 , t2 ) is:
t1 X+ t2 Y
MX,Y (t1 , t2 ) = E[e ]
∞ ∞
C ov(X, Y ) = ∑ ∑ x y p (x, y) −μ
X
μ
Y
i n the discrete case C ov(X, Y ) = ∫ ∫ x y f (x, y) dx d y − μ μ i n the continuous case
x y X Y
− ∞ − ∞
If X and Y are independent, then E[X ⋅ Y ] = E[X]E[Y ] C ov(X, Y ) = 0
C ov(X, Y )
ρ
X,Y
=
σ σ
X Y
E[ X ∣ Y ] = g ( Y ) V ar[X ∣ Y ] = E[X ∣ Y ]
2
E[X ∣ Y ]
2