Professional Documents
Culture Documents
Machine Learning
Math Essentials
Part 2
Jeff Howbert Introduction to Machine Learning Winter 2012 #
Most commonly used continuous probability
distribution
Also known as the normal distribution
Two parameters define a Gaussian:
Mean location of center
Variance o
2
width of curve
Gaussian distribution
Jeff Howbert Introduction to Machine Learning Winter 2012 #
2
2
2
) (
2 / 1 2
2
) 2 (
1
) , | (
o
to
o
=
x
e x
Gaussian distribution
In one dimension
Jeff Howbert Introduction to Machine Learning Winter 2012 #
2
2
2
) (
2 / 1 2
2
) 2 (
1
) , | (
o
to
o
=
x
e x
Gaussian distribution
In one dimension
Normalizing constant:
insures that distribution
integrates to 1
Controls width of curve
Causes pdf to decrease as
distance from center
increases
Jeff Howbert Introduction to Machine Learning Winter 2012 #
Gaussian distribution
= 0 o
2
= 1 = 2 o
2
= 1
= 0 o
2
= 5 = -2 o
2
= 0.3
2
2
1
2
1
x
e x
=
t
Jeff Howbert Introduction to Machine Learning Winter 2012 #
) ( ) (
2
1
2 / 1 2 /
1 T
| |
1
) 2 (
1
) , | (
x x
x
= e
d
t
Multivariate Gaussian distribution
In d dimensions
x and now d-dimensional vectors
gives center of distribution in d-dimensional space
o
2
replaced by E, the d x d covariance matrix
E contains pairwise covariances of every pair of features
Diagonal elements of E are variances o
2
of individual
features
E describes distributions shape and spread
Jeff Howbert Introduction to Machine Learning Winter 2012 #
Covariance
Measures tendency for two variables to deviate from
their means in same (or opposite) directions at same
time
Multivariate Gaussian distribution
n
o
c
o
v
a
r
i
a
n
c
e
h
i
g
h
(
p
o
s
i
t
i
v
e
)
c
o
v
a
r
i
a
n
c
e
Jeff Howbert Introduction to Machine Learning Winter 2012 #
In two dimensions
Multivariate Gaussian distribution
(
=
(
=
1 3 . 0
3 . 0 25 . 0
0
0
=
2 6 . 0
6 . 0 2
=
1 0
0 2
=
1 0
0 1