Professional Documents
Culture Documents
x3
Feature 2
x1 Class 1
x Class 3
x = 2
x
x d
x1 x2
Class 2
Feature 1
g Classifiers
n The goal of a classifier is to partition feature space into class-labeled
decision regions
n Borders between decision regions are called decision boundaries
R1
R1
R2 R3
R2
R3
R4
Feedback / Adaptation
Classification Class
algorithm assignment
Preprocessing Feature
The Clustering Cluster
Sensor and
“real world” extraction algorithm assignment
enhancement
Regression Predicted
algorithm variable(s)
CCD
camera
Conveyor
belt (salmon)
computer
Conveyor
belt
Robot
arm Conveyor
belt (bass)
*Adapted from Duda, Hart and Stork,
Pattern Classification, 2nd Ed.
g Classification
Decision
n Collect a set of examples from both species count
boundary
length
Decision
boundary
n We combine “length” and “average
intensity of the scales” to improve class
separability
n We compute a linear discriminant function
to separate the two classes, and obtain a
classification rate of 95.7% Sea bass Salmon
New
length
length
Decision
boundary Decision
boundary
length
network with five hidden layers, a
combination of logistic and hyperbolic
tangent activation functions, train it
with the Levenberg-Marquardt algorithm
and obtain an impressive classification
rate of 99.9975% with the following
decision boundary Sea bass Salmon
n Satisfied with our classifier, we integrate the system and deploy it to the
fish processing plant
g A few days later the plant manager calls to complain that the system is
misclassifying an average of 25% of the fish
g What went wrong?
probability
A2
A1 Probability
Law
A3 A1 A2 A3 A4 event
A4
g Conditional Probability
n If A and B are two events, the probability of event A when we already
know that event B has occurred P[A|B] is defined by the relation
P[A I B]
P[A | B] = for P[B] > 0
P[B]
g P[A|B] is read as the “conditional probability of A conditioned on B”,
or simply the “probability of A given B”
Intelligent Sensor Systems 11
Ricardo Gutierrez-Osuna
Wright State University
Review of probability theory
g Conditional probability: graphical interpretation
S S
N
P[A] = P[A | B1 ]P[B1 ] + ...P[A | BN ]P[B N ] = ∑ P[A | Bk ]P[B k ]
k =1
B3
B1 BN-1
B2 BN
∑ P[A | B ] ⋅ P[B
P[A]
k k ]
k =1
1
1
5/6
pdf 4/6
pmf
3/6
2/6
1/6
100 200 300 x(lb)
400 500
1 2 3 4 5 6 x
pdf for a person’s weight
pmf for rolling a (fair) dice
n Partial: Statistics
g Expectation
n The expectation represents the center of mass of a density
g Variance
n The variance represents the spread about the mean
g Covariance (only for random vectors)
n The tendency of each pair of features to vary together, i.e., to co-vary
Xi Xi Xi Xi Xi
C ik=-σiσk Cik=-½σiσ k Cik=0 Cik=+½σiσk Cik=σiσk
ρik=-1 ρik=-½ ρik=0 ρik=+½ ρik=+1
1 1
fX (x) = exp − (X −
T
∑ −1(X −
(2 π )n/2 ∑ 2
1/2
1 1 X - 2
fX (x) = exp−
2π σ 2 σ