Professional Documents
Culture Documents
Introduction to
Machine Learning
B. Solaiman
Basel.solaiman@imt-atlantique.fr
Départ. Image &Traitement de l’Information
Brest, France
2. Machine Learning
3. Learning Paradigm
4. Application examples
1
10/02/2020
1. Introduction:
Pattern Recognition Concepts
• Pattern Recognition :
Scientific discipline targeting to conceive systems/approaches/methods allowing to
make reasonable decisions assigning “observed patterns” from the real world to
some “categories/classes” of patterns
Assign an unknown pattern to one of several known categories (or classes)
2
10/02/2020
• Pattern Recognition :
Scientific discipline targeting to conceive systems/approaches/methods allowing to
make reasonable decisions assigning “observed patterns” from the real world to
some “categories/classes” of patterns
• Pattern Recognition :
Scientific discipline targeting to conceive systems/approaches/methods allowing to
make reasonable decisions concerning the belonging of “observed patterns” from
the real world to some “categories/classes” of patterns
What is a pattern?
3
10/02/2020
u1
u2 ….. uK
uk
4
10/02/2020
5
10/02/2020
Features
u1
u2 ….. uK
uk
C2
C1 …..
CM
Cm
6
10/02/2020
Sensing challenge
Availability / Accecebility of the input data that can be measured/observed from
objects to be recognized !
Features extraction challenge
Intraset features : features encoding objects which are common features to all
patterns belonging to each given class
Interest features : features allowing to discriminate between pattern classes
(discrimination capacity, dimensionality reduction, …)
Introduction to Machine Learning----------- B. Solaiman
7
10/02/2020
Discriminative Features
Classes’ models
Machine World object :
Feature Vector Recognition
DECISION
X ( x1 , x2 , , xN ) Algorithm
8
10/02/2020
CRUCIAL
STEP
C2 Mod2
C1 ….. Categories - Classes
Models Mod1….. ModM
CM (Classes Knowledge
Cm Modm
Representation)
9
10/02/2020
10
10/02/2020
Case-based reasoning
(in artificial intelligence)
Bénigne Maligne
11
10/02/2020
23
2. Machine Learning
2. Machine Learning 24
12
10/02/2020
2. Machine Learning 25
In machine learning:
- instead of injecting a massive knowledge to solve the problem (Recognition,
Estimation, Prediction…), a set of samples as well as a generic model are
injected to the “machine”
- learning refers to making the model tuned
- the resulting tuned / learned model is expected to solve the problem extremely
accurately
B Learned
MACHINE B1 Model
Generic
Test Samples
Model
2. Machine Learning 26
13
10/02/2020
2. Machine Learning 27
Training sample
dataset
2. Machine Learning 28
14
10/02/2020
2. Machine Learning 29
Validation Data
Training Data V
B ML Learned Model
Model
Generic Model Strategy Tuning 𝑓 ( . /B )
f(.)
Output Model
(program)
Expected Output
Model
30
3. Learning paradigm
15
10/02/2020
3. Learning paradigm 31
Learning A general term denoting the way in which people and computers:
- Acquire and organize knowledge and skills (by building, modifying and organizing
internal representations of some external reality); TRAINING
- Discover new knowledge (by creating hypotheses that explain some data or
phenomena); MINING
3. Learning paradigm 32
16
10/02/2020
3. Learning paradigm 33
3. Learning paradigm 34
Online Learning
Model
B
New Model
Example updating
17
10/02/2020
3. Learning paradigm 35
Informative Channels
(Explanatory Variables)
Data x1 x2 x3 x4 …….. xN y
N: Data dimension
Observations, measures, features..
(Qualitative, Quantitative, ….)
Target Variable
Introduction to Machine Learning----------- B. Solaiman
3. Learning paradigm 36
B = {(Xn,Yn) n = 1,..,N}
(Xn,Yn) i.i.d. N observations of the random pair of variables (X, Y)
18
10/02/2020
3. Learning paradigm 37
3. Learning paradigm 38
Training Data
D Real World B
(X,Y)
Data Sampling
V
Validation Data
Y
No Target Variable Y Unlabeled Training Data
19
10/02/2020
3. Learning paradigm 39
3. Learning paradigm 40
20
10/02/2020
3. Learning paradigm 41
Supervised Learning:
Learning with a series of labelled examples (i.e. target variables Yn, n = 1,2…N are
used for learning f) with direct feedback
3. Learning paradigm 42
• Supervised Learning
Set of Labeled
Training Data
21
10/02/2020
Sensor 2
x
x
x o
x x
x
Target feature : o x
Instance o x
Categorical (Label /Class) o o
o
f( ) = “Apple” o
f( ) = “Tomato” Sensor 1
x Class 1
f( ) = “Cow” o Class 2
Learned
Model
22
10/02/2020
Learning Step
Training
Labeled base od training
Labels
instances/patterns/samples/
examples…
Image Learned
Training
Features model
Testing/Generalization Step
Image Learned
Prediction
Features model
Test Image
Strong Learner :
A learning algorithm that is given a labelled training set and produces a
classifier which can be close to perfect performance (close to 100%)
Weak Learner :
A learning algorithm that is given a labelled training set and produces a
classifier which is only slightly better than random classification guess
(close to 50%°
23
10/02/2020
3. Learning paradigm 47
• Unsupervised Learning
Sensor 2
Set of Unlabeled
x x
Training Data x x x
x
x x
x
x
x x x
x x
x x
x
x
x
Instance / Example….
Sensor 1
• Explore the data to find some intrinsic structures in them and discover
patterns and structures in unlabeled data
3. Learning paradigm 48
• Semi-supervised Learning
• Labeled data is generally hard to obtain – Unlabeled data is generally cheap
24
10/02/2020
49
4. Application Examples
50
Intelligent Transportation
The Integrated Safety
Program: PreVent
Environment Sensing /
Sensor Fusion
25
10/02/2020
Intelligent Transportation 51
52
Intelligent Transportation
Estimated trajectory
doppler
Complementarity
26
10/02/2020
Intelligent Transportation 53
ASSISTED DRIVING
Set of - A reasonable
Labeled Training Algorithm / Model /
distance measure
Base Learner /
Training d(X,Y) Classifier / …
Data Inducer
- K
Classifier :
For a given instance X,
Get the top K dataset instances that are “nearest” to X
Inspect the category of these K instances
Choose the category C that represent the most instances
Conclude that X belongs to category C
27
10/02/2020
28
10/02/2020
Selecting K
Increase K:
Makes KNN less sensitive to noise
Decrease K:
Allows capturing finer structure of space
Time complexity
Suppose there are M instances and N features in the dataset
Nearest neighbor algorithm requires computing M distances
Each distance computation involves scanning through each
feature value
Running time complexity is, thus, proportional to M.N
29
10/02/2020
30
10/02/2020
31