Professional Documents
Culture Documents
Introduction
The hard margin is a one which clearly separate positive and negative points.
Soft margin is also called as noisy linear SVM which includes some miss-
classified points.
Solution to the soft margin is approximation of points which are miss-
classified in linear decision boundary.
Loss function
Conti..
=-1
=+1
Overtraining/overfitting 2
A measure of the risk of overtraining with SVM (there are also other
measures).
It can be shown that: The portion, n, of unseen data that will be
missclassified is bounded by:
n Number of support vectors / number of training examples
Ockham´s razor principle: Simpler system are better than more complex ones.
In SVM case: fewer support vectors mean a simpler representation of the
hyperplane.
The biggest limitation of SVM lies in the choice of the kernel (the best choice
of kernel for a given problem is still a research problem).
A second limitation is speed and size (mostly in training - for large training
sets, it typically selects a small number of support vectors, thereby
minimizing the computational requirements during testing).
The optimal design for multiclass SVM classifiers is also a research area.