Download presentation
Presentation is loading. Please wait.
1
A Kernel-based Support Vector Machine by Peter Axelberg and Johan Löfhede
2
Maximum Margin Classifier Decision boundary 22 11 Maximum Margin
3
Need for more complex decision boundaries In general, real-world applications require more complex decision boundaries than linear functions.
4
Dimension expansion The SVM offers a method where the input space is mapped by a non-linear function,, to a higher dimensional feature space where the classes are more likely to be linearly separable. (x)(x)
5
Input space High dimensional feature space x1x1 (x1)(x1) (Xi)(Xi) Decision space 11 22 xixi (Xi)(Xi) f( ( X i )) High dimensional feature space
6
Kernel function Instead of directly using the mapping vectors in the high dimentional feature space, a kernel function K(x,x i ) can be introduced in the input space according to the following substitution: (x)(x) where denotes the inner product
7
The Kernel function used by the decision boundary function
8
Examples of Kernel functions Polynomial Radial basis function (RBF) Sigmoidal
9
A real world example Classification of some fruits/vegetables using kernel-based SVM
10
Feature 1Feature 2Feature 3Feature 4
11
Circumference 1 (longest) (cm) Circumference 2 (shortest) (cm) Weight (g)Color (code) 601630035 19 8045 242316055 252218040 252318045 441316043 211316050 462540070
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.