Download presentation
Presentation is loading. Please wait.
Published byDorthy Grant Modified over 9 years ago
1
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr
2
Given : { [x 1, y 1 ], [x 2, y 2 ], … [x n, y n ] } Binary classification task x i - feature vector y i {-1, 1} - label The task is : To design y = H(x) that predicts a label y for a new x
3
Given : { [x 1, y 1 ], [x 2, y 2 ], … [x n, y n ] } x i - feature vector y i {-1, 1} - label The task is : To design y = H(x) that predicts a label y for a new x Several approaches proposed : Support Vector Machines, Boosting, Random Forests, Neural networks, Nearest Neighbour,.. Binary classification task
4
Linear SVMs Fast training and evaluation Applicable to large scale data sets Low discriminative power Kernel SVMs Slow training and evaluation Not feasible to too large data sets Much better performance for hard problems Linear vs. Kernel SVMs
5
The goal is to design an SVM with: Good trade-off between the performance and speed Scalability to large scale data sets (solvable using SGD) Motivation
6
Points approximated as a weighted sum of anchor points v – anchor points γ V (x) - coordinates Local codings
7
Points approximated as a weighted sum of anchor points Coordinates obtained using : Distance based methods (Gemert et al. 2008, Zhou et al. 2009) Reconstruction methods (Roweis et al.2000, Yu et al. 2009, Gao et al. 2010) v – anchor points γ V (x) - coordinates Local codings
8
For a normalised codings and any Lipschitz function f : (Yu et al. 2009) v – anchor points γ V (x) - coordinates Local codings
9
The form of the classifier Weights w and bias b obtained as : (Vapnik & Learner 1963, Cortes & Vapnik 1995) [x k, y k ] – training samples λ- regularisation weight S – number of samples Linear SVMs
10
The decision boundary should be smooth Approximately linear is sufficiently small region Locally Linear SVMs
11
The decision boundary should be smooth Approximately linear is sufficiently small region Locally Linear SVMs
12
The decision boundary should be smooth Approximately linear is sufficiently small region Locally Linear SVMs
13
The decision boundary should be smooth Approximately linear is sufficiently small region Curvature is bounded Functions w i (x) and b(x) are Lipschitz w i (x) and b(x) can be approximated using local coding Locally Linear SVMs
14
The decision boundary should be smooth Approximately linear is sufficiently small region Curvature is bounded Functions w i (x) and b(x) are Lipschitz w i (x) and b(x) can be approximated using local coding
15
The classifier takes the form : Weights W and biases b are obtained as : where Locally Linear SVMs
16
Optimised using stochastic gradient descent (Bordes et al. 2005) : Locally Linear SVMs
17
Generalisation of Linear SVM on x : - representable by W = (w w..) T and b = (b’ b’..) T Generalisation of Linear SVM on γ: - representable by W = 0 and b = w Generalisation of model selecting Latent(MI)-SVM : - representable by γ = (0, 0,.., 1,.. 0) Relation to other models
18
The finite kernel classifier takes the form where is the kernel function Weights W and b obtained as : Extension to finite kernels
19
MNIST, LETTER & USPS datasets Anchor points obtained using K-means clustering Coordinates evaluated on kNN (k = 8) (slow part) Coordinates obtained using weighted inverse distance Raw data used CALTECH-101 (15 training samples per class) Coordinates evaluated on kNN (k = 5) Approximated Intersection kernel used (Vedaldi & Zisserman 2010) Spatial pyramid of BOW features Coordinates evaluated based on histogram over whole image Experiments
20
MNIST Experiments
21
UIUC / LETTER Caltech-101 Experiments
22
We propose novel Locally Linear SVM formulation with Good trade-off between speed and performance Scalability to large scale data sets Easy to implement Optimal way of learning anchor points is an open question Questions ? Conclusions
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.