Download presentation
Presentation is loading. Please wait.
Published byAustin Davis Modified over 9 years ago
1
Support Vector Machines Project מגישים : גיל טל ואורן אגם מנחה : מיקי אלעד נובמבר 1999 הטכניון מכון טכנולוגי לישראל הפקולטה להנדסת חשמל המעבדה לעיבוד וניתוח תמונות
2
Introduction SVM is an emerging technique for supervised learning problems, which might replace neural-networks. Main features: –Good generalization error - maximal margin. –Convex optimization problem. –Linear and non-linear decision surfaces. Proposed initially by Vapnik (82’).
3
Project Objectives Learn the theory of SVM, Design an efficient training algorithm, Create a 2D demo in order to explain the features of the SVM, and the parameters involved, and Create a C++ software package which can serve as a platform for learnining problems.
4
Supervised Learning Input: where: input vectors, classification values. Purpose: Find a machine I(z) that classifies correctly the training data, and generalizes well to other inputs.
5
Neural Networks 1. Training involves a solution of non- convex optimiz. problem. 2. Generalization error is typically not satisfactory. 3. Hard to choose the structure of the net.
6
Support Vector Machine (SVM) Input Vectors are mapped to a high dimensional feature space Z. (1. How ?) In this space a hyper-plane decision surface is constructed. (2. How ?) This decision surface has special properties that ensure high generalization. (3. How ?) Training is done in a numerically feasible approach. (4. How ?)
7
1. Mapping to Higher Dimension Map the vectors from to a higher dimension (N>n) using a non-linear mapping function chosen a priori. Basic idea: a linear separation in the N-dim. is a non-linear separating surface in the n-dim.
8
Example: Non-Linear Mapping As a different example, if the input vectors has n=200, and we are using 5th order polynomial, N has BILLIONS OF ENTRIES There is a computational problem that must be taken care of
9
2. Separating Hyper-plane Input: The input is linearly separable if there exists a vector W and a scalar b such that: The separating hyperplane is given by
10
3. Optimal Hyper-plane 1. SVM defines the optimal hyper-plane as the one with maximal margin 2. It can be shown that the margin is given by QP Problem
11
LaGrange multipliers To do so we construct a Lagrangian: At the point of minimum we get: Most of the α’s are zeros. The non-zeros correspond to the points satisfing the inequalities as equalities These points are called the SUPPORT-VECTORS
12
Decision Law
13
Classification by SVM
14
4. Using Kernel functions Let us restrict the kind of functions such that Examples:
15
The QP Problem is the vector of weights (laGrange mult.), 1 is a vector of ones D is a matrix with the entries: Using kernel functions, the overall problem remain QP:
16
The Decision Rule Using Kernel functions, we are required to perform inner-products in the lower (n) dimension only, both for training and for applying it on input patterns. By solving for the optimal we actually find the support vector
17
Results 1. Write here about the software that you developed 2. Cut and paste an image which will show the application window 3. Add more examples (for example - show how the same non-linear problem is treated with growing d - the polynomial degree) 4. Say something about the algorithm that you have implemented (main features)
18
Example 1: Linear Classification
19
Example 2: Non-Linear Separation
20
Conclusions
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.