Presentation is loading. Please wait.

Presentation is loading. Please wait.

Learning in Feature Space (Could Simplify the Classification Task)  Learning in a high dimensional space could degrade generalization performance  This.

Similar presentations


Presentation on theme: "Learning in Feature Space (Could Simplify the Classification Task)  Learning in a high dimensional space could degrade generalization performance  This."— Presentation transcript:

1 Learning in Feature Space (Could Simplify the Classification Task)  Learning in a high dimensional space could degrade generalization performance  This phenomenon is called curse of dimensionality  By using a kernel function, that represents the inner product of training example in feature space, we never need to explicitly know what the nonlinear map is.  Even do not know the dimensionality of feature space  There is no free lunch  Deal with a huge and dense kernel matrix Reduced kernel can avoid this difficulty

2

3 Linear Machine in Feature Space Let be a nonlinear map from the input space to some feature space The classifier will be in the form ( Primal ): Make it in the dual form:

4 The Perceptron Algorithm (Dual Form) Given a linearly separable training setand Repeat: until no mistakes made within the for loop return:

5 Kernel: Represent Inner Product in Feature Space The classifier will become: Definition: A kernel is a function such that where

6 A Simple Example of Kernel Polynomial Kernel of Degree 2: Let and the nonlinear map defined by. Then.  There are many other nonlinear maps,, that satisfy the relation:

7 Power of the Kernel Technique Consider a nonlinear mapthat consists of distinct features of all the monomials of degree d. Then. For example:  Is it necessary? We only need to know !  This can be achieved

8 Basic Properties of Kernel Function  Symmetric (inherit from inner product)  Cauchy-Schwarz inequality  These conditions are not sufficient to guarantee the existence of a feature space

9 Characterization of Kernels Motivation in Finite Input Space Consider a finite space and is a symmetric function on. Letbe a matrix defined as following: There is an orthogonal matrix such that:

10 Characterization of Kernels Assume: Let Be Positive Semi-definite where

11 Mercer’s Conditions: Guarantee the Existence of Feature Space and is a symmetric function on. be a finite space Let Then is a kernel function if and only if is positive semi-definite.  What if is infinite (but compact)? Mercer’s conditions: Any finite subset of the corresponding matrix is positive semi-definite.

12 Making Kernels Kernels Satisfy a Number of Closure Properties Let Then the following functions are kernels: be kernels over be a kernel over and be a symmetric positive semi-definite.

13 Translation Invariant Kernels two inputs is unchanged if both are translated by the same vector.  The inner product (in the feature space) of  The kernels are in the form:  Some examples:  Gaussian RBF:  Multiquadric:  Fourier: see Example 3.9 on p. 37

14 A Negative Definite Kernel  Generalized Support Vector Machine  The kernelis negative definite  Does not satisfy Mercer ’ s conditions  Oliv L. Mangansarian used this kernel to solve XOR classification problem


Download ppt "Learning in Feature Space (Could Simplify the Classification Task)  Learning in a high dimensional space could degrade generalization performance  This."

Similar presentations


Ads by Google