Pattern Recognition: Statistical and Neural

Slides:



Advertisements
Similar presentations
Aula 3 Single Layer Percetron
Advertisements

Perceptron Lecture 4.
Introduction to Neural Networks Computing
G53MLE | Machine Learning | Dr Guoping Qiu
NEURAL NETWORKS Perceptron
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Linear Discriminant Functions Wen-Hung Liao, 11/25/2008.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Intro. ANN & Fuzzy Systems Lecture 8. Learning (V): Perceptron Learning.
4 1 Perceptron Learning Rule. 4 2 Learning Rules Learning Rules : A procedure for modifying the weights and biases of a network. Learning Rules : Supervised.
Simple Neural Nets For Pattern Classification
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Linear Learning Machines  Simplest case: the decision function is a hyperplane in input space.  The Perceptron Algorithm: Rosenblatt, 1956  An on-line.
Perceptron Learning Rule
An Illustrative Example
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Linear Discriminant Functions Chapter 5 (Duda et al.)
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
SVM by Sequential Minimal Optimization (SMO)
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 13 Oct 14, 2005 Nanjing University of Science & Technology.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 23 Nov 2, 2005 Nanjing University of Science & Technology.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 20 Oct 26, 2005 Nanjing University of Science & Technology.
Discriminant Functions
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 21 Oct 28, 2005 Nanjing University of Science & Technology.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 14 Oct 14, 2005 Nanjing University of Science & Technology.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
Linear Discrimination Reading: Chapter 2 of textbook.
Non-Bayes classifiers. Linear discriminants, neural networks.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 12 Sept 30, 2005 Nanjing University of Science & Technology.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 8 Sept 23, 2005 Nanjing University of Science & Technology.
Lecture 4 Linear machine
Linear Models for Classification
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 29 Nov 11, 2005 Nanjing University of Science & Technology.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
Chapter 2 Single Layer Feedforward Networks
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 7: Linear and Generalized Discriminant Functions.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall Perceptron Rule and Convergence Proof Capacity.
November 21, 2013Computer Vision Lecture 14: Object Recognition II 1 Statistical Pattern Recognition The formal description consists of relevant numerical.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 28 Nov 9, 2005 Nanjing University of Science & Technology.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
Linear Discriminant Functions Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Artificial Neural Networks
DIFFERENTIAL EQUATIONS
Chapter 2 Single Layer Feedforward Networks
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Perceptron as one Type of Linear Discriminants
Neural Networks Chapter 5
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Support Vector Machines
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Artificial Intelligence 9. Perceptron
PROGRAMME F5 LINEAR EQUATIONS and SIMULTANEOUS LINEAR EQUATIONS.
Chapter - 3 Single Layer Percetron
Lecture 8. Learning (V): Perceptron Learning
Linear Discrimination
Perceptron Learning Rule
Perceptron Learning Rule
Perceptron Learning Rule
Pattern Recognition: Statistical and Neural
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Pattern Recognition: Statistical and Neural Nanjing University of Science & Technology Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 17 Oct 21, 2005

Lecture 17 Topics View of Perceptron Algorithm in Pattern Space Fractional Correction Perceptron Algorithm Simplified Perceptron Algorithm 4. Derivation of the perceptron algorithm 5. Extension of Perceptron Algorithm to M Class Case: 3 Special Cases

from C1 from C2 Motivation

How do we find separating Hyperplane??? a “needle in the haystack” Question: How do we find separating Hyperplane??? a “needle in the haystack” Answer: The Perceptron Algorithm !!!! Other ways exist like random selection

from C1 from C2 Motivation Separating Hyperplane

Linear Discriminant Functions where Augmented Pattern Vector Weight vector C1 Decision Rule: if d(x) > 0 < C2

Finds a hyperplane that separates two sets of patterns Review Perceptron Algorithm Finds a hyperplane that separates two sets of patterns Algorithm

Review Perceptron Algorithm New Training Sample

Review Perceptron Algorithm wT(k)x(k) < 0 wT(k)x(k) > 0 If c is too large we may not get convergence

View of Perceptron correction in pattern space

Fractional Correction Perceptron Algorithm Weight Update Equation Where

Weight update in original pattern space

Weight update in original pattern space

Weight update in original pattern space

Simplified Perceptron Algorithm Given the following samples from two classes We want a weight vector w such that Negated patterns for class C2

We could define two classes as Consider the set of samples x of C11 U C21 then the weight vector update is as follows That is all of the samples should be on the positive side of the hyperplane boundary

This is a simplification of the perceptron algorithm in that it only contains one if branch Simplified algorithm: 1. Augment all patterns in C1 and C2 2. Negate the augmented patterns of C2 3. Combine the two sets above to form one set of samples 4. Iterate through this set using the weight update as

Derivation of the Perceptron Algorithm Define the following performance measure

Minimizing J for each sample will satisfy the conditions desired To minimize J we can use the gradient algorithm where the weight update is as follows

The partial derivatives with respect to each weight are determined as

Substituting these partials into the weight update equation yields the following for

Which can be written in the following vector form Which is the perceptron algorithm in augmented and negated form. (end of proof)

Other Perceptron like Algorithms If we use different performance measures in the preceding proof we get perceptron like algorithms to accomplish the separation of classes.

Other meaningful performance measures Each leads to different weight update equations. Each requires a convergence theorem to be useful.

So far only worked with separating two classes So far only worked with separating two classes. Most problems have more than two classes. Question: Can we modify the perceptron algorithm to work with more than two classes? Answer: Yes for certain special cases.

Case 1. Pattern classes Group separable

Case 1. K Pattern classes - Group separable Given: S1 is set of samples from class C1 S2 is set of samples from class C2 SK is set of samples from class CK S = S1 U S2 U . . . U SK … Define: Assume: Pattern classes are group separable S1 linearly separable from S1/ = S – S1 S2 linearly separable from S2/ = S – S2 SK linearly separable from SK/ = S – SK …

dk(x)=wk1x1 + wk2x2 + … + wknxn + wkn+1 Find K Linear Discriminant functions that separate the Sk in the following manner Discriminant Functions for k = 1, 2, … , K dk(x)=wk1x1 + wk2x2 + … + wknxn + wkn+1 Decision Rule: If dj(x) > 0 then decide Cj Solution: Find dk(x) by using the Perceptron algorithm on the two classes Sk and Sk/

Solution Continued (1) Find d1(x) by using the Perceptron algorithm on the two sets S1 and S1/ (2) Find d2(x) by using the Perceptron algorithm on the two sets S2 and S2/ … (K) Find dK(x) by using the Perceptron algorithm on the two sets SK and SK/

0 = wk1x1 + wk2x2 + … + wknxn + wkn+1 Solution Continued This gives the decision boundaries as: for k = 1, 2, … , K 0 = wk1x1 + wk2x2 + … + wknxn + wkn+1

Case 2. Pattern classes Pairwise separable

dkj(x)=wkj1x1 + wkj2x2 + … + wkjnxn + wkjn+1 Case 2. Pattern classes Pairwise separable Find Linear Discriminant functions that separate all pairs of Sk in the following manner: for k = 1, 2, … , K j = 1, 2, … , K j = k dkj(x)=wkj1x1 + wkj2x2 + … + wkjnxn + wkjn+1 Decision Rule: for k = 1, 2, … , K If dkj(x) > 0 for all j , j = k then decide Ck On Boundaries decide randomly

Solution: for k = 1, 2, … , K j = 1, 2, … , K , j = k Find dkj(x) by using the Perceptron algorithm on the two classes Sk and Sj . Notes: (1) dkj(x) = - djk(x) . (2) Requires determining K(K-1)/2 discriminant functions

Case 3. K Pattern classes separable by K discriminant functions

Flow Diagram for Case 3 Perceptron separability

Summary Lecture 17 View of Perceptron Algorithm in Pattern Space Fractional Correction Perceptron Algorithm Simplified Perceptron Algorithm 4. Derivation of the perceptron algorithm 5. Extension of Perceptron Algorithm to M Class Case: 3 Special Cases

End of Lecture 17