The Perceptron CS/CMPE 333 – Neural Networks. CS/CMPE 333 - Neural Networks (Sp 2002/2003) - Asim LUMS2 The Perceptron – Basics Simplest and one.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Perceptron Lecture 4.
Slides from: Doug Gray, David Poole
Introduction to Neural Networks Computing
G53MLE | Machine Learning | Dr Guoping Qiu
Perceptron Learning Rule
NEURAL NETWORKS Perceptron
Artificial Neural Networks
Self Organization: Competitive Learning
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Perceptron.
Simple Neural Nets For Pattern Classification
A Review: Architecture
BP - Review CS/CMPE 333 – Neural Networks. CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS2 Notation Consider a MLP with P input, Q hidden,
The back-propagation training algorithm
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Correlation Matrix Memory CS/CMPE 333 – Neural Networks.
Least-Mean-Square Algorithm CS/CMPE 537 – Neural Networks.
An Illustrative Example
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
Multilayer Perceptrons
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
CS 4700: Foundations of Artificial Intelligence
September 28, 2010Neural Networks Lecture 7: Perceptron Modifications 1 Adaline Schematic Adjust weights i1i1i1i1 i2i2i2i2 inininin …  w 0 + w 1 i 1 +
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Artificial neural networks:
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Where We’re At Three learning rules  Hebbian learning regression  LMS (delta rule) regression  Perceptron classification.
1 Mehran University of Engineering and Technology, Jamshoro Department of Electronic, Telecommunication and Bio-Medical Engineering Neural Networks Mukhtiar.
Artificial Neural Networks
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
ANNs (Artificial Neural Networks). THE PERCEPTRON.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 20 Oct 26, 2005 Nanjing University of Science & Technology.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
Linear Classification with Perceptrons
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
CS621 : Artificial Intelligence
Chapter 2 Single Layer Feedforward Networks
EEE502 Pattern Recognition
Neural Networks 2nd Edition Simon Haykin 柯博昌 Chap 3. Single-Layer Perceptrons.
Perceptrons Michael J. Watts
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Announcements 1. Textbook will be on reserve at library 2. Topic schedule change; modified reading assignment: This week: Linear discrimination, evaluating.
Lecture 12. Outline of Rule-Based Classification 1. Overview of ANN 2. Basic Feedforward ANN 3. Linear Perceptron Algorithm 4. Nonlinear and Multilayer.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Neural networks.
Fall 2004 Backpropagation CS478 - Machine Learning.
Chapter 2 Single Layer Feedforward Networks
Artificial neural networks:
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
CSSE463: Image Recognition Day 17
Neural Networks Advantages Criticism
Perceptron as one Type of Linear Discriminants
network of simple neuron-like computing elements
Neural Networks Chapter 5
Neural Network - 2 Mayank Vatsa
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Chapter - 3 Single Layer Percetron
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
Presentation transcript:

The Perceptron CS/CMPE 333 – Neural Networks

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS2 The Perceptron – Basics Simplest and one of the earliest neural network model proposed by Rosenblatt in 1958, 1962 It is based on the McCulloch-Pitts model of a neuron Characteristics  Single-layer feedforward network (A layer of input nodes and one layer of computation/output nodes)  Threshold activation function (or hard limiter function)  Performs classification of linearly separable patterns  Trained using error-correcting learning

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS3 The Perceptron Consider a single neuron perceptron

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS4 Linear Separability (1) The function of the single neuron perceptron is to classify the input x into one of two classes, C 1 and C 2. In general, a q neuron perceptron can classify the input x into 2 q classes For the two classes case, the decision boundary is defined by the hyperplane Σ i=1 p w i x i – Θ = 0

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS5 Linear Separability (2) When p = 2 (i.e. two inputs), the decision boundary is a line

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS6 Error-Correction Learning and Pattern Classification

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS7 Pattern Classification (1) Let n = number of training samples (set X); X 1 = set of training sample belonging to C 1 ; X 2 = set of training sample belonging to C 2 For a given sample n x(n) = [-1, x 1 (n),…, x p (n)] T = input vector w(n) = [Θ(n), w 1 (n),…, w p (n)] T = weight vector Net activity level v(n) = w T (n)x(n) Output y(n) = +1 if v(n) >= 0 and y(n) = -1 otherwise The decision hyperplane separates classes C 1 and C 2

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS8 Pattern Classification (2) If the two classes C 1 and C 2 are linearly separable, then there exists a weight vector w such that w T x ≥ 0 for all x belonging to class C 1 w T x < 0 for all x belonging to class C 2

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS9 Error-Correction Learning Update rule: w(n + 1) = w(n) + Δw(n) Learning process  If x(n) is correctly classified by w(n), then w(n + 1) = w(n)  Otherwise, the weight vector is updated as follows w(n + 1) = w(n) – η(n)x(n) if w(n) T x(n) ≥ 0 and x(n) belongs to C 2 and w(n + 1) = w(n) + η(n)x(n) if w(n) T x(n) < 0 and x(n) belongs to C 1

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS10 Perceptron Convergence Algorithm (1) Variables and parameters  x(n) = [-1, x 1 (n),…, x p (n)]; w(n) = [Θ(n), w 1 (n),…,w p (n)]  y(n) = actual response (output); d(n) = desired response  η = learning rate, a positive number less than 1 Step 1: Initialization  Set w(0) = 0, then do the following for n = 1, 2, 3, … Step 2: Activation  Activate the perceptron by applying input vector x(n) and desired output d(n) Step 3: Computation of actual response  y(n) = sgn[w T (n)x(n)]  Where sgn(.) is the signum function

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS11 Perceptron Convergence Algorithm (2) Step 4: Adaptation of weight vector w(n+1) = w(n) + η[d(n) – y(n)]x(n) Where d(n) = +1 if x(n) belongs to C 1 d(n) = -1 if x(n) belongs to C 2 Step 5  Increment n by 1, and go back to step 2

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS12 Performance Measure (1) A learning rule is designed to optimize a performance measure  However, in the development of the perceptron convergence algorithm we did not mention a performance measure Intuitively, what would be an appropriate performance measure for a classification neural network? Define the performance measure J = -E[e(n)v(n)] Or, as an instantaneous estimate J’(n) = -e(n)v(n)  e(n) = error at iteration n = d(n) – y(n); v(n) = linear combiner output at iteration n; E = expectation operator

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS13 Performance Measure (2) Can we derive our learning rule by minimizing this performance function… Now v(n) = w T (n)x(n), thus Learning rule

CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS14 Concluding Remarks A single layer perceptron can perform pattern classification only on linearly separable patterns, regardless of the type of nonlinearity (hard limiter, signoidal) Papert and Minsky in 1969 elucidated limitations of Rosenblatt’s single layer perceptron (e.g. requirement of linear separability, inability to solve XOR problem) and cast doubt on the viability of neural networks However, multilayer perceptron and the back- propagation algorithm overcomes many of the shortcomings of the single layer perceptron