Ch 2. Concept Map ⊂ ⊂ Single Layer Perceptron = McCulloch – Pitts Type Learning starts in Ch 2 Architecture, Learning Adaline : Linear Learning.

Slides:



Advertisements
Similar presentations
Aula 3 Single Layer Percetron
Advertisements

Slides from: Doug Gray, David Poole
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Introduction to Neural Networks Computing
NEURAL NETWORKS Perceptron
Artificial Neural Networks
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
CS Perceptrons1. 2 Basic Neuron CS Perceptrons3 Expanded Neuron.
Perceptron.
Overview over different methods – Supervised Learning
Performance Optimization
Simple Neural Nets For Pattern Classification
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
The Perceptron CS/CMPE 333 – Neural Networks. CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS2 The Perceptron – Basics Simplest and one.
Before we start ADALINE
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
Data Mining with Neural Networks (HK: Chapter 7.5)
September 28, 2010Neural Networks Lecture 7: Perceptron Modifications 1 Adaline Schematic Adjust weights i1i1i1i1 i2i2i2i2 inininin …  w 0 + w 1 i 1 +
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Where We’re At Three learning rules  Hebbian learning regression  LMS (delta rule) regression  Perceptron classification.
1 Mehran University of Engineering and Technology, Jamshoro Department of Electronic, Telecommunication and Bio-Medical Engineering Neural Networks Mukhtiar.
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Artificial Intelligence Chapter 3 Neural Networks Artificial Intelligence Chapter 3 Neural Networks Biointelligence Lab School of Computer Sci. & Eng.
Non-Bayes classifiers. Linear discriminants, neural networks.
ADALINE (ADAptive LInear NEuron) Network and
Chapter 2 Single Layer Feedforward Networks
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
SUPERVISED LEARNING NETWORK
Supervised learning network G.Anuradha. Learning objectives The basic networks in supervised learning Perceptron networks better than Hebb rule Single.
IE 585 History of Neural Networks & Introduction to Simple Learning Rules.
Perceptrons Michael J. Watts
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Machine Learning Supervised Learning Classification and Regression
Fall 2004 Backpropagation CS478 - Machine Learning.
Learning with Perceptrons and Neural Networks
Chapter 2 Single Layer Feedforward Networks
One-layer neural networks Approximation problems
Artificial neural networks:
第 3 章 神经网络.
Real Neurons Cell structures Cell body Dendrites Axon
Pipelined Adaptive Filters
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Data Mining with Neural Networks (HK: Chapter 7.5)
Biological and Artificial Neuron
Classification Neural Networks 1
Chapter 3. Artificial Neural Networks - Introduction -
Biological and Artificial Neuron
Artificial Intelligence Chapter 3 Neural Networks
Perceptron as one Type of Linear Discriminants
Ch2: Adaline and Madaline
Biological and Artificial Neuron
Neural Networks Chapter 5
Artificial Neural Networks
Lecture 7. Learning (IV): Error correcting Learning and LMS Algorithm
Lecture Notes for Chapter 4 Artificial Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Chapter - 3 Single Layer Percetron
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Ch 2. Concept Map ⊂ ⊂ Single Layer Perceptron = McCulloch – Pitts Type Learning starts in Ch 2 Architecture, Learning Adaline : Linear Learning Perceptron : Nonlinear Learning Real valued output Binary output Binary output Objective : W x > 0 , x ω₁ W x < 0 , x ω₂ T ⊂ T T Objective : W x = Any Desired Real Value Can be used for Both Classification / Regression Objective : W x = 1 +- T ⊂ Correction for Correct Sign, too No correction for correct sign Always Finds Perfect hyperplanes for Linearly Separable Cases All Patterns x Move Hyperplanes by an Amount Proportional to x Finds Optimum Hyperplanes for Linearly Nonseparable Cases Error Patterns Push & Pull Hyperplanes

Chapter 2. Single Layer Perceptron Ref. Perceptron, Madaline and Backpropagation, Widrow, Proc.IEEE 90. Rosenblatt - Perceptron Learning 1 2 D Adjusted Training Set Adjustable Weights x (n) 1 ) ( 1 n Adder s(n) 2 x (n) ) ( n y ) ( 2 n å M M M (n)=Bias D+1 x (n) D 1 (n) D + ) ( n e - Error signal ) ( n d Desired response

α = Learning Rate that scales x. Perceptron Learning – Error Correction Learning Initialize to zero or some random vector. 2) Input and Adjust the weights as follows: . α = Learning Rate that scales x. Rosenblatt set α = 1. The choice does not affect stability but it affects convergence time only if w(0) ≠ 0. A. Case 1: d(n) – y(n) = 1 or d(n) = 1 and y(n) =0 ; y(n+1) y(n) B. Case 2: d(n) – y(n) = –1 or d(n) = 0 and y(n) = 1 ; C. Case 3: d(n) = y(n) Do nothing. 3) Repeat 1)2) until No Classification Error for a Linearly Separable Case.

(2) Patterns of Classification Error - push and pull the decision hyperplanes. 1 ( x D ( 1 ) ) 1 ( w ) 2 ( w ) 2 ( D Same as (1) except

(3) Example – OR Learning x(1) pulls down to , x(2) pulls up to . 1 2 2 3 or w (0) X = 0 T 1 : Error at x(1) 1 w (1) X = 0 2 T w (2) X = 0 T 3 Error at x(2) – Overcorrection 3 2 No Error for All Patterns– Terminate Learning

Of AND Function

2. Adaptive Linear Element – ADALINE = Linear Neuron If d(n) = Real, Regression. If d(n) = Integer (0/1 or -1/+1), Classification. Perceptron Adaline s(n) y(n) x(n) - Perceptron error å D + D e(n) d(n)

- LMS Learning Rule – Normalized, controls stability and speed of conv. In general, Data Set is not completely specified in advance. LMS is good for training data stream drawn from stationary distribution at least for x(n) In practice 0.1 < <1 cf. -LMS : Instantaneous Gradient Descent

(2) Minimum Disturbance Principle – Adapt to reduce the Output Error for the Current Training Pattern, with Minimal disturbance to Responses already learned x(n) w(n+1) D w(n) w(n)

(3) Learning Rate Scheduling (Annealing) α, μ = Learning Rate, Step Size μ(n+1) = μ(n) – β or μ(n) = c/n or μ0 / [1 + n/τ] The Perceptron Learning Rule can also be derived from the LMS-Rule :

Perceptron Learning LMS Learning Works with binary (bipolar) outputs Works with both binary and analog Always converges for linearly separable cases - Theorem May not converge even for linearly separable cases May be unstable for linearly nonseparable cases – Use pocket Alg. Also works well for linearly nonseparable cases – finds minimum error solution No correction for correct classification Always corrects Objective : > 0 or < 0 Objective: = 1 or − 1 Nonlinear rule Linear rule

Graphical Representation of Learning For classification with Adaline, one could just stop learning when the signs are all correct – way to speed up Adaline learning. Perceptron ω 1 1 1 ω 1 Adaline After Learning -1 -1 ω 2 ω 2

To Noise Cancellation, Equalization, Echo Cancellation (4) ADALINE To Noise Cancellation, Equalization, Echo Cancellation

Student Questions -2005 Adaline Learning takes longer than Perceptron in general. But, still it may have an edge over Perceptron in some aspect ? Diff. between α–LMS and μ–LMS. Self-normalizing vs. const. Latter always converges in the mean to the minimum MSE solution. Anyway to learn without desired outputs ? What does it mean when Perceptron learning can be derived from LMS using J = … Not clear on Regression. How can we guarantee a whole system convergence ? Which of Perceptron and Adaline performs better, is more efficient ?