Neuro-Computing Lecture 2 Single-Layer Perceptrons

Slides:



Advertisements
Similar presentations
Perceptron Lecture 4.
Advertisements

Slides from: Doug Gray, David Poole
Introduction to Neural Networks Computing
G53MLE | Machine Learning | Dr Guoping Qiu
Perceptron Learning Rule
NEURAL NETWORKS Perceptron
Artificial Neural Networks
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Perceptron.
Overview over different methods – Supervised Learning
Simple Neural Nets For Pattern Classification
The Perceptron CS/CMPE 333 – Neural Networks. CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS2 The Perceptron – Basics Simplest and one.
September 21, 2010Neural Networks Lecture 5: The Perceptron 1 Supervised Function Approximation In supervised learning, we train an ANN with a set of vector.
Perceptron Learning Rule
An Illustrative Example
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
Data Mining with Neural Networks (HK: Chapter 7.5)
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Artificial Neural Network
Neural Networks Lecture 8: Two simple learning algorithms
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
1 Mehran University of Engineering and Technology, Jamshoro Department of Electronic, Telecommunication and Bio-Medical Engineering Neural Networks Mukhtiar.
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
Linear Discrimination Reading: Chapter 2 of textbook.
Non-Bayes classifiers. Linear discriminants, neural networks.
Linear Classification with Perceptrons
ADALINE (ADAptive LInear NEuron) Network and
Chapter 2 Single Layer Feedforward Networks
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Supervised learning network G.Anuradha. Learning objectives The basic networks in supervised learning Perceptron networks better than Hebb rule Single.
EEE502 Pattern Recognition
Chapter 8: Adaptive Networks
Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.
Neural Networks 2nd Edition Simon Haykin 柯博昌 Chap 3. Single-Layer Perceptrons.
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
Perceptrons Michael J. Watts
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Neural networks.
Learning with Perceptrons and Neural Networks
Chapter 2 Single Layer Feedforward Networks
Artificial neural networks:
第 3 章 神经网络.
Ranga Rodrigo February 8, 2014
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Ch 2. Concept Map ⊂ ⊂ Single Layer Perceptron = McCulloch – Pitts Type Learning starts in Ch 2 Architecture, Learning Adaline : Linear Learning.
Dr. Unnikrishnan P.C. Professor, EEE
Data Mining with Neural Networks (HK: Chapter 7.5)
Classification Neural Networks 1
Chapter 3. Artificial Neural Networks - Introduction -
Neuro-Computing Lecture 4 Radial Basis Function Network
Artificial Intelligence Chapter 3 Neural Networks
Perceptron as one Type of Linear Discriminants
Neural Network - 2 Mayank Vatsa
Lecture 7. Learning (IV): Error correcting Learning and LMS Algorithm
Perceptrons Introduced in1957 by Rosenblatt
Neural Networks.
Artificial Intelligence Lecture No. 28
Artificial Intelligence Chapter 3 Neural Networks
Chapter - 3 Single Layer Percetron
Artificial Intelligence Chapter 3 Neural Networks
Computer Vision Lecture 19: Object Recognition III
Lecture 8. Learning (V): Perceptron Learning
Perceptron Learning Rule
Perceptron Learning Rule
Perceptron Learning Rule
Presentation transcript:

Neuro-Computing Lecture 2 Single-Layer Perceptrons 4/20/2019 Neuro-Computing Lecture 2 Single-Layer Perceptrons

Suggested Reading Perceptrons: - Haykin: Sections 3.5 - 3.9

Introduction A structure of a Perceptron is Input signals, xi, are assumed to have real values. The activation function, σ , is a unipolar step function, therefore, the output signal is binary: y ε { 0, 1 } .

Introduction The input signals and weights are arranged in the following column and row vectors, respectively: Aggregation of the “proper” input signals results in the activation potential, v, which can be expressed as the inner product of “proper” input signals and related weights:

For each input signal, the output is determined as: Introduction For each input signal, the output is determined as: Hence, a Perceptron works as a threshold element, the output being “active” if the activation potential exceeds the threshold.

Perceptron Architecture

A Perceptron as a Pattern Classifier A single Perceptron classifies input patterns, x, into two classes. A linear combination of signals and weights for which the augmented activation potential is zero, v=0, describes a decision surface which partitions the input space into two regions. The decision surface is a hyperplane specified as: An input vector (pattern)

A Perceptron as a Pattern Classifier The input patterns that can be classified by a single Perceptron into two distinct classes are called linearly separable patterns.

Example — a three-synapse Perceptron Consider a three-input Perceptron with weights w = [2 3 -6] The total (augmented) activation potential is determined as:

Example — a three-synapse Perceptron The input space, [x1 x2], is 2-dimensional and the decision plane is reduced to a straight line: The 2-D weight vector , is perpendicular to the decision line, that is, to the vector which is parallel to (lies in) the decision line, because the inner product , that is:

Selection of weights for the Perceptron In order for the Perceptron to perform a desired task, its weights must be properly selected. In general two basic methods can be employed to select a suitable weight vector: -By off- line calculation of weights. -By learning procedure.

Selection of weights for the Perceptron Off- line calculation of weights -If the problem is relatively simple it is often possible to calculate the weight vector from the specification of the problem. Learning procedure -The weight vector is determined from a given (training) set of input- output vectors (exemplars) in such a way to achieve the best classification of the training vectors. Once the weights are selected (the encoding process), the Perceptron performs its desired task (e. g., pattern classification) (the decoding process).

Selection of weights by off-line calculations — Example Consider the problem of building the NAND gate using the Perceptron. In this case the desired input output relationship is specified by the following truth table and related plot: The input patterns (vectors) belong to two classes and are marked in the input space (the plane (x1,x2) ) with Δ and ◎, respectively.

Selection of weights by off-line calculations — Example The decision plane is the straight line described by the following equation: In order for the weight vector to point in the direction of y = 1 patterns we select it as: For each input vector the output signal is now calculated as: Verify that such a Perceptron does implement a two- input NAND gate.

The Perceptron & learning Learning is a recursive procedure of modifying weights from a given set of input- output patterns. For a single Perceptron, the objective of the learning (encoding) procedure is to find the decision plane, (that is, the related weight vector), which separates two classes of given input- output training vectors. Once the learning is finalized, every input vector will be classified into an appropriate class. A single Perceptron can classify only the linearly separable patterns. The Perceptron learning procedure is an example of a supervised error- correcting learning law.

The Perceptron & learning More specifically, the supervised learning procedure can be described as follows given the set of N training patterns consisting of input signals x( n) and the related desired output signals, d( n): obtain the correct decision plane specified by the weight vector w.

The training set The training patterns are arranged in a training set which consists of a p × N input matrix, X, and an N-element output vector, D: We can assume that an untrained Perceptron can generate an incorrect output y( n) ≠ d( n) due to incorrect values of weights. We can think about the actual output y( n) as an “estimate” of the correct, desired output d( n).

The Perceptron learning law Project # 1: part_1 OR gate, Initial weights: W=[.7;-.3;0.5]; LR=0.2; N_epochs=15 The Perceptron learning law The Perceptron learning law proposed by Rosenblatt in 1959 also known as the Perceptron convergence procedure, can be described as follows: 1. weights are initially set to “small” random values w(0) = rand 2. for n = 1, 2, . . . , N training input vectors, x(n), are presented to the Perceptron and the actual output y(n), is compared with the desired output, d(n), and the error ε(n) is calculated as follows: Note that because

The Perceptron learning law 3. at each step, the weights are adapted as follows [Least-Mean-Square(LMS) Algorithm]: where 0 < η < 1 is a learning parameter (gain), which controls the adaptation rate. The gain must be adjusted to ensure the convergence of the algorithm. Note that the weight update is zero if the error ε(n) = 0. For convenience the weight update equation (3) is presented in this table:

The Perceptron learning law The sketch of the geometric relationships involved in the above algorithm is presented in the following Figure. we can identify a current weight vector, w(n), the next weight vector, w(n+1), and the correct weight vector, w*. Related decision planes are orthogonal to these vectors and are depicted as straight lines.

Convergence of the Perceptron learning law During the learning process the current weight vector w(n) is modified in the direction of the current input vector x(n), if the input pattern is misclassified, that is, if the error is non-zero. Presenting the Perceptron with enough training vectors, the weight vector w(n) will tend to the correct value w*. Rosenblatt proved that if input patterns are linearly separable, then the Perceptron learning law converges, and the hyperplane separating two classes of input patterns can be determined.

Least-Mean-Square(LMS) Algorithm From Adaptive filters viewpoint: “Minimize the cost function E(w) with respect to the weight vector w” The necessary condition for optimality is: where is the gradient operator Accordingly, the steepest descent algorithm: Where is a positive constant called the step size or learning-rate parameter, and g(n) is the gradient vector evaluated at the point w(n)

Least-Mean-Square(LMS) Algorithm The least-mean-square (LMS) algorithm is based on the use of instantaneous values for the cost function, namely, Where e(n) is the error signal measured at time n the LMS algorithm operates with a linear neuron so we may express the error signal as Therefor

Apples/Bananas Sorter

Prototype Vectors

Apple / Banana Example

Apple / Banana Example, Second iteration

Apple / Banana Example, Check Project # 1: part_2 XOR gate, Initial weights: W=[.7;-.3;0.5]; LR=0.1; N_epochs=20 Apple / Banana Example, Check