A Review: Architecture

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Perceptron Lecture 4.
Introduction to Neural Networks Computing
G53MLE | Machine Learning | Dr Guoping Qiu
Artificial Neural Networks (1)
NEURAL NETWORKS Perceptron
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
CS Perceptrons1. 2 Basic Neuron CS Perceptrons3 Expanded Neuron.
Neural Networks (II) Simple Learning Rule
Overview over different methods – Supervised Learning
Simple Neural Nets For Pattern Classification
The back-propagation training algorithm
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
PERCEPTRON. Chapter 3: The Basic Neuron  The structure of the brain can be viewed as a highly interconnected network of relatively simple processing.
The Perceptron CS/CMPE 333 – Neural Networks. CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS2 The Perceptron – Basics Simplest and one.
20.5 Nerual Networks Thanks: Professors Frank Hoffmann and Jiawei Han, and Russell and Norvig.
September 21, 2010Neural Networks Lecture 5: The Perceptron 1 Supervised Function Approximation In supervised learning, we train an ANN with a set of vector.
September 14, 2010Neural Networks Lecture 3: Models of Neurons and Neural Networks 1 Visual Illusions demonstrate how we perceive an “interpreted version”
Perceptron Learning Rule
An Illustrative Example
Perceptron Learning Rule Assuming the problem is linearly separable, there is a learning rule that converges in a finite time Motivation A new (unseen)
Before we start ADALINE
Data Mining with Neural Networks (HK: Chapter 7.5)
CS 4700: Foundations of Artificial Intelligence
COMP305. Part I. Artificial neural networks.. Topic 3. Learning Rules of the Artificial Neural Networks.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CS532 Neural Networks Dr. Anwar Majid Mirza Lecture No. 3 Week2, January 22 nd, 2008 National University of Computer and Emerging.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Where We’re At Three learning rules  Hebbian learning regression  LMS (delta rule) regression  Perceptron classification.
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Networks
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
DIGITAL IMAGE PROCESSING Dr J. Shanbehzadeh M. Hosseinajad ( J.Shanbehzadeh M. Hosseinajad)
Chapter 9 Neural Network.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Techniques Multilayer Perceptrons.
1 Pattern Classification X. 2 Content General Method K Nearest Neighbors Decision Trees Nerual Networks.
IE 585 Associative Network. 2 Associative Memory NN Single-layer net in which the weights are determined in such a way that the net can store a set of.
The Perceptron. Perceptron Pattern Classification One of the purposes that neural networks are used for is pattern classification. Once the neural network.
Feed-Forward Neural Networks 主講人 : 虞台文. Content Introduction Single-Layer Perceptron Networks Learning Rules for Single-Layer Perceptron Networks – Perceptron.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 8: Neural Networks.
Linear Discrimination Reading: Chapter 2 of textbook.
Linear Classification with Perceptrons
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Chapter 2 Single Layer Feedforward Networks
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Supervised learning network G.Anuradha. Learning objectives The basic networks in supervised learning Perceptron networks better than Hebb rule Single.
November 21, 2013Computer Vision Lecture 14: Object Recognition II 1 Statistical Pattern Recognition The formal description consists of relevant numerical.
Perceptrons Michael J. Watts
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
Start with student evals. What function does perceptron #4 represent?
CS532 Neural Networks Dr. Anwar Majid Mirza Lecture No. 2 January 17 th, 2008 National University of Computer and Emerging Sciences.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Neural networks.
Chapter 2 Single Layer Feedforward Networks
Real Neurons Cell structures Cell body Dendrites Axon
Hebb and Perceptron.
Neural Networks Chapter 5
Neural Networks.
The Naïve Bayes (NB) Classifier
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

A Review: Architecture Basic architecture : Input layer, output layer Input Unit Output Unit 1 b X1 w1 Y wi Xi wn Xn Single-layer net for pattern classification

Biases & Thresholds  xiwi if net  0 -1 if net < 0 A bias acts exactly as a weight on a connection from a unit whose activation is always 1 Increasing the bias increases the net input to the unit If a bias is included, the activation function is typically taken to be f(net) = if net  0 -1 if net < 0 Where net = b +  xiwi i

Biases & Thresholds  xiwi Some do not use bias weight, but use a fixed threshold  for the activation function if net   -1 if net <  Where net =  xiwi f(net) = i

Biases & Thresholds Input Unit Output Unit 1 b X1 w1 Y w2 X2 Single-layer net for pattern classification

Linear Separable For a particular output unit, the desired response is a ‘yes’(output signal 1) if the input pattern is the member of its class and ‘no’(output signal –1) We want one of two responses Activation function – step function From the net input formula we can draw a line, a plane or hyperplane y_in = b +  xi wi b +  xi wi = 0

Linear Separable b +  xi wi = 0 The problem is said ‘linearly separable’ if there are weights (and a bias) so that all of the training input vectors for which the correct response is +1 lie on one side of the decision boundary and all of the training input vectors for which the correct response is –1 lie on the other side Draw the decision boundary using this equation b +  xi wi = 0

Linear Separability + - x1 - -

Linear Separability + + x1 - +

Linear Separability - + x1 - +

Linear Separability In the case of an elementary perceptron, the n-dimensional space is divided by a hyperplane into two decision regions. The hyperplane is defined by the linearly separable function:

Linear Separability

Linear Separability A perceptron can learn the operations AND and OR, but not Exclusive-OR.

The HEBB NET The earliest and simplest learning rule Proposed that learning occurs by modification of the synapse strengths (weights). If two interconnected neurons are both “on” or “off” at the same time, then the weight between those neurons should be increased. Hebb net also used for training other specific nets

The HEBB NET A single layer nets Interconnected neurons will be between an input unit and one output unit. Suit for bipolar form of data (1, -1) Limitation for binary data (see examples 2.5 & 2.6)

The HEBB NET A single layer nets Input Unit Output Unit 1 b X1 w1 Y w2 Single-layer net for pattern classification Output Unit Input Unit

The HEBB NET bipolar form of data (1, -1) Weight update wi(new) = wi(old) + xi y

The Algorithm Step 0: Initialize all weights: wi = 0 (i= 1 to n) X1 X2 Y b w2 Single-layer net for pattern classification Output Unit Input Unit Step 0: Initialize all weights: wi = 0 (i= 1 to n) Step 1: For each input training vector and target output pair, s:t, do steps 2-4 Step 2. Set activations for input units: xi = si (i = 1 to n). Step 3. Set activation for output unit: y = t (t= target) Step 4. Adjust the weights for wi(new) = wi(old) + wi (i = 1 to n).  wi =xi y Adjust the bias: b(new) = b(old) + y.

The Application Hebb net for And function: binary inputs and binary targets Hebb net for And function: binary inputs and bipolar targets Hebb net for And function: bipolar inputs and bipolar targets

Hebb net for And function: binary inputs and binary targets w1 1 X1 X2 Y b w2 Output Unit Input Unit Input Target   (x1 x2 1) (1 1 1) (1 0 1) (0 1 1) (0 0 1) 1 w1 = x1t w2 = x2 t b = t Weight Changes Weights (1 1 1) (1 0 1) (0 1 1) (0 0 1) ( w1 w2 b) (1 1 1) (0 0 0) (w1 w2 b) (0 0 0)

Hebb net for And function: binary inputs and binary targets Weight Changes Weights (x1 x2 1)   (1 1 1) (1 0 1) (0 1 1) (0 0 1) 1 ( w1 w2 b) (1 1 1) (0 0 0) (w1 w2 b) (0 0 0) The response of the net correct for the first input pattern but not for the 2, 3, 4th pattern because the target values is 0, no learning occurs. Using binary target values prevents the net from learning any pattern for which the target is “off”

How a net learns By adjusting weights If no weight adjusted :: no learning occurs w1 1 X1 X2 Y b w2 Output Unit Input Unit

Hebb net for And function: binary inputs and binary targets Weight Changes Weights (x1 x2 1)   (1 1 1) (1 0 1) (0 1 1) (0 0 1) 1 ( w1 w2 b) (1 1 1) (0 0 0) (w1 w2 b) (0 0 0) No Learning occur

Hebb net for And function: binary inputs and binary targets x2 w1 1 X1 X2 Y b w2 x1 -1 + b +  xi wi = 1 + x1(1)+ x2(1)=0 Separating lines x2= -x1-1 Basic formula to draw separating line b +  xi wi = 0

Hebb net for And function: binary inputs and bipolar targets   (x1 x2 1) (1 1 1) (1 0 1) (0 1 1) (0 0 1) 1 -1 w1 = x1t w2 = x2 t b = t Weight Changes Weights (1 1 1) (1 0 1) (0 1 1) (0 0 1) ( w1 w2 b) (1 1 1) (-1 0 -1) (0 -1 -1) (0 0 -1) (w1 w2 b) (0 0 0) (0 1 0) (0 0 -1) (0 0 -2) w1 1 X1 X2 Y b w2

Hebb net for And function: binary inputs and bipolar targets (x1 x2 1)   (1 1 1) (1 0 1) (0 1 1) (0 0 1) 1 -1 ( w1 w2 b) (1 1 1) (-1 0 -1) (0 -1 -1) (0 0 -1) (w1 w2 b) (0 0 0) (0 1 0) (0 0 -1) (0 0 -2) The response of the net correct for the first input pattern and for the 2, 3, 4th patterns shows that learning continues for each of these since the target value is now -1. However, these weights do not provide the correct response for the first input pattern

Hebb net for And function: binary inputs and bipolar targets x2 w1 1 X1 X2 Y b w2 - + -1 - - x1 -1 b +  xi wi = 1 + x1(1)+ x2(1)=0 Separating lines x2= -x1-1 Basic formula to draw separating line b +  xi wi = 0

Hebb net for And function: bipolar inputs and bipolar targets   (x1 x2 1) (1 1 1) (1 -1 1) (-1 1 1) (-1 -1 1) 1 -1 w1 = x1t w2 = x2 t b = t Weight Changes Weights (1 1 1) (1 -1 1) (-1 1 1) (-1 -1 1) ( w1 w2 b) (1 1 1) (-1 1 -1) (1 -1 -1) (1 1 -1) (w1 w2 b) (0 0 0) (1 1 1) (0 2 0) (1 1 -1) (2 2 -2) w1 1 X1 X2 Y b w2

Hebb net for And function: bipolar inputs and bipolar targets (x1 x2 1)   (1 1 1) (1 -1 1) (-1 1 1) (-1 -1 1) 1 -1 ( w1 w2 b) (1 1 1) (-1 1 -1) (1 -1 -1) (1 1 -1) (w1 w2 b) (0 0 0) (1 1 1) (0 2 0) (1 1 -1) (2 2 -2) The response of the net correct for the first input pattern and the 2, 3, 4th patterns. w1 1 X1 X2 Y b w2

Hebb net for And function: bipolar inputs and bipolar targets x2 - - 1 + -1 1 x1 - -1 - b +  xi wi = -2 + x1(2)+ x2(2)=0 Separating lines x2= -x1+1 Basic formula to draw separating line b +  xi wi = 0

Application: Character Recognition # . . . # . # . # . . . # . . . # . # . . # # # . # . . . # Pattern 1 Pattern 2 1 –1 –1 –1 1, -1 1 –1 1 –1, -1 –1 1 –1 –1, -1 1 –1 1 –1, 1 –1 –1 –1 1 -1 1 1 1 –1, 1 –1 –1 –1 1 1 –1 –1 –1 1, 1 –1 –1 –1 1, 1 –1 –1 –1 1, -1 1 1 1 -1