CSE4403 3.0 & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks Videos Brief Review The Next Generation Neural Networks - Geoff Hinton.

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
Hopefully a clearer version of Neural Network. I1 O2 O1 H1 H2I2.
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Neural Networks  A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
B.Macukow 1 Lecture 12 Neural Networks. B.Macukow 2 Neural Networks for Matrix Algebra Problems.
Simple Neural Nets For Pattern Classification
The back-propagation training algorithm
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
September 30, 2010Neural Networks Lecture 8: Backpropagation Learning 1 Sigmoidal Neurons In backpropagation networks, we typically choose  = 1 and 
November 9, 2010Neural Networks Lecture 16: Counterpropagation 1 Unsupervised Learning So far, we have only looked at supervised learning, in which an.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Hopefully a clearer version of Neural Network. With Actual Weights.
MLP Exercise (2006) Become familiar with the Neural Network Toolbox in Matlab Construct a single hidden layer, feed forward network with sigmoidal units.
November 21, 2012Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms III 1 Learning in the BPN Gradients of two-dimensional functions:
Image Compression Using Neural Networks Vishal Agrawal (Y6541) Nandan Dubey (Y6279)
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Traffic Sign Recognition Using Artificial Neural Network Radi Bekker
CS623: Introduction to Computing with Neural Nets (lecture-10) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
Appendix B: An Example of Back-propagation algorithm
CONTENTS:  Introduction  What is neural network?  Models of neural networks  Applications  Phases in the neural network  Perceptron  Model of fire.
NEURAL NETWORKS FOR DATA MINING
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
BACKPROPAGATION: An Example of Supervised Learning One useful network is feed-forward network (often trained using the backpropagation algorithm) called.
Multi-Layer Perceptron
Soft Computing Lecture 8 Using of perceptron for image recognition and forecasting.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
CS621 : Artificial Intelligence
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 32: sigmoid neuron; Feedforward.
EEE502 Pattern Recognition
C - IT Acumens. COMIT Acumens. COM. To demonstrate the use of Neural Networks in the field of Character and Pattern Recognition by simulating a neural.
Previous Lecture Perceptron W  t+1  W  t  t  d(t) - sign (w(t)  x)] x Adaline W  t+1  W  t  t  d(t) - f(w(t)  x)] f’ x Gradient.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Essential components of the implementation are:  Formation of the network and weight initialization routine  Pixel analysis of images for symbol detection.
Neural Networks Lecture 11: Learning in recurrent networks Geoffrey Hinton.
Multinomial Regression and the Softmax Activation Function Gary Cottrell.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Neural networks.
Multiple-Layer Networks and Backpropagation Algorithms
The Gradient Descent Algorithm
A Simple Artificial Neuron
CSE 473 Introduction to Artificial Intelligence Neural Networks
Neural Networks A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
CSE P573 Applications of Artificial Intelligence Neural Networks
CS621: Artificial Intelligence
Prof. Carolina Ruiz Department of Computer Science
Artificial Neural Network & Backpropagation Algorithm
CSE 573 Introduction to Artificial Intelligence Neural Networks
network of simple neuron-like computing elements
Artificial Intelligence 12. Two Layer ANNs
Artificial Neural Networks
CSC321: Neural Networks Lecture 11: Learning in recurrent networks
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
CS621: Artificial Intelligence Lecture 18: Feedforward network contd
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks Videos Brief Review The Next Generation Neural Networks - Geoff Hinton

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Brief Review

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm Most people would consider the Back Propagation network to be the quintessential Neural Net. Actually, Back Propagation is the training or learning algorithm rather than the network itself. Let’s consider what Back Propagation is and how to use it. A Back Propagation network learns by example. You give the algorithm examples of what you want the network to do and it changes the network’s weights so that, when training is finished, it will give you the required output for a particular input. Back Propagation networks are ideal for simple Pattern Recognition and Mapping Tasks. To train the network you need to give it examples of what you want - the output you want (called the Target) for a particular input.

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm Once the network is trained, it will provide the desired output for any of the input patterns. Let’s now look at how the training works. The network is first initialised by setting up all its weights to be small random numbers - say between -1 and +1. Next, the input pattern is applied and the output calculated (this is called the forward pass). The calculation gives an output which is completely different to what you want (the Target), since all the weights are random. We then calculate the Error of each neuron, which is essentially: Target - Actual Output (i.e., What you want - What you actually get). This error is then used mathematically to change the weights in such a way that the error will get smaller. In other words, the Output of each neuron will get closer to its Target (this part is called the reverse pass). The process is repeated again and again until the error is minimal.

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm Let's do an example with an actual network to see how the process works. We’ll just look at one connection initially, between a neuron in the output layer and one in the hidden layer. Single connection learning in a Back Propagation network. The connection we’re interested in is between neuron A (a hidden layer neuron) and neuron B (an output neuron) and has the weight W AB. The diagram also shows another connection, between neuron A and C, but we’ll return to that later. The algorithm works like this:

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm 1. First apply the inputs to the network and work out the output - remember this initial output could be anything, as the initial weights were random numbers. 2. Next work out the error for neuron B. The error is What you want - What you actually get, in other words: Error B = Output B (1-Output B )(Target B - Output B ) The “Output(1-Output)” term is necessary in the equation because of the Sigmoid Function - if we were only using a threshold neuron it would just be (Target - Output). 3. Change the weight. Let W+ AB be the new (trained) weight and W AB be the initial weight. W+ AB = W AB + (Error B x Output A ) Notice that it is the output of the connecting neuron (neuron A) we use (not B). We update all the weights in the output layer in this way.

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm 4. Calculate the Errors for the hidden layer neurons. Unlike the output layer we can’t calculate these directly (because we don’t have a Target), so we Back Propagate them from the output layer (hence the name of the algorithm). This is done by taking the Errors from the output neurons and running them back through the weights to get the hidden layer errors. For example if neuron A is connected as shown to B and C then we take the errors from B and C to generate an error for A. Error A = Output A (1 - Output A )(Error B W AB + ErrorC W AC ) Again, the factor “Output (1 - Output )” is present because of the sigmoid squashing function. 5. Having obtained the Error for the hidden layer neurons now proceed as in stage 3 to change the hidden layer weights. By repeating this method we can train a network of any number of layers.

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm This may well have left some doubt in your mind about the operation, so let’s clear that up by explicitly showing all the calculations for a full sized network with 2 inputs, 3 hidden layer neurons and 2 output neurons as shown below. W+ represents the new, recalculated, weight, whereas W represents the old weight.

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm - worked example

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm - worked example

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Back Propagation Algorithm

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks – Geoff Hinton’s Lecture Please view the 59 minute video at

CSE & CSE6002E - Soft Computing Winter Semester, 2011 Concluding Remarks T. T. T. Put up in a place where it's easy to see the cryptic admonishment T. T. T. When you feel how Depressingly slowly you climb, it's well to remember that Things Take Time.