Presentation is loading. Please wait.

Presentation is loading. Please wait.

November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation.

Similar presentations


Presentation on theme: "November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation."— Presentation transcript:

1 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation Network (BPN) Supervised Learning in the BPN The Self-Organizing Map (SOM) Unsupervised Learning in the SOM

2 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 2 The Backpropagation Network The backpropagation network (BPN) is the most popular type of ANN for applications such as classification or function approximation. Like other networks using supervised learning, the BPN is not biologically plausible. The structure of the network is as follows: Three layers of neurons, Three layers of neurons, Only feedforward processing: input layer  hidden layer  output layer, Only feedforward processing: input layer  hidden layer  output layer, Sigmoid activation functions Sigmoid activation functions

3 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 3 neuron i An Artificial Neuron o1o1 o2o2 onon … w i1 w i2 … w in oioi net input signal synapses output signal

4 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 4 Sigmoidal Neurons The output (spike frequency) of every neuron is simulated as a value between 0 (no spikes) and 1 (maximum frequency). 1 0 1 o i (t) o i (t) net i (t)  = 1  = 0.1

5 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 5 The Backpropagation Network BPN structure: input vector x I1I1I1I1 output vector o /desired output vector y I2I2I2I2 IIIIIIII H1H1H1H1 H2H2H2H2 H3H3H3H3 HJHJHJHJ O1O1O1O1 OKOKOKOK … … …

6 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 6 Supervised Learning in the BPN In supervised learning, we train an ANN with a set of vector pairs, so-called exemplars. Each pair (x, y) consists of an input vector x and a corresponding output vector y. Whenever the network receives input x, we would like it to provide output y. The exemplars thus describe the function that we want to “teach” our network. Besides learning the exemplars, we would like our network to generalize, that is, give plausible output for inputs that the network had not been trained with.

7 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 7 Supervised Learning in the BPN Before the learning process starts, all weights (synapses) in the network are initialized with pseudorandom numbers. We also have to provide a set of training patterns (exemplars). They can be described as a set of ordered vector pairs {(x 1, y 1 ), (x 2, y 2 ), …, (x P, y P )}. Then we can start the backpropagation learning algorithm. This algorithm iteratively minimizes the network’s error by finding the gradient of the error surface in weight-space and adjusting the weights in the opposite direction (gradient-descent technique).

8 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 8 Supervised Learning in the BPN Gradient-descent example: Finding the absolute minimum of a one-dimensional error function f(x): f(x)x x0x0x0x0 slope: f’(x 0 ) x 1 = x 0 -  f’(x 0 ) Repeat this iteratively until for some x i, f’(x i ) is sufficiently close to 0.

9 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 9 Supervised Learning in the BPN Gradients of two-dimensional functions: The two-dimensional function in the left diagram is represented by contour lines in the right diagram, where arrows indicate the gradient of the function at different locations. Obviously, the gradient is always pointing in the direction of the steepest increase of the function. In order to find the function’s minimum, we should always move against the gradient.

10 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 10 Supervised Learning in the BPN In the BPN, learning is performed as follows: 1.Randomly select a vector pair (x p, y p ) from the training set and call it (x, y). 2.Use x as input to the BPN and successively compute the outputs of all neurons in the network (bottom-up) until you get the network output o. 3.Compute the error of the network, i.e., the difference between the desired output y and the actual output o. 4.Apply the backpropagation learning rule to update the weights in the network so that its output o for input x is closer to the desired output y.

11 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 11 Supervised Learning in the BPN Repeat steps 1 to 4 for all vector pairs in the training set; this is called a training epoch. Run as many epochs as required to reduce the network error E to fall below a threshold that you set beforehand.

12 November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 12 Supervised Learning in the BPN Now our BPN is ready to go! If we choose the type and number of neurons in our network appropriately, after training the network should show the following behavior: If we input any of the training vectors, the network should yield the expected output vector (with some margin of error). If we input any of the training vectors, the network should yield the expected output vector (with some margin of error). If we input a vector that the network has never “seen” before, it should be able to generalize and yield a plausible output vector based on its knowledge about similar input vectors. If we input a vector that the network has never “seen” before, it should be able to generalize and yield a plausible output vector based on its knowledge about similar input vectors.


Download ppt "November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation."

Similar presentations


Ads by Google