Adaptive Resonance Theory (ART) networks perform completely unsupervised learning. Their competitive learning algorithm is similar to the first (unsupervised)

Slides:



Advertisements
Similar presentations
NEURAL NETWORKS Backpropagation Algorithm
Advertisements

Artificial Neural Networks (1)
Marković Miljan 3139/2011
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
B.Macukow 1 Lecture 12 Neural Networks. B.Macukow 2 Neural Networks for Matrix Algebra Problems.
Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Self Organization: Competitive Learning
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Kohonen Self Organising Maps Michael J. Watts
Unsupervised Learning with Artificial Neural Networks The ANN is given a set of patterns, P, from space, S, but little/no information about their classification,
Artificial neural networks:
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Simple Neural Nets For Pattern Classification
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Radial Basis Functions
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation.
September 30, 2010Neural Networks Lecture 8: Backpropagation Learning 1 Sigmoidal Neurons In backpropagation networks, we typically choose  = 1 and 
November 9, 2010Neural Networks Lecture 16: Counterpropagation 1 Unsupervised Learning So far, we have only looked at supervised learning, in which an.
September 14, 2010Neural Networks Lecture 3: Models of Neurons and Neural Networks 1 Visual Illusions demonstrate how we perceive an “interpreted version”
An Illustrative Example
-Artificial Neural Network- Adaptive Resonance Theory(ART) 朝陽科技大學 資訊管理系 李麗華 教授.
September 16, 2010Neural Networks Lecture 4: Models of Neurons and Neural Networks 1 Capabilities of Threshold Neurons By choosing appropriate weights.
Chapter 6: Multilayer Neural Networks
Before we start ADALINE
ART (Adaptive Resonance Theory)
November 24, 2009Introduction to Cognitive Science Lecture 21: Self-Organizing Maps 1 Self-Organizing Maps (Kohonen Maps) In the BPN, we used supervised.
Neural Networks Lecture 17: Self-Organizing Maps
Lecture 09 Clustering-based Learning
November 21, 2012Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms III 1 Learning in the BPN Gradients of two-dimensional functions:
Radial Basis Function Networks
Artificial Neural Networks
Artificial Neural Network Unsupervised Learning
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
NEURAL NETWORKS FOR DATA MINING
Adaptive Resonance Theory
Chapter 5. Adaptive Resonance Theory (ART) ART1: for binary patterns; ART2: for continuous patterns Motivations: Previous methods have the following problem:
Soft Computing Lecture 14 Clustering and model ART.
Ming-Feng Yeh1 CHAPTER 16 AdaptiveResonanceTheory.
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
Neural Networks Steven Le. Overview Introduction Architectures Learning Techniques Advantages Applications.
CSE & CSE6002E - Soft Computing Winter Semester, 2011 Neural Networks Videos Brief Review The Next Generation Neural Networks - Geoff Hinton.
1 Adaptive Resonance Theory. 2 INTRODUCTION Adaptive resonance theory (ART) was developed by Carpenter and Grossberg[1987a] ART refers to the class of.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
381 Self Organization Map Learning without Examples.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
November 21, 2013Computer Vision Lecture 14: Object Recognition II 1 Statistical Pattern Recognition The formal description consists of relevant numerical.
Perceptrons Michael J. Watts
Intelligent Database Systems Lab Advisor : Dr. Hsu Graduate : Yu Cheng Chen Author : Yongqiang Cao Jianhong Wu 國立雲林科技大學 National Yunlin University of Science.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Soft Computing Lecture 15 Constructive learning algorithms. Network of Hamming.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
April 5, 2016Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II 1 Capabilities of Threshold Neurons By choosing appropriate.
Chapter 5 Unsupervised learning
Self-Organizing Network Model (SOM) Session 11
Adaptive Resonance Theory (ART)
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
Counter propagation network (CPN) (§ 5.3)
Dr. Unnikrishnan P.C. Professor, EEE
of the Artificial Neural Networks.
Adaptive Resonance Theory
Capabilities of Threshold Neurons
The Naïve Bayes (NB) Classifier
Adaptive Resonance Theory
Dr. Unnikrishnan P.C. Professor, EEE
Presentation transcript:

Adaptive Resonance Theory (ART) networks perform completely unsupervised learning. Their competitive learning algorithm is similar to the first (unsupervised) phase of CPN learning. However, ART networks are able to grow additional neurons if a new input cannot be categorized appropriately with the existing neurons. A vigilance parameter  determines the tolerance of this matching process. A greater value of  leads to more, smaller clusters (= input samples associated with the same winner neuron). Adaptive Resonance Theory (ART) November 23, 20101Neural Networks Lecture 19: Adaptive Resonance Theory

ART networks consist of an input layer and an output layer. We will only discuss ART-1 networks, which receive binary input vectors. Bottom-up weights are used to determine output-layer candidates that may best match the current input. Top-down weights represent the “prototype” for the cluster defined by each output neuron. A close match between input and prototype is necessary for categorizing the input. Finding this match can require multiple signal exchanges between the two layers in both directions until “resonance” is established or a new neuron is added. Adaptive Resonance Theory (ART) November 23, 20102Neural Networks Lecture 19: Adaptive Resonance Theory

ART networks tackle the stability-plasticity dilemma: Plasticity: They can always adapt to unknown inputs (by creating a new cluster with a new weight vector) if the given input cannot be classified by existing clusters. Stability: Existing clusters are not deleted by the introduction of new inputs (new clusters will just be created in addition to the old ones). Problem: Clusters are of fixed size, depending on . Adaptive Resonance Theory (ART) November 23, 20103Neural Networks Lecture 19: Adaptive Resonance Theory

Input layer Output layer with inhibitory connections The ART-1 Network November 23, 20104Neural Networks Lecture 19: Adaptive Resonance Theory

A.Initialize each top-down weight t l,j (0) = 1; B.Initialize bottom-up weight b j,l (0) = ; C. While the network has not stabilized, do 1.Present a randomly chosen pattern x = (x 1,…,x n ) for learning 2.Let the active set A contain all nodes; calculate y j = b j,1 x 1 +…+b j,n x n for each node j A; 3. Repeat a)Let j* be a node in A with largest y j, with ties being broken arbitrarily; b)Compute s * = (s * 1,…,s * n ) where s * l = t l,j* x l ; c)Compare similarity between s* and x with the given vigilance parameter   if <  then remove j* from set A else associate x with node j* and update weights: b j*l (new) = t l,j* (new) = Until A is empty or x has been associated with some node j 4. If A is empty, then create new node whose weight vector coincides with current input pattern x; end-while November 23, 20105Neural Networks Lecture 19: Adaptive Resonance Theory

For this example, let us assume that we have an ART-1 network with 7 input neurons (n = 7) and initially one output neuron (n = 1). Our input vectors are {(1, 1, 0, 0, 0, 0, 1), (0, 0, 1, 1, 1, 1, 0), (1, 0, 1, 1, 1, 1, 0), (0, 0, 0, 1, 1, 1, 0), (1, 1, 0, 1, 1, 1, 0)} and the vigilance parameter  = 0.7. Initially, all top-down weights are set to t l,1 (0) = 1, and all bottom-up weights are set to b 1,l (0) = 1/8. ART Example Computation November 23, 20106Neural Networks Lecture 19: Adaptive Resonance Theory

For the first input vector, (1, 1, 0, 0, 0, 0, 1), we get: Clearly, y 1 is the winner (there are no competitors). Since we have: the vigilance condition is satisfied and we get the following new weights: ART Example Computation November 23, 20107Neural Networks Lecture 19: Adaptive Resonance Theory

Also, we have: We can express the updated weights as matrices: Now we have finished the first learning step and proceed by presenting the next input vector. ART Example Computation November 23, 20108Neural Networks Lecture 19: Adaptive Resonance Theory

For the second input vector, (0, 0, 1, 1, 1, 1, 0), we get: Of course, y 1 is still the winner. However, this time we do not reach the vigilance threshold: This means that we have to generate a second node in the output layer that represents the current input. Therefore, the top-down weights of the new node will be identical to the current input vector. ART Example Computation November 23, 20109Neural Networks Lecture 19: Adaptive Resonance Theory

The new unit’s bottom-up weights are set to zero in the positions where the input has zeroes as well. The remaining weights are set to: 1/( ) This gives us the following updated weight matrices: ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

For the third input vector, (1, 0, 1, 1, 1, 1, 0), we have: Here, y 2 is the clear winner. This time we exceed the vigilance threshold again: Therefore, we adapt the second node’s weights. Each top-down weight is multiplied by the corresponding element of the current input. ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

The new unit’s bottom-up weights are set to the top- down weights divided by ( ). It turns out that, in the current case, these updates do not result in any weight changes at all: ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

For the fourth input vector, (0, 0, 0, 1, 1, 1, 0), it is: Again, y 2 is the winner. The vigilance test succeeds once again: Therefore, we adapt the second node’s weights. As usual, each top-down weight is multiplied by the corresponding element of the current input. ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

The new unit’s bottom-up weights are set to the top- down weights divided by ( ). This gives us the following new weight matrices: ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

Finally, the fifth input vector, (1, 1, 0, 1, 1, 1, 0), gives us: Once again, y 2 is the winner. The vigilance test fails this time: This means that the active set A is reduced to contain only the first node, which becomes the uncontested winner. ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

The vigilance test fails for the first unit as well: We thus have to create a third output neuron, which gives us the following new weight matrices: ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

In the second epoch, the first input vector, (1, 1, 0, 0, 0, 0, 1), gives us: Here, y 1 is the winner, and the vigilance test succeeds: Since the current input is identical to the winner’s top- down weights, no weight update happens. ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

The second input vector, (0, 0, 1, 1, 1, 1, 0), results in: Now y 2 is the winner, and the vigilance test succeeds: Again, because the current input is identical to the winner’s top-down weights, no weight update occurs. ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

The third input vector, (1, 0, 1, 1, 1, 1, 0), give us: Once again, y 2 is the winner, but this time the vigilance test fails: This means that the active set is reduced to A = {1, 3}. Since y 3 > y 1, the third node is the new winner. ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

The third node does satisfy the vigilance threshold: This gives us the following updated weight matrices: ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

For the fourth vector, (0, 0, 0, 1, 1, 1, 0), the second node wins, passes the vigilance test, but no weight changes occur. The fifth vector, (1, 1, 0, 1, 1, 1, 0), makes the second unit win, which fails the vigilance test. The new winner is the third output neuron, which passes the vigilance test but does not lead to any weight modifications. Further presentation of the five sample vectors do not lead to any weight changes; the network has thus stabilized. ART Example Computation November 23, Neural Networks Lecture 19: Adaptive Resonance Theory

Adaptive Resonance Theory November 23, Neural Networks Lecture 19: Adaptive Resonance Theory Illustration of the categories (or clusters) in input space formed by ART networks. Notice that increasing  leads to narrower cones and not to wider ones as suggested by the figure.

A problem with ART-1 is the need to determine the vigilance parameter  for a given problem, which can be tricky. Furthermore, ART-1 always builds clusters of the same size, regardless of the distribution of samples in the input space. Nevertheless, ART is one of the most important and successful attempts at simulating incremental learning in biological systems. Adaptive Resonance Theory November 23, Neural Networks Lecture 19: Adaptive Resonance Theory