Ming-Feng Yeh1 CHAPTER 16 AdaptiveResonanceTheory.

Slides:



Advertisements
Similar presentations
© Negnevitsky, Pearson Education, Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Advertisements

Multi-Layer Perceptron (MLP)
Introduction to Neural Networks Computing
Artificial Neural Networks (1)
Perceptron Learning Rule
Marković Miljan 3139/2011
Instar and Outstar Learning Laws Adapted from lecture notes of the course CN510: Cognitive and Neural Modeling offered in the Department of Cognitive and.
Ming-Feng Yeh1 CHAPTER 13 Associative Learning. Ming-Feng Yeh2 Objectives The neural networks, trained in a supervised manner, require a target signal.
Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Self Organization: Competitive Learning
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Kohonen Self Organising Maps Michael J. Watts
Artificial neural networks:
Adaptive Resonance Theory (ART) networks perform completely unsupervised learning. Their competitive learning algorithm is similar to the first (unsupervised)
The back-propagation training algorithm
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
1 Pertemuan 15 ADAPTIVE RESONANCE THEORY Matakuliah: H0434/Jaringan Syaraf Tiruan Tahun: 2005 Versi: 1.
An Illustrative Example
Instar Learning Law Adapted from lecture notes of the course CN510: Cognitive and Neural Modeling offered in the Department of Cognitive and Neural Systems.
Artificial Neural Networks Ch15. 2 Objectives Grossberg network is a self-organizing continuous-time competitive network.  Continuous-time recurrent.
Neural Networks based on Competition
ART (Adaptive Resonance Theory)
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Lecture 09 Clustering-based Learning
Ming-Feng Yeh1 CHAPTER 3 An Illustrative Example.
Supervised Hebbian Learning
© Negnevitsky, Pearson Education, Lecture 7 Artificial neural networks: Supervised learning Introduction, or how the brain works Introduction, or.
Artificial Neural Networks
Artificial Neural Network Unsupervised Learning
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
NEURAL NETWORKS FOR DATA MINING
Adaptive Resonance Theory
15 1 Grossberg Network Biological Motivation: Vision Eyeball and Retina.
George F Luger ARTIFICIAL INTELLIGENCE 6th edition Structures and Strategies for Complex Problem Solving Machine Learning: Connectionist Luger: Artificial.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Chapter 5. Adaptive Resonance Theory (ART) ART1: for binary patterns; ART2: for continuous patterns Motivations: Previous methods have the following problem:
Soft Computing Lecture 14 Clustering and model ART.
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
UNSUPERVISED LEARNING NETWORKS
1 Adaptive Resonance Theory. 2 INTRODUCTION Adaptive resonance theory (ART) was developed by Carpenter and Grossberg[1987a] ART refers to the class of.
381 Self Organization Map Learning without Examples.
Chapter 2 Single Layer Feedforward Networks
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
EEE502 Pattern Recognition
13 1 Associative Learning Simple Associative Network.
CHAPTER 14 Competitive Networks Ming-Feng Yeh.
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
Neural networks.
Chapter 5 Unsupervised learning
Adaptive Resonance Theory (ART)
Chapter 2 Single Layer Feedforward Networks
Pertemuan 7 JARINGAN INSTAR DAN OUTSTAR
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
Counter propagation network (CPN) (§ 5.3)
Dr. Unnikrishnan P.C. Professor, EEE
Associative Learning.
Grossberg Network.
Competitive Networks.
Adaptive Resonance Theory
Artificial Neural Networks
Clustering Techniques
Grossberg Network.
Competitive Networks.
Adaptive Resonance Theory
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Dr. Unnikrishnan P.C. Professor, EEE
Associative Learning.
Presentation transcript:

Ming-Feng Yeh1 CHAPTER 16 AdaptiveResonanceTheory

Ming-Feng Yeh2 Objectives There is no guarantee that, as more inputs are applied to the competitive network, the weight matrix will eventually converge. Present a modified type of competitive learning, called adaptive resonance theory (ART), which is designed to overcome the problem of learning stability.

Ming-Feng Yeh3 Theory & Examples A key problem of the Grossberg network and the competitive network is that they do NOT always from stable clusters (or categories). The learning instability occurs because of the network’s adaptability (or plasticity), which causes prior learning to be eroded by more recent learning.

Ming-Feng Yeh4 Stability / Plasticity How can a system be receptive to significant new patterns and yet remain stable in response to irrelevant patterns? Grossberg and Carpenter developed the ART to address the stability/plasticity dilemma. The ART networks are based on the Grossberg network of Chapter 15.

Ming-Feng Yeh5 Key Innovation The key innovation of ART is the use of “expectations.” As each input is presented to the network, it is compared with the prototype vector that is most closely matches (the expectation). If the match between the prototype and the input vector is NOT adequate, a new prototype is selected. In this way, previous learned memories (prototypes) are not eroded by new learning.

Ming-Feng Yeh6 Overview Basic ART architecture Grossberg competitive network

Ming-Feng Yeh7 Grossberg Network The L1-L2 connections are instars, which performs a clustering (or categorization) operation. When an input pattern is presented, it is multiplied (after normalization) by the L1-L2 weight matrix. A competition is performed at Layer 2 to determine which row of the weight matrix is closest to the input vector. That row is then moved toward the input vector. After learning is complete, each row of the L1-L2 weight matrix is a prototype pattern, which represents a cluster (or a category) of input vectors.

Ming-Feng Yeh8 ART Networks -- 1 Learning of ART networks also occurs in a set of feedback connections from Layer 2 to Layer 1. These connections are outstars which perform pattern recall. When a node in Layer 2 is activated, this reproduces a prototype pattern (the expectation) at layer 1. Layer 1 then performs a comparison between the expectation and the input pattern. When the expectation and the input pattern are NOT closely matched, the orienting subsystem causes a reset in Layer 2.

Ming-Feng Yeh9 ART Networks -- 2 The reset disables the current winning neuron, and the current expectation is removed. A new competition is then performed in Layer 2, while the previous winning neuron is disable. The new winning neuron in Layer 2 projects a new expectation to Layer 1, through the L2-L1 connections. This process continues until the L2-L1 expectation provides a close enough match to the input pattern.

Ming-Feng Yeh10 ART Subsystems Layer 1 Comparison of input pattern and expectation. L1-L2 Connections (Instars) Perform clustering operation. Each row of W 1:2 is a prototype pattern. Layer 2 Competition (Contrast enhancement) L2-L1 Connections (Outstars) Perform pattern recall (Expectation). Each column of W 2:1 is a prototype pattern Orienting Subsystem Causes a reset when expectation does not match input pattern Disables current winning neuron

Ming-Feng Yeh11 Layer 1

Ming-Feng Yeh12 Layer 1 Operation Equation of operation of Layer 1: Output of Layer 1: Excitatory input: Input pattern + L1-L2 expectation Inhibitory input: Gain control from L2

Ming-Feng Yeh13 Excitatory Input to L1 The excitatory input: Assume that the jth neuron in Layer 2 has won the competition, i.e., The excitatory input to Layer 1 is the sum of the input pattern and the L2-L1 expectation.

Ming-Feng Yeh14 Inhibitory Input to L1 The inhibitory input – the gain control The inhibitory input to each neuron in Layer 1 is the sum of all of the outputs of Layer 2. The gain control to Layer 1 will be one when Layer 2 is active (one neuron has won the competition), and zero when Layer 2 is inactive (all neurons having zero output).

Ming-Feng Yeh15 Steady State Analysis -- 1 The response of neuron i in Layer 1: Case 1: Layer 2 is inactive – each In steady state: If then If then The output of Layer 1 is the same as the input pattern

Ming-Feng Yeh16 Steady State Analysis -- 2 Case 2: Layer 2 is active – and In steady state: Layer 1 is to combine the input vector with the expectation from Layer 2. Since both the input and the expectation are binary pattern, we will use a logic AND operation to combine the two vectors. if either or is equal to 0  if both and are equal to 1 

Ming-Feng Yeh17 Layer 1 Example Let Assume that Layer 2 is active and neuron 2 of Layer 2 wins the competition.

Ming-Feng Yeh18 Response of Layer 1

Ming-Feng Yeh19 Layer 2 From the orienting subsystem

Ming-Feng Yeh20 Layer 2 Operation Equation of operation of Layer 2: The rows of adaptive weights, after training, will represent the prototype patterns. on-center feedback adaptive instar excitatory input off-surround feedback inhibitory input

Ming-Feng Yeh21 Layer 2 Example Let

Ming-Feng Yeh22 Response of Layer 2 t

Ming-Feng Yeh23 Orienting Subsystem Determine if there is a sufficient match between the L2-L1 expectation ( a 1 ) and the input pattern ( p )

Ming-Feng Yeh24 Orienting Subsyst. Operat. Equation of operation of the Orienting Subsystem: excitatory input: inhibitory input: Whenever the excitatory input is larger than the inhibitory input, the Orienting Subsystem will be driven on. excitatory input inhibitory input

Ming-Feng Yeh25 Steady State Operation Steady state: Let, then if, or if (vigilance) The condition that will cause a reset of Layer 2.

Ming-Feng Yeh26 Vigilance Parameter. The term  is called the vigilance parameter and must fall in the range If  is close to 1, a reset will occur unless is close to If  is close to 0, need not be close to to present a reset., whenever Layer 2 is active. The orienting subsystem will cause a reset when there is enough of a mismatch between and

Ming-Feng Yeh27 Orienting Subsystem Ex. Suppose that In this case a reset signal will be sent to Layer 2, since is positive. t

Ming-Feng Yeh28 Learning Law Two separate learning laws: one for the L1-L2 connections, (instar) and another for L2-L1 connections (outstar). Both L1-L2 connections and L2-L1 connections are updated at the same time. Whenever the input and the expectation have an adequate match. The process of matching, and subsequent adaptation, is referred to as resonance.

Ming-Feng Yeh29 Subset / Superset Dilemma Suppose that, so that the prototype patterns are If the output of Layer 1 is then the input to Layer 2 will be Both prototype vectors have the same inner product with a 1, even though the 1 st prototype is identical to a 1 and the 2 nd prototype is not. This is called subset/superset dilemma.

Ming-Feng Yeh30 Subset / Superset Solution One solution to the subset/superset dilemma is to normalize the prototype patterns. The input to Layer 2 will then be The first prototype has the largest inner product with a 1. The first neuron in Layer 2 will be active.

Ming-Feng Yeh31 Learning Law: L1-L2 Instar learning with competition: When neuron i of Layer 2 is active, the i th row of,, is moved in the direction of a 1. The learning law is that the elements of compete, and therefore is normalized.

Ming-Feng Yeh32 Fast Learning For fast learning, we assume that the outputs of Layer 1 and Layer 2 remain constant until the weights reach steady state. assume that and set Case 1: Case 2: Summary:

Ming-Feng Yeh33 Learning Law: L2-L1 Typical outstar learning: If neuron j in Layer 2 is active (has won the competition), then column j of is moved toward a 1. Fast learning: assume that and  Column j of converges to the output of Layer 1, a 1, which is a combination of the input pattern and the appropriate prototype pattern. The prototype pattern is modified to incorporate the current input pattern.

Ming-Feng Yeh34 ART1 Algorithm Summary 0. Initialization: The initial is set to all 1’s. Every elements of the initial is set to. 1. Present an input pattern to the network. Since Layer 2 is NOT active on initialization, the output of Layer 1 is. 2. Compute the input to Layer 2,, and activate the neuron in Layer 2 with the largest input In case of tie, the neuron with the smallest index is declared the winner.

Ming-Feng Yeh35 Algorithm Summary Cont. 3. Compute the L2-L1 expectation (assume that neuron j of Layer 2 is activated): 4. Layer 2 is active. Adjust the Layer 1 output to include the L2-L1 expectation: 5. Determine the degree of match between the input pattern and the expectation (Orienting Subsystem): 6. If, then set, inhibit it until an adequate match occurs (resonance), and return to step 1. If, then continue with step 7.

Ming-Feng Yeh36 Algorithm Summary Cont. 7. Update row j of when resonance has occurred: 8. Update column j of : 9. Remove the input pattern, restore all inhibited neurons in Layer 2, and return to step 1. The input patterns continue to be applied to the network until the weights stabilize (do not change). ART1 network can only be used for binary input patterns.

Ming-Feng Yeh37 Solved Problem: P16.5 Train an ART1 network using the parameters and, and choosing (3 categories), and using the following three input vectors: Initial weights: 1-1: Compute the Layer 1 response:

Ming-Feng Yeh38 P16.5 Continued 1-2: Compute the input to Layer 2 Since all neurons have the same input, pick the first neuron as winner. 1-3: Compute the L2-L1 expectation

Ming-Feng Yeh39 P16.5 Continued 1-4: Adjust the Layer 1 output to include the expectation 1-5: Determine the match degree: Therefore (no reset) 1-6: Since, continued with step : Resonance has occurred, update row 1 of

Ming-Feng Yeh40 P16.5 Continued 1-8: Update column 1 of : 2-1: Compute the new Layer 1 response (Layer 2 inactive): 2-2: Compute the input to Layer 2: Since neurons 2 and 3 have the same input, pick the second neuron as winner:

Ming-Feng Yeh41 P16.5 Continued 2-3: Compute the L2-L1 expectation: 2-4: Adjust the Layer 1 output to include the expectation 2-5: Determine the match degree: Therefore (no reset) 2-6: Since, continued with step 7.

Ming-Feng Yeh42 P16.5 Continued 2-7: Resonance has occurred, update row 2 of 2-8: Update column 2 of : 3-1: Compute the new Layer 1 response: 3-2: Compute the input to Layer 2:

Ming-Feng Yeh43 P16.5 Continued 3-3: Compute the L2-L1 expectation: 3-4: Adjust the Layer 1 output to include the expectation 3-5: Determine the match degree: Therefore (no reset) 3-6: Since, continued with step 7.

Ming-Feng Yeh44 P16.5 Continued 3-7: Resonance has occurred, update row 1 of 3-8: Update column 2 of : This completes the training, since if you apply any of the three patterns again they will not change the weights. These patterns have been successfully clustered.

Ming-Feng Yeh45 Solved Problem: P16.6 Repeat Problem P16.5, but change the vigilance parameter to. The training will proceed exactly as in Problem P16.5, until pattern p 3 is presented. 3-1: Compute the Layer 1 response: 3-2: Compute the input to Layer 2:

Ming-Feng Yeh46 P16.6 Continued 3-3: Compute the L2-L1 expectation: 3-4: Adjust the Layer 1 output to include the expectation 3-5: Determine the match degree: Therefore (reset) 3-6: Since, set, inhibit it until an adequate match occurs (resonance), and return to step 1.

Ming-Feng Yeh47 P16.6 Continued 4-1: Recompute the Layer 1 response: (Layer 2 inactive) 4-2: Compute the input to Layer 2: Since neuron 1 is inhibited, neuron 2 is the winner: 4-3: Compute the L2-L1 expectation: 4-4: Adjust the Layer 1 output to include the expectation

Ming-Feng Yeh48 P16.6 Continued 4-5: Determine the match degree: Therefore (reset) 4-6: Since, set, inhibit it until an adequate match occurs (resonance), and return to step : Recompute the Layer 1 response: 5-2: Compute the input to Layer 2: Since neurons 1 & 2 are inhibited, neuron 3 is the winner:

Ming-Feng Yeh49 P16.6 Continued 5-3: Compute the L2-L1 expectation: 5-4: Adjust the Layer 1 output to include the expectation 5-5: Determine the match degree: Therefore (no reset) 5-6: Since, continued with step 7.

Ming-Feng Yeh50 P16.6 Continued 5-7: Resonance has occurred, update row 3 of 5-8: Update column 2 of : This completes the training, since if you apply any of the three patterns again they will not change the weights. These patterns have been successfully clustered.

Ming-Feng Yeh51 Solved Problem: P16.7 Train an ART1 network using the following input vectors. Present the vectors in the order p 1 - p 2 - p 3 - p 1 - p 4. Use the parameters and, and choose (three categories). Train the network until the weights have converged. The initial matrix is an matrix of 1’s. The initial matrix is an matrix, with equal to blue square  1; white square  0 5  5 grids  25-dimensional vectors

Ming-Feng Yeh52 P16.7 Continued Training sequence: p 1 - p 2 - p 3 - p 1 - p 4  : resonance √: reset