Recurrent Networks A recurrent network is characterized by

Slides:



Advertisements
Similar presentations
Bioinspired Computing Lecture 16
Advertisements

Chapter3 Pattern Association & Associative Memory
Memristor in Learning Neural Networks
Computational Intelligence
Lecture 19: Parallel Algorithms
Introduction to Neural Networks Computing
1 Neural networks 3. 2 Hopfield network (HN) model A Hopfield network is a form of recurrent artificial neural network invented by John Hopfield in 1982.
Correlation Matrix Memory CS/CMPE 333 – Neural Networks.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Un Supervised Learning & Self Organizing Maps Learning From Examples
November 30, 2010Neural Networks Lecture 20: Interpolative Associative Memory 1 Associative Networks Associative networks are able to store a set of patterns.
December 7, 2010Neural Networks Lecture 21: Hopfield Network Convergence 1 The Hopfield Network The nodes of a Hopfield network can be updated synchronously.
CS623: Introduction to Computing with Neural Nets (lecture-10) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Chapter 6 Associative Models. Introduction Associating patterns which are –similar, –contrary, –in close proximity (spatial), –in close succession (temporal)
Jochen Triesch, UC San Diego, 1 Short-term and Long-term Memory Motivation: very simple circuits can store patterns of.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
10/6/20151 III. Recurrent Neural Networks. 10/6/20152 A. The Hopfield Network.
Neural Networks Architecture Baktash Babadi IPM, SCS Fall 2004.
1 Chapter 6: Artificial Neural Networks Part 2 of 3 (Sections 6.4 – 6.6) Asst. Prof. Dr. Sukanya Pongsuparb Dr. Srisupa Palakvangsa Na Ayudhya Dr. Benjarath.
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
Hebbian Coincidence Learning
Recurrent Network InputsOutputs. Motivation Associative Memory Concept Time Series Processing – Forecasting of Time series – Classification Time series.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
Neural Networks and Fuzzy Systems Hopfield Network A feedback neural network has feedback loops from its outputs to its inputs. The presence of such loops.
IE 585 Associative Network. 2 Associative Memory NN Single-layer net in which the weights are determined in such a way that the net can store a set of.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
So Far……  Clustering basics, necessity for clustering, Usage in various fields : engineering and industrial fields  Properties : hierarchical, flat,
Ming-Feng Yeh1 CHAPTER 16 AdaptiveResonanceTheory.
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam
Optimization with Neural Networks Presented by: Mahmood Khademi Babak Bashiri Instructor: Dr. Bagheri Sharif University of Technology April 2007.
Markov Chains and Random Walks. Def: A stochastic process X={X(t),t ∈ T} is a collection of random variables. If T is a countable set, say T={0,1,2, …
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam en Universiteit Utrecht
Lecture 9 Model of Hopfield
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Computational Intelligence Winter Term 2015/16 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Lecture 39 Hopfield Network
J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control Artificial Neural Networks II - Outline Cascade Nets and Cascade-Correlation.
The Acceptance Problem for TMs
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
Hopfield Networks MacKay - Chapter 42.
CSC321 Lecture 18: Hopfield nets and simulated annealing
Visual Recognition Tutorial
New Characterizations in Turnstile Streams with Applications
Chapter 6 Associative Models
Ch7: Hopfield Neural Model
Real Neurons Cell structures Cell body Dendrites Axon
Neural Networks A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Lecture 22: Parallel Algorithms
Degree and Eigenvector Centrality
Unsupervised learning
Disadvantages of Discrete Neurons
Hidden Markov Models Part 2: Algorithms
Modern Control Systems (MCS)
O I Sinit Mealy machine LOGIC STATE 1-Dec-18
Perceptron as one Type of Linear Discriminants
Perceptrons Introduced in1957 by Rosenblatt
Guest Lecture by David Johnston
Boltzmann Machine (BM) (§6.4)
Lecture 39 Hopfield Network
The Selection Problem.
Computational Intelligence
16. Mean Square Estimation
Sanguthevar Rajasekaran University of Connecticut
Computational Intelligence
CS623: Introduction to Computing with Neural Nets (lecture-11)
Chapter 2. Simplex method
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Recurrent Networks A recurrent network is characterized by The connection graph of the network has cycles, i.e. the output of a neuron can influence its input There are no natural input and output nodes Initially each neuron has a given input state Neurons change state using some update rule The network evolves until some stable situation is reached The resulting state is the output of the network 22-Feb-19 Rudolf Mak TU/e Computer Science

Pattern Recognition Recurrent networks can be used for pattern recognition in the following way: The stable states represent the patterns to be recognized The initial state is a noisy or otherwise mutilated version of one of the patterns The recognition process consists of the network evolving from its initial state to a stable state 22-Feb-19 Rudolf Mak TU/e Computer Science

Pattern Recognition Example 22-Feb-19 Rudolf Mak TU/e Computer Science

Pattern Recognition Example (cntd) Noisy image Recognized pattern 22-Feb-19 Rudolf Mak TU/e Computer Science

Bipolar Data Encoding sgn(x ) = x xTx = n In bipolar encoding firing of a neuron is repre-sented by the value 1, and non-firing by the value –1 In bipolar encoding the transfer function of the neurons is the sign function sgn A bipolar vector x of dimension n satisfies the equations sgn(x ) = x xTx = n 22-Feb-19 Rudolf Mak TU/e Computer Science

Binary versus Bipolar Encoding The number of orthogonal vector pairs is much larger in case of bipolar encoding. In an n- dimensional vector space: For binary encoding For bipolar encoding 22-Feb-19 Rudolf Mak TU/e Computer Science

Hopfield Networks A recurrent network is a Hopfield network when The neurons have discrete output (for convenience we use bipolar encoding) Each neuron has a threshold Each pair of neurons is connected by a weighted connection. The weight matrix is symmetric and has a zero diagonal (no connection from a neuron to itself) 22-Feb-19 Rudolf Mak TU/e Computer Science

Network states If a Hopfield network has n neurons, then the state of the network at time t is the vector x(t) 2 {-1, 1}n with components x i (t) that describe the states of the individual neurons. Time is discrete, so t 2 N The state of the network is updated using a so-called update rule. (Not) firing of a neuron at time t+1 will depend on the sign of the total input at time t 22-Feb-19 Rudolf Mak TU/e Computer Science

Update Strategies In a sequential network only one neuron at a time is allowed to change its state. In the asyn-chronous update rule this neuron is randomly selected. In a parallel network several neurons are allowed to change their state simultaneously. Limited parallelism: only neurons that are not connected can change their state simultaneously Unlimited parallelism: also connected neurons may change their state simultaneously Full parallelism: all neurons change their state simul-taneously 22-Feb-19 Rudolf Mak TU/e Computer Science

Asynchronous Update 22-Feb-19 Rudolf Mak TU/e Computer Science

Asynchronous Neighborhood The asynchronous neighborhood of a state x is defined as the set of states Because wkk = 0 , it follows that for every pair of neighboring states x* 2 Na(x) 22-Feb-19 Rudolf Mak TU/e Computer Science

Synchronous Update This update rule corresponds to full parallelism 22-Feb-19 Rudolf Mak TU/e Computer Science

Sign-assumption In order for both update rules to be applica-ble, we assume that for all neurons i Because the number of states is finite, it is always possible to adjust the thresholds such that the above assumption holds. 22-Feb-19 Rudolf Mak TU/e Computer Science

Stable States A state x is called a stable state, when For both the synchronous and the asyn-chronous update rule we have: a state is a stable state if and only if the update rule does not lead to a different state. 22-Feb-19 Rudolf Mak TU/e Computer Science

Cyclic behavior in asymmetric RNN -1 1 -1 1 1 -1 1 -1 1 22-Feb-19 Rudolf Mak TU/e Computer Science

Basins of Attraction stable state initial state state space 22-Feb-19 Rudolf Mak TU/e Computer Science

Consensus and Energy The consensus C(x) of a state x of a Hopfield network with weight matrix W and bias vector b is defined as The energy E(x) of a Hopfield network in state x is defined as 22-Feb-19 Rudolf Mak TU/e Computer Science

For any pair of vectors x and x* we have Consensus difference For any pair of vectors x and x* we have 22-Feb-19 Rudolf Mak TU/e Computer Science

Asynchronous Convergence If in an asynchronous step the state of the network changes from x to x-2xkek, then the consensus increases. Since there are only a finite number of states, the consensus serves as a variant function that shows that a Hopfield network evolves to a stable state, when the asynchronous update rule is used. 22-Feb-19 Rudolf Mak TU/e Computer Science

Stable States and Local maxima A state x is a local maximum of the consensus function when Theorem: A state x is a local maximum of the consensus function if and only if it is a stable state. 22-Feb-19 Rudolf Mak TU/e Computer Science

Stable equals local maximum 22-Feb-19 Rudolf Mak TU/e Computer Science

Modified Consensus The modified consensus of a state x of a Hopfield network with weight matrix W and bias vector b is defined as Let x , x*, and x** be successive states obtained with the synchronous update rule. Then 22-Feb-19 Rudolf Mak TU/e Computer Science

Synchronous Convergence Suppose that x, x*, and x** are successive states obtained with the synchronous update rule. Then Hence a Hopfield network that evolves using the synchronous update rule will arrive either in a stable state or in a cycle of length 2. 22-Feb-19 Rudolf Mak TU/e Computer Science

Storage of a Single Pattern How does one determine the weights of a Hopfield network given a set of desired sta- ble states? First we consider the case of a single stable state. Let x be an arbitrary vector. Choos-ing weight matrix W and bias vector b as makes x a stable state. 22-Feb-19 Rudolf Mak TU/e Computer Science

Proof of Stability 22-Feb-19 Rudolf Mak TU/e Computer Science

Example 22-Feb-19 Rudolf Mak TU/e Computer Science

State encoding 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 x1 -1 x2 x3 x4 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 4z1 -1 -3 4z2 4z3 4z4 22-Feb-19 Rudolf Mak TU/e Computer Science

Finite state machine for async update 22-Feb-19 Rudolf Mak TU/e Computer Science

Weights for Multiple Patterns Let {x(p) j 1 · p · P } be a set of patterns, and let W(p) be the weight matrix corresponding to pattern number p. Choose the weight matrix W and the bias vector b for a Hopfield network that must recognize all P patterns as Question: Is x(p) indeed a stable state? 22-Feb-19 Rudolf Mak TU/e Computer Science

Remarks It is not guaranteed that a Hopfield network with weight matrix as defined on the previous slide indeed has the patterns as it stable states The disturbance caused by other patterns is called crosstalk. The closer the patterns are, the larger the crosstalk is This raises the question how many patterns there can be stored in a network before crosstalk gets the overhand 22-Feb-19 Rudolf Mak TU/e Computer Science

Input of neuron i in state x(p) 22-Feb-19 Rudolf Mak TU/e Computer Science

Crosstalk The crosstalk term is defined by Neuron i is stable when , because 22-Feb-19 Rudolf Mak TU/e Computer Science

Spurious States Besides the desired stable states the network can have additional undesired (spurious) stable states If x is stable and b = 0, then –x is also stable. Some combinations of an odd number of stable states can be stable. Moreover there can be more complicated additional stable states (spin glass states) that bare no relation to the desired states. 22-Feb-19 Rudolf Mak TU/e Computer Science

Storage Capacity Question: How many stable states P can be stored in a network of size n ? Answer: That depends on the probability of instability one is willing to accept. Experi- mentally P ¼ 0.15n has been found (by Hopfield) to be a reasonable value. 22-Feb-19 Rudolf Mak TU/e Computer Science

Probabilistic analysis 1 Assume that all components of the patterns are random variables with equal probability of being 1 and -1 Then it can be shown that has ap- proximately the standard normal distribu- tion N(0, 1). 22-Feb-19 Rudolf Mak TU/e Computer Science

Probabilistic Analysis 2 From these assumptions it follows that Application of the central limit theorem yields 22-Feb-19 Rudolf Mak TU/e Computer Science

Standard Normal Distribution The shaded area under the bell-shaped curve gives the probability Pr[y ¸ 1.5] 22-Feb-19 Rudolf Mak TU/e Computer Science

Probability of Instability 0.05 1.645 0.370 0.01 2.326 0.185 0.005 2.576 0.151 0.001 3.090 0.105 22-Feb-19 Rudolf Mak TU/e Computer Science

Topics Not Treated Reduction of crosstalk for correlated patterns Stability analysis for correlated patterns Methods to eliminate spurious states Continuous Hopfield models Different associative memories Binary Associative Memory (Kosko) Brain State in a Box (Kawamoto, Anderson) 22-Feb-19 Rudolf Mak TU/e Computer Science