It’s raining outside; want to go to the pub? It’s dry outside; want to go to the pub? Sure; I’ll grab the umbrella. What, are you insane? I’ll grab the.

Slides:



Advertisements
Similar presentations
Introduction to Neural Networks
Advertisements

Chrisantha Fernando & Sampsa Sojakka
Introduction to Neural Networks Computing
Modelling and Identification of dynamical gene interactions Ronald Westra, Ralf Peeters Systems Theory Group Department of Mathematics Maastricht University.
Omri Barak Collaborators: Larry Abbott David Sussillo Misha Tsodyks Sloan-Swartz July 12, 2011 Messy nice stuff & Nice messy stuff.
MATHEMATICAL MODELS OF SPIKE CODES AS DETERMINISTIC DYNAMICAL SYSTEMS Eleonora Catsigeras Universidad de la República Uruguay Neural Coding Montevideo.
Modeling Uncertainty over time Time series of snapshot of the world “state” we are interested represented as a set of random variables (RVs) – Observable.
III-28 [122] Spike Pattern Distributions in Model Cortical Networks Joanna Tyrcha, Stockholm University, Stockholm; John Hertz, Nordita, Stockholm/Copenhagen.
Spike timing-dependent plasticity: Rules and use of synaptic adaptation Rudy Guyonneau Rufin van Rullen and Simon J. Thorpe Rétroaction lors de l‘ Intégration.
Marseille, Jan 2010 Alfonso Renart (Rutgers) Jaime de la Rocha (NYU, Rutgers) Peter Bartho (Rutgers) Liad Hollender (Rutgers) Néstor Parga (UA Madrid)
BIOLOGICAL NEURONAL NETWORKS AS DETERMINISTIC DYNAMICAL SYSTEMS Eleonora Catsigeras Universidad de la República Uruguay
Introduction to Mathematical Methods in Neurobiology: Dynamical Systems Oren Shriki 2009 Modeling Conductance-Based Networks by Rate Models 1.
Stable Propagation of Synchronous Spiking in Cortical Neural Networks Markus Diesmann, Marc-Oliver Gewaltig, Ad Aertsen Nature 402: Flavio Frohlich.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Pattern Recognition using Hebbian Learning and Floating-Gates Certain pattern recognition problems have been shown to be easily solved by Artificial neural.
Levels in Computational Neuroscience Reasonably good understanding (for our purposes!) Poor understanding Poorer understanding Very poorer understanding.
September 14, 2010Neural Networks Lecture 3: Models of Neurons and Neural Networks 1 Visual Illusions demonstrate how we perceive an “interpreted version”
The Decisive Commanding Neural Network In the Parietal Cortex By Hsiu-Ming Chang ( 張修明 )
How facilitation influences an attractor model of decision making Larissa Albantakis.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Connected Populations: oscillations, competition and spatial continuum (field equations) Lecture 12 Course: Neural Networks and Biological Modeling Wulfram.
Artificial Neural Network
Neural basis of Perceptual Learning Vikranth B. Rao University of Rochester Rochester, NY.
Unsupervised learning
Ising Models for Neural Data John Hertz, Niels Bohr Institute and Nordita work done with Yasser Roudi (Nordita) and Joanna Tyrcha (SU) Math Bio Seminar,
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
10/6/20151 III. Recurrent Neural Networks. 10/6/20152 A. The Hopfield Network.
Lecture 10: Mean Field theory with fluctuations and correlations Reference: A Lerchner et al, Response Variability in Balanced Cortical Networks, q-bio.NC/ ,
Biological Modeling of Neural Networks Week 8 – Noisy input models: Barrage of spike arrivals Wulfram Gerstner EPFL, Lausanne, Switzerland 8.1 Variation.
Dynamical network motifs: building blocks of complex dynamics in biological networks Valentin Zhigulin Department of Physics, Caltech, and Institute for.
Deriving connectivity patterns in the primary visual cortex from spontaneous neuronal activity and feature maps Barak Blumenfeld, Dmitri Bibitchkov, Shmuel.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Geographic Information Science
Lecture 9: Introduction to Neural Networks Refs: Dayan & Abbott, Ch 7 (Gerstner and Kistler, Chs 6, 7) D Amit & N Brunel, Cerebral Cortex 7, (1997)
Modeling Neural Networks Christopher Krycho Advisor: Dr. Eric Abraham May 14, 2009.
Synchronization in complex network topologies
Image Stabilization by Bayesian Dynamics Yoram Burak Sloan-Swartz annual meeting, July 2009.
Neural Networks with Short-Term Synaptic Dynamics (Leiden, May ) Misha Tsodyks, Weizmann Institute Mathematical Models of Short-Term Synaptic plasticity.
פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב
CAP6938 Neuroevolution and Developmental Encoding Evolving Adaptive Neural Networks Dr. Kenneth Stanley October 23, 2006.
Chapter 3. Stochastic Dynamics in the Brain and Probabilistic Decision-Making in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning.
Neural Synchronization via Potassium Signaling. Neurons Interactions Neurons can communicate with each other via chemical or electrical synapses. Chemical.
Mean Field Theories in Neuroscience B. Cessac, Neuromathcomp, INRIA.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 15: Mixtures of Experts Geoffrey Hinton.
Network Models (2) LECTURE 7. I.Introduction − Basic concepts of neural networks II.Realistic neural networks − Homogeneous excitatory and inhibitory.
Biological Modeling of Neural Networks: Week 10 – Neuronal Populations Wulfram Gerstner EPFL, Lausanne, Switzerland 10.1 Cortical Populations - columns.
CSC321: Neural Networks Lecture 1: What are neural networks? Geoffrey Hinton
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
Ghent University Compact hardware for real-time speech recognition using a Liquid State Machine Benjamin Schrauwen – Michiel D’Haene David Verstraeten.
Ghent University An overview of Reservoir Computing: theory, applications and implementations Benjamin Schrauwen David Verstraeten and Jan Van Campenhout.
1 Computational Vision CSCI 363, Fall 2012 Lecture 32 Biological Heading, Color.
Chaos Analysis.
CSC2535: Computation in Neural Networks Lecture 11 Extracting coherent properties by maximizing mutual information across space or time Geoffrey Hinton.
Deep Feedforward Networks
Real Neurons Cell structures Cell body Dendrites Axon
How Neurons Do Integrals
Computer Science Department Brigham Young University
Neuro-Computing Lecture 4 Radial Basis Function Network
OCNC Statistical Approach to Neural Learning and Population Coding ---- Introduction to Mathematical.
Classification functions of chaotic pulse-coupled networks
Brendan K. Murphy, Kenneth D. Miller  Neuron 
Volume 40, Issue 6, Pages (December 2003)
Volume 36, Issue 5, Pages (December 2002)
A Dynamic System Analysis of Simultaneous Recurrent Neural Network
H.Sebastian Seung, Daniel D. Lee, Ben Y. Reis, David W. Tank  Neuron 
Yann Zerlaut, Alain Destexhe  Neuron 
Active, dynamic, interactive, system
CSC 578 Neural Networks and Deep Learning
Continuous attractor neural networks (CANNs)
Presentation transcript:

It’s raining outside; want to go to the pub? It’s dry outside; want to go to the pub? Sure; I’ll grab the umbrella. What, are you insane? I’ll grab the umbrella. Typical exchanges in London

The present state of networks depend on past input. For many task, “past” means 10s of seconds. Goal: understand how a single network can do this. Use an idea suggested by Jaeger (2001) and Maass et al. (2002).

A particular input – and only that input – strongly activates an output unit. Time varying input drives a randomly connected recurrent network. Output is a linear combination of activity; many linear combinations are possible in the same network. time output The idea: Can randomly connected networks like this one do a good job classifying input? In other words: can randomly connected networks tell that two different inputs really are different?

Answer can be visualized by looking at trajectories in activity space: r1r1 r3r3 r2r2 Activity space (N-dimensional) input 3 input 1 input 2 time There is a subtlety involving time:

r1r1 r3r3 r2r2 time r1r1 r3r3 r2r2 distinguishable inputs indistinguishable inputs τ How big can we make τ before the inputs are indistinguishable? inputs are the same starting here

τ -T-T0 Three main regimes: r1r1 r3r3 r2r2 t = -T t = 0 t = τ converging input: r1r1 r3r3 r2r2 t = -T t = 0 t = τ neutral r1r1 r3r3 r2r2 t = -T t = 0 t = τ diverging Can we build a network that operates here?

Reduced model*: x i (t+1) = sign [ ∑ j w ij x j (t) + u i (t) ] Question: what happens to nearby trajectories? *Bertschinger and Natschläger (2004): low connectivity. Our network: high connectivity. Analysis is virtually identical. random matrix mean = 0 variance = σ 2 /N number of neurons = N temporally uncorrelated input

Two trajectories: x 1,i (t) and x 2,i (t) (different initial conditions) Normalized Hamming distance: d(t) = (1/N) ∑ i |x 1,i (t)-x 2,i (t)|/2 How does d(t) evolve in time? For small d, d(t+1) ~ d(t) 1/2 This leads to very rapid growth of small separations: d(t) ~ d(0) 1/2 => d(t) ~ 1 when t ~ log log [1/d(0)].. d(t)d(t).. d(t+1).. d(t+2) t d(t)d(t) d(t+1) simulations Analysis

h P(h)P(h) 0 σ x i (t+1) = sign [ h i (t) + u ] ∑ j w ij x j (t) -u What happens if one neuron (neuron k) is different between the two trajectories? x 1k = -x 2,k h 1,i = h 2,i ± 2w ik = h 2,i + Order(σ/N 1/2 ) => N O(σ/N 1/2 )/σ = O(N 1/2 ) neurons are different on the next time step. In other words, d(0) = 1/N d(1) ~ N 1/2 /N = N -1/2 = d(0) 1/2 ~σ/N 1/2 threshold “Derivation”

V w m Real neurons: spike generation surface: small differences in initial conditions are strongly amplified (=> chaos). van Vreeswijk and Sompolinsky (1996) Banerjee (2001) Operation in the neutral regime (on the edge of chaos) is not an option in realistic networks.

Implications r1r1 r3r3 r2r2 t = -1 t = τ t = 0 Trajectories evolve onto chaotic attractors (blobs). Different initial conditions will lead to different points on the attractor. What is the typical distance between points on an attractor? How does that compare the typical distance between attractors? τ -T-T0 input:

r1r1 r3r3 r2r2 t = -1 t = τ t = d(t+1) d(t)d(t) f(d)f(d) stable equilibrium, d* near attractor, d(t+1)-d* = f'(d*) (d(t)-d*) => d(t)-d* ~ exp[t log(f'(d*))] Distance between attractors is d 0 > d* After a long time, the distance between attractors decays to d*. At that point, inputs are no longer distinguishable (with caveat). Typical distance between points on an attractor: d*. Typical distance between attractors: d 0 at time 0; d* at long times.

τ -T-T0 d*d* d0d0 input differentsame All points on the attractor are a distance d*+O(1/N 1/2 ) apart. Distance between attractors is d*+(d(0)-d*)exp[t log(f'(d*))]+O(1/N 1/2 ). State of the network no longer provides reliable information about the input when exp[τ log(f'(d*))] ~ 1/N 1/2, or: τ ~ log N -2 log(f'(d*)) distance between attractors distance within attractors indistinguishable when O(1/N 1/2 ) fraction correct τ simulations predictions Linear readout n=

Conclusions 1.Expanding on a very simple model proposed by Bertschinger and Natschläger (2004), we found that randomly connected networks cannot exhibit a temporal memory that extends much beyond the time constants of the individual neurons. 2.Scaling with the size of the network is not favorable: memory scales as log N. 3.Our arguments were based on the observation that high connectivity, recurrent networks are chaotic (Banerjee, 2001), and so our conclusions should be very general.

Mean field limit: d(t) = prob { sign [ ∑ j w ij x 1,j (t) + u i (t) ] ≠ sign [ ∑ j w ij x 2,j (t) + u i (t) ]} Define: h k,i = ∑ j w ij x k,j (t), k=1, 2 h k,i is a zero mean Gaussian random variable. Covariance matrix: R kl = = (1/N)∑ i ∑ jj' w ij x k,j (t) w ij ' x l,j '(t) = (σ 2 /N) ∑ j x k,j (t)x l,j (t) = σ 2 [ 1 – 2d(t) (1-δ kl ) ] Technical details σ 2 δ jj '

More succinctly: R = σ 2 Can compute d(t+1) as a function of d(t) by doing Gaussian integrals: The d 1/2 scaling is generic; it comes from the fact that the Gaussian ellipse has width d 1/2 in the narrow direction. (-u, -u) d 1/2 integral is over these regions 1 1-2d 1-2d 1 ()

This scaling also holds for more realistic reduced models with excitatory and inhibitory cells and synaptic and cellular time constants. x i (t+1) = sign [ ∑ j w xx,ij z xj (t) - ∑ j w xy,ij z yj (t) + u i (t) ] + (1-α)x i (t) y i (t+1) = sign [ ∑ j w yx,ij z xj (t) - ∑ j w yy,ij z yj (t) + u i (t) ] + (1-β)y i (t) z xi (t+1) = x i (t) + (1-κ) z xi (t) z yi (t+1) = x i (t) + (1-γ) z yi (t) leaky integrator synapses with temporal dynamics References: H. Jaeger, German National Research Center for Information Technology, GMD Report 148 (2001). W. Maass, T. Natschläger, and H. Markram, Neural Computation 14: (2004). Bertschinger and Natschläger, Neural Computation 16: (2004). C. van Vreeswijk and H. Sompolinsky, Science 274: (1996). A. Banerjee, Neural Computation 13: , (2001).