Computational Neuroscience 03 Lecture 8

Slides:



Advertisements
Similar presentations
© Negnevitsky, Pearson Education, Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Advertisements

Multi-Layer Perceptron (MLP)
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Spike timing dependent plasticity Homeostatic regulation of synaptic plasticity.
Biological and Artificial Neurons Michael J. Watts
Overview over different methods
Network of Neurons Computational Neuroscience 03 Lecture 6.
Simple Neural Nets For Pattern Classification
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Self Organization: Hebbian Learning CS/CMPE 333 – Neural Networks.
Contents Hebb. Learn. Patt. Assoc. Associator Correlations CS 476: Networks of Neural Computation, CSD, UOC, 2009 Examples Conclusions WK7 – Hebbian Learning.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
Pattern Recognition using Hebbian Learning and Floating-Gates Certain pattern recognition problems have been shown to be easily solved by Artificial neural.
Synapses are everywhere neurons synapses Synapse change continuously –From msec –To hours (memory) Lack HH type model for the synapse.
September 16, 2010Neural Networks Lecture 4: Models of Neurons and Neural Networks 1 Capabilities of Threshold Neurons By choosing appropriate weights.
Perceptron Learning Rule Assuming the problem is linearly separable, there is a learning rule that converges in a finite time Motivation A new (unseen)
The three main phases of neural development 1. Genesis of neurons (and migration). 2. Outgrowth of axons and dendrites, and synaptogenesis. 3. Refinement.
A globally asymptotically stable plasticity rule for firing rate homeostasis Prashant Joshi & Jochen Triesch
Artificial neural networks.
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #31 4/17/02 Neural Networks.
1 Activity-dependent Development (2) Hebb’s hypothesis Hebbian plasticity in visual system Cellular mechanism of Hebbian plasticity.
COMP305. Part I. Artificial neural networks.. Topic 3. Learning Rules of the Artificial Neural Networks.
Unsupervised learning
Supervised Hebbian Learning
Learning and Stability. Learning and Memory Ramón y Cajal, 19 th century.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
The BCM theory of synaptic plasticity.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Artificial Neural Network Unsupervised Learning
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
HEBB’S THEORY The implications of his theory, and their application to Artificial Life.
Plasticity and learning Dayan and Abbot Chapter 8.
Neural Networks and Fuzzy Systems Hopfield Network A feedback neural network has feedback loops from its outputs to its inputs. The presence of such loops.
7 1 Supervised Hebbian Learning. 7 2 Hebb’s Postulate “When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Unsupervised learning
Biological Modeling of Neural Networks Week 6 Hebbian LEARNING and ASSOCIATIVE MEMORY Wulfram Gerstner EPFL, Lausanne, Switzerland 6.1 Synaptic Plasticity.
Neural Networks Steven Le. Overview Introduction Architectures Learning Techniques Advantages Applications.
Unsupervised Learning Motivation: Given a set of training examples with no teacher or critic, why do we learn? Feature extraction Data compression Signal.
Neural Networks 2nd Edition Simon Haykin
1 4. Associators and Synaptic Plasticity Lecture Notes on Brain and Computation Byoung-Tak Zhang Biointelligence Laboratory School of Computer Science.
1 Financial Informatics –XVII: Unsupervised Learning 1 Khurshid Ahmad, Professor of Computer Science, Department of Computer Science Trinity College, Dublin-2,
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Adaptive Cooperative Systems Chapter 8 Synaptic Plasticity 8.11 ~ 8.13 Summary by Byoung-Hee Kim Biointelligence Lab School of Computer Sci. & Eng. Seoul.
EEE502 Pattern Recognition
Keeping the neurons cool Homeostatic Plasticity Processes in the Brain.
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 2 ARTIFICIAL.
Jochen Triesch, UC San Diego, 1 Part 3: Hebbian Learning and the Development of Maps Outline: kinds of plasticity Hebbian.
Introduction to Connectionism Jaap Murre Universiteit van Amsterdam en Universiteit Utrecht
Biointelligence Laboratory, Seoul National University
Real Neurons Cell structures Cell body Dendrites Axon
Simple learning in connectionist networks
Financial Informatics –XVII: Unsupervised Learning
Hebb and Perceptron.
Emre O. Neftci  iScience  Volume 5, Pages (July 2018) DOI: /j.isci
Volume 40, Issue 6, Pages (December 2003)
Backpropagation.
Feature space tansformation methods
Artificial neurons Nisheeth 10th January 2019.
Simple learning in connectionist networks
Introduction to Neural Network
Supervised Hebbian Learning
Presentation transcript:

Computational Neuroscience 03 Lecture 8 Learning Rules 1 Computational Neuroscience 03 Lecture 8

Last week showed how to model neurons and networks of neurons using firing rate models. And we then discussed how to add them together to form networks of neurons

Feedforward and Recurrent networks Where the weight vector is replaced by a matrix. Also often replace feedforward input with a vector

Recurrent networks can also do this but have much more complex dynamics than feedforward nets. Also more difficult to analyse Much analysis focuses on looking at the eigenvectors of the matrix M Can show for instance that networks can exhibit selective amplification if there is one dominant eigenvector (cf PCA)

Or if an eigenvalue is exactly equal to 1 and others < 1can get integration of inputs and therefore persistent activity as activity does not stop when input stops But how are we to generate such precise weight changes? Need some synaptic modification rules

Hebb’s postulate Hebb's postulate of learning (or simply Hebb's rule) (1949), is the following: "When an axon of cell A is near enough to excite cell B and repeatedly or persistently takes part in firing it, some growth processes or metabolic changes take place in one or both cells such that A's efficiency as one of the cells firing B, is increased". This rule forms the basis of much of the research on role of synaptic plasticity in memory and learning Has been generalised to include decreases of strength when neuron A repeatedly fails to be involved in activation of B and generally look at the correlation or covariance of activities of pre-and postsynaptic neurons

Learning in the hippocampus Here see examples of long-term potentiation and depression (LTP and LTD). High frequency stimulation induces LTP while long-lasting low frequency leads to LTD ‘Long Term’ refers to changes > 10 mins

Learning Rules We will consider 3 types of learning and will focus on Hebbian type learning though others exist (modifcations on pre/postsynaptic activity only etc) Unsupervised learning: Network responds during training solely as a result of its connections and intrinsic dynamics. Net self-organises in a manner dependent on inputs and synaptic plasticity rule Supervised learning: Here the network also has a “teacher” in the form of a set of desired ‘target’ outputs for each input. Not especially biologically plausible, but good for existence proofs Reinforcement learning: Somewhere in between the 2. The net does not know the target output but gets feedback via reward punishment

Unsupervised Learning Start with single postsynaptic neuron and a linear activation function. As synaptic changes will be much longer time scale than these dynamics firing rate equation reduces to: Start with simplest Hebbian style plasticity for a single neuron:

As u, v denote firing rates vu can be interpreted as a probability that pre- and post fire together. Each different set of u is known as an input pattern. As weight changes are slow, rather than summing all changes separately can average the input patterns and thus compute the average change. To do this use < > to denote averages over the ensemble of input patterns. Thus get: Remembering that v = w.u this gives the correlation-based rule Where Q is correlation matrix: Qbb’ = <ubub’>

eg for 2 patterns u1 and u2, Qij = ½(u1i u1j + u2i u2j) So suppose u1= (1, 0) and u2= (0, 1) then Alternatively u1= (1, 1) and u2= (1, 0) then Suppose tw=10 and w=(0.1,0.1) and we have the 2nd matrix then: As wn+1= wn + Dw, w1=(0.115, 0.11), w2=(0.13, 0.12) …. Which leads to instability and uncontrolled growth of w

W=(0.1, 0.1) W=(0.1, -0.3) Outcome dependent on eiegenvectors of Q and initial conditions, but unstable

To avoid unbounded growth can impose a saturation constraint. However, this means all weights go to max or min and thus we have no competition between different synapses This means that the neuron cannot distinguish between presynaptic inputs Also since u, v are firing rates and therefore positive, Hebb rule only describes LTP. However, earlier figure showed that synapses can depress in strength if presynaptic activity is accompanied by a low level of postsynaptic activity Can also get results where opposite is true

Analysis focuses on looking at eigenvectors of Q the correlation matrix of the input vectors For instance, can show that after training Hebbian rule leads to v a e1.u for arbitary vector u and that weight vector expressed as a sum of eigenvectors is dominated by e1 ie w a e1 That is v is projection of u onto the principal eigenvector of Q Eg for a Q with principal eigenvector (1, -1)/sqrt(2) we would expect w to end up as (wmax,0) or (0, wmax). However because of saturation constraints can get (wmax, wmax)

Covariance rule To get LTD can introduce a postsynaptic or presynaptic threshold: Below thresholds get depression, above potentiation. A convenient choice for the thresholds is the average pre/post synaptic input <u> or <v>. If we now replace v with w.u we get the covariance rule Where C is the covariance matrix of the input data ie:

Note that as <v> keeps changing we need to keep updating the postsynaptic threshold while the presynaptic one is independent of the weights Although both average to the same thing, they do have differences. The postsynaptic threshold means that only modifies weights for non-zero presynaptic activities. If v is below threshold then this results in homosynaptic depression Alternatively, presynaptic threshold reduces the strength of inactive synapses for v>0: heterosynaptic depression Although the covariance rule allows LTD it is still unstable due to positive feedback Also we do not have competition, but this can be introduced to allowing threshold to slide as follows

BCM rule As covariance rule allows LTD without postsynaptic/presynaptic activity, Bienenstock, Cooper and Munro (82) proposeed an alternative for which there is experimental evidence where the postsynaptic threshold is dynamic

This is again unstable if q is fixed. However, if we allow the threshold to grow faster than v we get stability. For instance use q as low pass filtered version of v2 Usually set tq to be less than tw so that changes in q faster than changes in v Now get competition between synapses since strengthening some synapses results in threshold increasing meaning that it is harder for others to be strengthened

Synaptic normalisation A more direct way of enforcing competition is through synaptic normalisation Idea is that postsynaptic neuron can only support a certain amount of total synaptic weight so strengthening one leads to weakening others Can either hold the sum of weights constant if all are +ve or –ve or can constrain the sum of squares of the weights (cf ANN network pruning) 2 types: subtractive normalisation and multiplicative normalisation

Subtractive normalisation Where n is a vector of ones of length Nu so n.u is the sum of all the inputs u. Thus the second term is simply a vector Nu long with the same values in ie (k, k, ….., k) whose sum over all the elements is equal to the sum over all the elements of vu. Thus the total increase in the weights is 0 This rule must be augmented by a saturation constraint to prevent the weights becoming negative, that if a weight becomes zero, it is not moved downwards Also, without upper saturation often leads to all weights bar one being zero. Note also rule involves global knowledge of weights

Multiplicative normalisation Where a is a +ve constant: known as Oja’s rule (1982) This rule is more local than previous as it only involves the weight in question and pre-and post synaptic activities. However, its form is based on theoretical arguments rather than experimental data Previous rule was rigid as it had to be satisfied at all times whereas this is more dynamic with |w|2 gradually relaxing to 1/ a This induces competition as if one weight increases, the maintenance of constant length of the weight vector forces others to decrease

Both Hebbian and Oja’s rule run for long enough generate vectors parallel to principal eigenvector of correlation matrix as in A This is basically principal component analaysis (PCA) which is theoretically the optimal in terms of retaining info way to encode high dimensional info onto lower dimensional subspace However, B shows what happens if input vectors don’t have zero mean (as in real systems), but this problem is alleviated by using covariance-based rules

Timing based rules Previous rules don’t take timing into account Can be crucial since if pre-synaptic spike occurs after postsynaptic get LTD rather than LTP

Therfore need to integrate over time as in the following: Where H is a function like the solid line in previous figure Such functions still require saturation constraints but timing can generate competition

Multiple Postsynaptic Neurons Can extend the rules defined previously to nets with multiple postsynaptic neurons. In these networks the output rates are: Thus where And the Hebbian rule becomes:

Can also use feature-based models where the net is indexed by input features rather than buy individual inputs Network models can have adaptive feedforward wieghts and fixed recurrent ones, or vice versa or both layers can be adaptive Can get competition through mainly inhibitory recurrent connections Can then get eg self-organising maps and elastic nets where K can have forms like: