Presentation is loading. Please wait.

Presentation is loading. Please wait.

STDP Finds the Start of Repeating Patterns in Continuous Spike Trains Timothee Masquelier, Rudy Guyonneau, Simon J. Thorpe.

Similar presentations


Presentation on theme: "STDP Finds the Start of Repeating Patterns in Continuous Spike Trains Timothee Masquelier, Rudy Guyonneau, Simon J. Thorpe."— Presentation transcript:

1 STDP Finds the Start of Repeating Patterns in Continuous Spike Trains Timothee Masquelier, Rudy Guyonneau, Simon J. Thorpe

2 Timothee Masquelier Computational Neuroscience Group, Universitat Pompeu Fabra, Barcelona, Spain STDP in neural networks, implications for neural coding and information processing. Applications in the visual system for object and motion recognition Rudy Guyonneau The Brain and Cognition Research Center, Toulouse, France Biologically-inspired learning in asynchronously spiking neural networks Simon J. Thorpe Research Director CNRS, Toulouse, France Understanding the phenomenal processing speed achieved by the visual system About Authors

3 Discreet spike patterns learning* A neuron is presented successively with discrete volleys of input spikes STDP concentrates synaptic weights on afferents that consistently fire early Postsynaptic spike latency decreases Postsynaptic spike latency reaches a minimal and stable value Limitations: Input spikes arrive in discrete volleys (spike waves) Explicit time reference for coding allowing specification of latency for afferents Activity between volleys is assumed to be much weaker Patterns presented in all volleys (no ‘distractor’ volleys) * Example: Masquelier T., Thorpe S.J. Unsupervised learning of visual features through STDP

4 Patterns learning with continuous spike firing Afferents (sensory simulators) fire continuously with a constant population rate Only a fraction of the recorded neurons are involved in the pattern Pattern of spike repeats at irregular intervals, but is hidden within the variable background firing of the whole population Nothing in terms of population firing rate characterizes the periods when the pattern is present Nothing is unusual about the firing rates of the neurons involved in the pattern Detecting the pattern requires taking the spike times into account… Is STDP is still able to find and learn spike patterns among the inputs? What happens if pattern appears at unpredictable times separated by ‘distractor’ periods? Is it OK to use the beginning of pattern as a time reference? Does the postsynaptic latency still decrease with respect to the beginning of the pattern as time reference?

5 Single receiving STDP neuron with a 10ms membrane constant Randomly generated ‘distractor’ spike trains (Poisson process with variable firing rates) 450s long Arbitrary pattern inserted at various times (half of afferents left random mode for 50ms and adopted a precise firing pattern) Same spike density of pattern and ‘distractor’ parts making pattern invisible in terms of firing rates Unsupervised learning Simulation Environment Afferents1 STDP neuron

6 Each afferent emits spikes independently using non-homogenous Poisson process: Firing rate r varies between 0 and 90 Hz (dr = s.dt) Max rate change s is selected to enable neuron going from 0 to 50 in 50ms Random change is applied to s Distractor Spike Trains

7 Spatio-temporal spike pattern Nothing characterizes the averages in terms of firing rates. Detecting the pattern thus requires taking the spike times into account The population-averaged firing rates over 10 ms time bins (membrane time constant of the neuron used in the simulations) A repeating 50 ms long pattern that concerns 50 afferents among 100 Individual firing rates averaged over the whole period. Neurons involved in the pattern are shown in red

8 Additional Difficulties Permanent 10 Hz Poissonian spontaneous activity Arbitrary pattern inserted at various times (half of afferents left random mode for 50ms and adopted a precise firing pattern) Same spike density of pattern and ‘distractor’ parts making pattern invisible in terms of firing rates A 1ms jitter is added to the pattern Is it still possible to develop pattern selectivity in these conditions?

9 Happy Simulation Results One single Leaky Integrate-and-Fire neuron is perfectly able to solve the problem and learns to respond selectively to the start of the repeating pattern: Implements STDP Acts as coincidence detector ‘Successful’ simulation definition: Postsynaptic latency inferior to 10 ms Hit rate superior to 98% No false alarms Observed in 96% of the cases (from 100 simulations) Limitation: Only excitation is considered, no inhibition in all these simulations

10 Leaky Integrate and Fire Neuron

11 LIF is modeled using Gerstner’s Spike Response Model (SRM) In SRM instead of solving the membrane potential diff. equation kernels are used to model the effect of presynaptic and postsynaptic spikes on the membrane potential Refractoriness can be modelled as a combination of increased threshold, hyperpolarizing afterpotential, and reduced responsiveness after a spike, as observed in real neurons (Badel at al. 2008) Spike Response Model

12 Membrane potential is given as:  t firing time of the last spike of the neuron  η response of the membrane to its own spike  k response of the neuron to the incoming short pulse  I(t) stimulating current The next spike occurs if the membrane potential u hits a threshold in which case firing time of the last spike is updated Spike Response Model

13 It is possible to show that LIF is a special case of SRM Take the differential equation of the LIF: Integration of the differential equation for arbitrary input I(t), yields: If we define k to be: it brings us to the general model equation Connection between LIF and SRM

14 In case the input is generated by synaptic current pulses caused by presynaptic spike arrival the model can be rewritten as: Spike Response Model =>

15 The kernel ε(t-t j ) can be interpreted as the time course of a postsynaptic potential evoked by the firing of a presynaptic neuron j at time t j (EPSP in our case) Spike Response Model = 10ms – membrane time constant = 2.5ms – synapse time constant Heaviside step function: K – multiplicative constant to scale the max value of the kernel to 1

16 The kernel ε(t-t j ) can be interpreted as the time course of a postsynaptic potential evoked by the firing of a presynaptic neuron j at time t j (EPSP in our case) The responsiveness ε to an input pulse depends on the time since last spike, because typically the effective membrane time constant after a spike is shorter, since many ion channels are open The time course of the response is presented as a combination of exponentials with different time constants Spike Response Model

17 The kernel η(t-t i ) describes the standard form of an action potential of neuron i including the negative overshoot which typically follows a spike (after-potential) T=500 - the threshold of the neuron K1=2, K2=4 Resting potential is zero Both ε and η are rounded to zero when respectively t-t j and t-t i are greater than 7 * w j – excitatory synaptic weights Spike Response Model Positive pulse Negative spike after-potential

18 STDP Modification Function It was found that small learning rates led to more robust learning

19 Coincidence Detection Coincidence detection relies on separate inputs converging on a common target If each input neuron's EPSP is a sub-threshold for an action potential at target neuron, then it will not fire unless the n inputs from X 1, X 2, …, X n are close together in time For n=2, if the two inputs arrive too far apart, the depolarization of the first input may have time to drop significantly, preventing the membrane potential of the target neuron from reaching the action potential threshold

20 Simulation Data At the beginning of the simulation the neuron is non- selective Synaptic weights are all equal (w=0.475) Neuron fires periodically, both inside and outside the pattern

21 After 70 Pattern Presentations and 700 Discharges No false alarms High hit rate Neuron fires even twice when pattern is present Chance determines which part(s) of the pattern the neuron becomes selective to

22 End of Simulation The system is converged (by saturation) Hit rate 99.1% with no false alarms Postsynaptic spike latency reduction to 4ms (converged to saturation) Fully specified neurons might have very low spontaneous rates Higher spontaneous rates might characterize less well specified cells

23 Convergence by Saturation All the spikes in the pattern that precede the postsynaptic spike already correspond to maximally potentiated synapses All these spikes are needed to reach the threshold Synapses corresponding to the spikes outside the pattern are fully depressed by STDP Bimodal Weight Distribution

24 Converged State Afferents out of pattern Afferents involved in pattern Pattern boundaries STDP has potentiated most of the synapses that correspond to the earliest spikes of the pattern Sudden increase in membrane potential corresponds to the white spikes cluster

25 Latency Reduction neuron is non-selective selectivity has emerged and STDP is ‘tracking back’ through the pattern the system has converged

26 Resistance to Degradation Additional batches of 100 simulations were performed to investigate the impact on successful performance of the following parameters: Relative pattern frequency Pattern duration Amount of jitter Proportion of afferents involved in the pattern Initial weights Proportion of missing spikes Membrane constant

27 Pattern Frequency Pattern frequency - sum of pattern durations over total duration ratio Pattern needs to be consistently present for the STDP to establish selectivity. Later, longer distractor periods does not perturb the learning at all Detection difficulty is related to the delay between two pattern presentations, not to the decrease in pattern duration Baseline condition – 0.25

28 For jitter larger than 3ms: Spike coincidences are lost STDP weight updates are inaccurate Learning is impaired Millisecond spiking precision has been reported for brain Amount of Jitter Baseline condition – 1ms

29 With fewer afferents involved in the pattern, it becomes harder to detect Detected more than 50% with 1/3 afferents only => ‘developmental exuberance’ Proportion of Afferents Involved in Pattern Developmental exuberance: Activity-driven mechanisms could select a small set of ‘interesting’ afferents among a much bigger set of initially connected afferents, probably specified genetically Baseline condition – 1/2

30 For too many false discharges in a row the threshold may no longer be reachable as they lead to decrease of synaptic weights Initial Weight Baseline condition – 0.475 High initial value for the weights increases the resistance to discharges outside the pattern High initial weights also cause the neuron to discharge at a high rate at the beginning of the learning process

31 As expected number of successfully learned patterns decreases with the proportion of spikes deleted However the system is robust to spike deletion (82% success with a 10% deletion) Proportion of Missing Spikes Baseline condition – 0%

32 Conclusions Main results previously obtained for STDP based learning with the highly simplified scheme of discrete spike volleys still stand in current more challenging continuous framework It is not necessary to know a time reference in order to decode a temporal code, recurrent spike patterns in the inputs are enough, even embedded in equally dense ‘distractor’ spike trains A neuron will gradually respond faster and faster when the pattern is presented, by potentiating synapses that correspond to the earliest spikes of the patterns, and depressing the others (substantial amount of information could be available very rapidly, in the very first spikes evoked by a stimuli)

33 Additional Observations LIF neuron can never be selective to the whole pattern (membrane constant – 10ms, pattern duration – 50ms) LIF is selective to ‘one coincidence’ of the pattern at a time, that is, selective to the nearly simultaneous arrival of certain earliest spikes, just as it occurs in one subdivision of the pattern (chance determines which one) LIF can serve as ‘earliest predictor’ of the subsequent spike events: Risk of triggering a false alarm Benefit of being very reactive

34 Additional Observations – Cont. If more than one repeating pattern is present in the input a postsynaptic neuron picks one (chance determines which one), and becomes selective to it and only to it To learn the other patterns other neurons are needed. A competitive mechanism could ensure they optimally cover all the different patterns and avoid learning the same ones (inhibitory horizontal connections between neurons) A long input pattern can be coded by the successive firings of several STDP neurons, each selective to a different part of the pattern Further research is needed to evaluate this form of competitive network

35 Final Conclusion One single LIF neuron equipped with STDP is consistently able to detect one arbitrary repeating spatio-temporal spike pattern embedded in equally dense ‘distractor’ spike trains, which is a remarkable demonstration of the potential for such a scheme


Download ppt "STDP Finds the Start of Repeating Patterns in Continuous Spike Trains Timothee Masquelier, Rudy Guyonneau, Simon J. Thorpe."

Similar presentations


Ads by Google