IJCNN, July 27, 2004 Extending SpikeProp Benjamin Schrauwen Jan Van Campenhout Ghent University Belgium.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Beyond Linear Separability
Backpropagation Learning Algorithm
NEURAL NETWORKS Backpropagation Algorithm
Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Introduction to Neural Networks Computing
NEURAL NETWORKS Perceptron
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Mehran University of Engineering and Technology, Jamshoro Department of Electronic Engineering Neural Networks Feedforward Networks By Dr. Mukhtiar Ali.
Introduction to Artificial Intelligence (G51IAI)
Artificial Spiking Neural Networks
Artificial Neural Networks - Introduction -
Classification Neural Networks 1
Simple Neural Nets For Pattern Classification
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
The back-propagation training algorithm
Chapter 6: Multilayer Neural Networks
Multi Layer Perceptrons (MLP) Course website: The back-propagation algorithm Following Hertz chapter 6.
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #31 4/17/02 Neural Networks.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial neural networks:
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Evolving a Sigma-Pi Network as a Network Simulator by Justin Basilico.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Networks
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
DIGITAL IMAGE PROCESSING Dr J. Shanbehzadeh M. Hosseinajad ( J.Shanbehzadeh M. Hosseinajad)
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Non-Bayes classifiers. Linear discriminants, neural networks.
Introduction to Neural Networks. Biological neural activity –Each neuron has a body, an axon, and many dendrites Can be in one of the two states: firing.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Backpropagation Training
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
July 23, BSA, a Fast and Accurate Spike Train Encoding Scheme Benjamin Schrauwen.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
Ghent University Compact hardware for real-time speech recognition using a Liquid State Machine Benjamin Schrauwen – Michiel D’Haene David Verstraeten.
Meta-controlled Boltzmann Machine toward Accelerating the Computation Tran Duc Minh (*), Junzo Watada (**) (*) Institute Of Information Technology-Viet.
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Ghent University Backpropagation for Population-Temporal Coded Spiking Neural Networks July WCCI/IJCNN 2006 Benjamin Schrauwen and Jan Van Campenhout.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Ghent University Accelerating Event Based Simulation for Multi-Synapse Spiking Neural Networks September 13, ICANN 2006 Michiel D'Haene, Benjamin.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Back Propagation and Representation in PDP Networks
Neural networks.
Back Propagation and Representation in PDP Networks
Artificial Neural Networks
27 April ESANN 2006 Benjamin Schrauwen and Jan Van Campenhout
Artificial neural networks:
CSE P573 Applications of Artificial Intelligence Neural Networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
Perceptron as one Type of Linear Discriminants
CSE 573 Introduction to Artificial Intelligence Neural Networks
Neural Network - 2 Mayank Vatsa
Neuro-RAM Unit in Spiking Neural Networks with Applications
Back Propagation and Representation in PDP Networks
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

IJCNN, July 27, 2004 Extending SpikeProp Benjamin Schrauwen Jan Van Campenhout Ghent University Belgium

IJCNN, July 27, Overview ● Introduction ● SpikeProp ● Improvements ● Results ● Conclusions

IJCNN, July 27, Introduction ● Spiking neural networks get increased attention: ● Biologically more plausible ● Computationally stronger (W. Maass) ● Compact and fast implementation in hardware possible (analogue and digital) ● Have temporal nature ● Main problem: supervised learning algorithms

IJCNN, July 27, SpikeProp ● Introduced by S. Bohte et al. in 2000 ● An error-backpropagation learning algorithm ● Only for SNN using “time-to-first-spike” coding t ~1/a

IJCNN, July 27, Architecture of SpikeProp ● Originally introduced by Natschläger and Ruf ● Every connection consists of several synaptic connections ● All 16 synaptic connections have enumerated delays (1- 16ms) and different weights, originally same filter

IJCNN, July 27, SRM neuron ● Modified Spike Response Model (Gerstner) t Neuron reset of no interest because only one spike needed !

IJCNN, July 27, Idea behind SpikeProp Minimize SSE between actual output spike time and desired output spike time Change weight along negative direction of the gradient

IJCNN, July 27, Math of SpikeProp Only output layer given Linearise around threshold crossing time

IJCNN, July 27, Problems with SpikeProp ● Overdetermined architecture ● Tendency to get stuck when a neuron stops firing ● Problems with weight initialisation

IJCNN, July 27, Solving some of the problems ● Instead of enumerating parameters: learn them ● Delays ● Synaptic time constants ● Thresholds ● We can use much more limited architecture ● Add specific mechanism to keep neurons firing: decrease threshold

IJCNN, July 27, Learn more parameters ● Quite similar to weight update rule ● Gradient of error with respect to parameter ● Parameter specific learning rate

IJCNN, July 27, Math of the improvements - delays Delta is the same as for weight rule, thus different delta formula for output as for inner layers.

IJCNN, July 27, Math of the improvements – synaptic time constants

IJCNN, July 27, Math of the improvements - thresholds

IJCNN, July 27, What if training gets stuck? ● If one of the neurons in the network stops firing: training rule stops working ● Solution: actively lower threshold of neuron whenever it stops firing (multiply by 0.9) ● Same as scaling all the weights up ● Improves convergence

IJCNN, July 27, What about weight initialisation ● Weight initialisation is a difficult problem ● Original publication has vague description of process ● S. M. Moore contacted S. Bohte personally for clarifying the subject for his masters thesis ● Weight initialisation is done by a complex procedure ● Moore concluded that: ”weights should be initialized in such a way that every neuron initially fires, and that its membrane potential doesn’t surpass the threshold too much”

IJCNN, July 27, What about weight initialisation ● In this publication we chose a very simple initialisation procedure ● Initialise all weights randomly ● Afterwards, set a weight such that the sum of all weights is equal to 1.5 ● Convergence rates could be increased by using more complex initialisation procedure

IJCNN, July 27, Problem with large delays During the testing of the algorithm a problem arose when the trained delays got very large: delay learning stopped If input is preceded by output: problem Solved by constraining delays Output of neuron Input of neuron

IJCNN, July 27, Results ● Tested for binary XOR (MSE = 1ms) Bohte: architecture 16 synaptic terminals 20*16 = 320 weights 250 training cycles Improvements: architecture 2 synaptic terminals 20*2 = 40 weights 130 training cycles 90% convergence architecture 2 synaptic terminals 12*2 = 24 weights 320 training cycles 60% convergence

IJCNN, July 27, Results ● Optimal learning rates (found by experiment): ● ● Some rates seem very high, but that is because the values we work with are times expressed in ms ● Idea that learning rate must be approx. 0.1 is only correct when input and weights are normalised !!

IJCNN, July 27, Conclusions ● Because parameters can be learned, no enumeration is necesarry, thus architectures are much smaller ● For XOR: ● 8 times less weights needed ● Learning converges faster (50% of original) ● No complex initialisation functions ● Positive and negative weights can be mixed ● But convergence deteriorate with further reduction of weights

IJCNN, July 27, Conclusions ● Technique only tested on small problem, should be tested on real world applications ● But, we are currently preparing a journal paper on a new backprop rule that: ● supports a multitude of coding hypotheses (population coding, convolution coding,...) ● better convergence ● simpler weight initialisation ●...