Synaptic Dynamics: Unsupervised Learning

Slides:



Advertisements
Similar presentations
Clustering. How are we doing on the pass sequence? Pretty good! We can now automatically learn the features needed to track both people But, it sucks.
Advertisements

Random Processes Introduction (2)
Chapter3 Pattern Association & Associative Memory
Introduction to Neural Networks Computing
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Random Variables ECE460 Spring, 2012.
K Means Clustering , Nearest Cluster and Gaussian Mixture
How to do control theory and discrete maths at the same time?
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Unsupervised Learning with Artificial Neural Networks The ANN is given a set of patterns, P, from space, S, but little/no information about their classification,
Artificial neural networks:
Observers and Kalman Filters
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Self Organization: Hebbian Learning CS/CMPE 333 – Neural Networks.
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Un Supervised Learning & Self Organizing Maps Learning From Examples
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Basics of regression analysis
Lecture II-2: Probability Review
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Principles of Pattern Recognition
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Network Unsupervised Learning
Chapter 8 Fuzzy Associative Memories Li Lin
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
The Logistic Growth SDE. Motivation  In population biology the logistic growth model is one of the simplest models of population dynamics.  To begin.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
DIGITAL COMMUNICATIONS Linear Block Codes
NEURONAL DYNAMICS 2: ACTIVATION MODELS
CHAPTER 5 SIGNAL SPACE ANALYSIS
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Unsupervised Learning Motivation: Given a set of training examples with no teacher or critic, why do we learn? Feature extraction Data compression Signal.
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
Architecture and Equilibra 结构和平衡 Chapter Chapter 6 Architecture and Equilibria Perface lyaoynov stable theorem.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Synaptic Dynamics: Unsupervised Learning
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Supervised learning network G.Anuradha. Learning objectives The basic networks in supervised learning Perceptron networks better than Hebb rule Single.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Discrete-time Random Signals
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
CY3A2 System identification Input signals Signals need to be realisable, and excite the typical modes of the system. Ideally the signal should be persistent.
Architecture and Equilibria 结构和平衡 Chapter 6 神经网络与模糊系统 学生: 李 琦 导师:高新波.
Rate Distortion Theory. Introduction The description of an arbitrary real number requires an infinite number of bits, so a finite representation of a.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
NEURAL NETWORK THEORY NEURONAL DYNAMICS Ⅰ : ACTIVATIONS AND SIGNALS
Assocative Neural Networks (Hopfield) Sule Yildirim 01/11/2004.
J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control Artificial Neural Networks II - Outline Cascade Nets and Cascade-Correlation.
Chapter 5 Unsupervised learning
The Viterbi Decoding Algorithm
Probability Theory and Parameter Estimation I
LECTURE 11: Advanced Discriminant Analysis
Ch7: Hopfield Neural Model
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
NEURONAL DYNAMICS 2: ACTIVATION MODELS
Spectral Clustering.
Neuro-Computing Lecture 4 Radial Basis Function Network
EE513 Audio Signals and Systems
Stability Analysis of Linear Systems
Generally Discriminant Analysis
数据的矩阵描述.
Architecture and Equilibria 结构和平衡 学生:郑巍 导师:刘三阳
NEURAL DYNAMIC1: ACTIVATIONHS AND SIGNALS
16. Mean Square Estimation
Presentation transcript:

Synaptic Dynamics: Unsupervised Learning Part Ⅱ Wang Xiumei 2018/9/20

1.Stochastic unsupervised learning and stochastic equilibrium; 2.Signal Hebbian Learning; 3.Competitive Learning. 2018/9/20

1.Stochastic unsupervised learning and stochastic equilibrium ⑴ The noisy random unsupervised learning law; ⑵ Stochastic equilibrium; ⑶ The random competitive learning law; ⑷ The learning vector quantization system. 2018/9/20

The noisy random unsupervised learning law The random-signal Hebbian learning law: (4-92) denotes a Browian-motion diffusion process, each term in (4-92)demotes a separate random process. 2018/9/20

The noisy random unsupervised learning law Using noise relationship: we can rewrite (4-92): (4-93) We assume the zero-mean, Gaussian white-noise process ,and use equation : 2018/9/20

The noisy random unsupervised learning law We can get a noisy random unsupervised learning law (4-94) Lemma: (4-95) is finite variance. proof: P132 2018/9/20

The noisy random unsupervised learning law The lemma implies two points: 1, stochastic synapses vibrate in equilibrium, and they vibrate at least as much as the driving noise process vibrates; 2,the synaptic vector changes or vibrate at every instant t, and equals a constant value. wanders in a brownian motion about the constant value E[ ]. 2018/9/20

Stochastic equilibrium When synaptic vector stops moving, synaptic equilibrium occurs in “steady state”, (4-101) synaptic vector reaches synaptic equilibrium when only the random noise vector change : (4-103) 2018/9/20

The random competitive learning law The random linear competitive learning law 2018/9/20

The learning vector quantization system. 2018/9/20

The self-organizing map system The self-organizing map system equations: 2018/9/20

The self-organizing map system The self-organizing map is a unsupervised clustering algorithm. Compared with traditional clustering algorithms, its centroid can be mapped a curve or plain, and it remains topological structure. 2018/9/20

2.Signal Hebbian Learning ⑴ Recency effects and forgetting; ⑵ Asymptotic correlation encoding; ⑶ Hebbian correlation decoding. 2018/9/20

Signal Hebbian Learning The deterministic first-order signal Hebbian learning law: (4-132) (4-133) 2018/9/20

Recency effects and forgetting Hebbian synapses learn an exponentially weighted average of sampled patterns. the forgetting term is . The simplest local unsupervised learning law: 2018/9/20

Asymptotic correlation encoding The synaptic matrix of long-term memory traces asymptotically approaches the bipolar correlation matrix : X and Y denotes the bipolar signal vectors and . 2018/9/20

Asymptotic correlation encoding In practice we use a diagonal fading-memory exponential matrix W compensates for the inherent exponential decay of learned information: (4-142) 2018/9/20

Hebbian correlation decoding First we consider the bipolar correlation encoding of the M bipolar associations ,and turn bipolar associations into binary vector associations . replace -1s with 0s 2018/9/20

Hebbian correlation decoding The Hebbian encoding of the bipolar associations corresponds to the weighted Hebbian encoding scheme if the weight matrix W equals the (4-143) 2018/9/20

Hebbian correlation decoding We use the Hebbian synaptic M for bidirectional processing of and neuronal signals, and pass neural signal through M in the forward direction, in the backward direction. 2018/9/20

Hebbian correlation decoding Signal-noise decomposition: 2018/9/20

Hebbian correlation decoding Correction coefficients : (4-149) They can make each vector resemble in sign as much as possible. The same correction property holds in the backward direction . 2018/9/20

Hebbian correlation decoding We define the Hamming distance between binary vectors and 2018/9/20

Hebbian correlation decoding [number of common bits] - [number of different bits ] 2018/9/20

Hebbian correlation decoding Suppose binary vector is close to , Then ,geometrically, the two patterns are less than half their space away from each other, So . In the extreme case ;so . The rare case that result in , and the correction coefficients should be discarded. 2018/9/20

Hebbian correlation decoding 3) Suppose is far away from , . In the extreme case: , . 2018/9/20

Hebbian encoding method binary vector bipolar vector sum contiguous correlation -encoded associations: Hebbian encoding method T 2018/9/20

Hebbian encoding method Example(P144): consider the three-step limit cycle: convert bit vectors to bipolar vectors: 2018/9/20

Hebbian encoding method Produce the asymmetric TAM matrix T: 2018/9/20

Hebbian encoding method Passing the bit vectors through T in the forward direction produces: Produce the forward limit cycle: 2018/9/20

Competitive Learning The deterministic competitive learning law: (4-165) (4-166) We see that the competitive learning law uses the nonlinear forgetting term: . 2018/9/20

Competitive Learning term . So the two laws differ in how they Heb learning law uses the linear forgetting term . So the two laws differ in how they forget, not in how they learn. In both cases when -when the jth competing neuron wins-the synaptic value encodes the forcing signal and encodes it exponentially quickly. 2018/9/20

3.Competitive Learning. ⑴ Competition as Indication; ⑵ Competition as correlation detection; ⑶ Asymptotic centroid estimation; ⑷ Competitive covariance estimation. 2018/9/20

Competition as indication Centroid estimation requires that the competitive signal approximate the indicator function of the locally sampled pattern class : (4-168) 2018/9/20

Competition as indication If sample pattern X comes from region , the jth competing neuron in should win, and all other competing neurons should Lose. In practice we usually use the random linear competitive learning law and a simple additive model. (4-169) 2018/9/20

Competition as indication the inhibitive-feedback term equals the additive sum of synapse-weighted signal: (4-170) if the jth neuron wins, and to if instead the kth neuron wins. 2018/9/20

Competition as correlation detection The metrical indicator function: (4-171) If the input vector X is closer to synaptic vector than to all other stored synaptic vectors, the jth competing neuron will win. 2018/9/20

Competition as correlation detection Using equinorm property, we can get the equivalent equalities(P147): (4-174) (4-178) (4-179) 2018/9/20

Competition as correlation detection From the above equality, we can get: The jth Competing neuron wins iff the input signal or pattern correlates maximally with . The cosine law: (4-180) 2018/9/20

Asymptotic centroid estimation The simpler competitive law: (4-181) If we use the equilibrium condition: (4-182) 2018/9/20

Asymptotic centroid estimation Solving for the equilibrium synaptic vector: (4-186) It show that equals the centroid of . 2018/9/20

Competitive covariance estimation Centroids provides a first-order Estimate of how the unknown probability Density function behaves in the regions , and local covariances provide a second-order description. 2018/9/20

Competitive covariance estimation Extend the competitive learning laws to asymptotically estimate the local conditional covariance matrices : (4-187) (4-189) denotes the centriod. 2018/9/20

Competitive covariance estimation The fundamental theorem of estimation theory [Mendel 1987]: (4-190) is Borel-measurable random vector function 2018/9/20

Competitive covariance estimation At each iteration we estimate the unknown centroid as the current synaptic vector ,In this sense becomes an error conditional covariance matrix . the stochastic difference-equation algorithm: (4-191-192) 2018/9/20

Competitive covariance estimation denotes an appropriately decreasing sequence of learning coefficients in(4-192). If the ith neuron loses the metrical competition 2018/9/20

Competitive covariance estimation The algorithm(4-192) corresponds to the stochastic differential equation: (4-195) (4-199) 2018/9/20