An extended Hebbian model of unsupervised learning

Slides:



Advertisements
Similar presentations
Computational Neuroscience 03 Lecture 8
Advertisements

Chapter3 Pattern Association & Associative Memory
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Ming-Feng Yeh1 CHAPTER 13 Associative Learning. Ming-Feng Yeh2 Objectives The neural networks, trained in a supervised manner, require a target signal.
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Linear Transformations
Pontifical Catholic University of the Rio Grande do Sul Brazil Applying Artificial Neural Networks to Energy Quality Measurement Fernando Soares dos Reis.
Correlation Matrix Memory CS/CMPE 333 – Neural Networks.
MACHINE LEARNING - Doctoral Class - EDIC EPFL A.. Billard MACHINE LEARNING Information Theory and The Neuron - II Aude.
Self Organization: Hebbian Learning CS/CMPE 333 – Neural Networks.
Contents Hebb. Learn. Patt. Assoc. Associator Correlations CS 476: Networks of Neural Computation, CSD, UOC, 2009 Examples Conclusions WK7 – Hebbian Learning.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Synapses are everywhere neurons synapses Synapse change continuously –From msec –To hours (memory) Lack HH type model for the synapse.
Perceptron Learning Rule Assuming the problem is linearly separable, there is a learning rule that converges in a finite time Motivation A new (unseen)
The Perceptron. 0T Afferents V thr V rest t max 0 What does a neuron do? spike no spike.
A globally asymptotically stable plasticity rule for firing rate homeostasis Prashant Joshi & Jochen Triesch
6 1 Linear Transformations. 6 2 Hopfield Network Questions.
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #31 4/17/02 Neural Networks.
PCA NETWORK Unsupervised Learning NEtWORKS. PCA is a Representation Network useful for signal, image, video processing.
1 Statistical Analysis Professor Lynne Stokes Department of Statistical Science Lecture 5QF Introduction to Vector and Matrix Operations Needed for the.
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
Principle Component Analysis (PCA) Networks (§ 5.8) PCA: a statistical procedure –Reduce dimensionality of input vectors Too many features, some of them.
Unsupervised learning
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
Artificial Neural Network Unsupervised Learning
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
Plasticity and learning Dayan and Abbot Chapter 8.
The Boltzmann Machine Psych 419/719 March 1, 2001.
7 1 Supervised Hebbian Learning. 7 2 Hebb’s Postulate “When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part.
Unsupervised learning
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
Adaptive Algorithms for PCA PART – II. Oja’s rule is the basic learning rule for PCA and extracts the first principal component Deflation procedure can.
Unsupervised Learning Motivation: Given a set of training examples with no teacher or critic, why do we learn? Feature extraction Data compression Signal.
Bayes Theorem The most likely value of x derived from this posterior pdf therefore represents our inverse solution. Our knowledge contained in is explicitly.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Adaptive Cooperative Systems Chapter 8 Synaptic Plasticity 8.11 ~ 8.13 Summary by Byoung-Hee Kim Biointelligence Lab School of Computer Sci. & Eng. Seoul.
Status of Reference Network Simulations John Dale ILC-CLIC LET Beam Dynamics Workshop 23 June 2009.
September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
Jochen Triesch, UC San Diego, 1 Part 3: Hebbian Learning and the Development of Maps Outline: kinds of plasticity Hebbian.
CSC2535: Computation in Neural Networks Lecture 7: Independent Components Analysis Geoffrey Hinton.
Pairwise comparisons: Confidence intervals Multiple comparisons Marina Bogomolov and Gili Baumer.
General Considerations
J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control Artificial Neural Networks II - Outline Cascade Nets and Cascade-Correlation.
LECTURE 04: DECISION SURFACES
Principle Component Analysis (PCA) Networks (§ 5.8)
Biointelligence Laboratory, Seoul National University
Synaptic Dynamics: Unsupervised Learning
Financial Informatics –XVII: Unsupervised Learning
Hebb and Perceptron.
Unsupervised learning
A principled way to principal components analysis
Optimal prediction of x
Bayesian belief networks 2. PCA and ICA
Techniques for studying correlation and covariance structure
Outline Associative Learning: Hebbian Learning
OCNC Statistical Approach to Neural Learning and Population Coding ---- Introduction to Mathematical.
Matrix Algebra and Random Vectors
Feature space tansformation methods
Artificial neurons Nisheeth 10th January 2019.
NONLINEAR AND ADAPTIVE SIGNAL ESTIMATION
NONLINEAR AND ADAPTIVE SIGNAL ESTIMATION
Lecture 16. Classification (II): Practical Considerations
Supervised Hebbian Learning
Presentation transcript:

An extended Hebbian model of unsupervised learning Anca Radulescu, University of Colorado at Boulder 11/11/2019 An extended Hebbian model of unsupervised learning Recent work in Long Term Potentiation in brain slices shows that Hebb’s rule is not completely synapse-specific, possibly due to intersynapse Calcium diffusion. We extend the classical Oja unsupervised learning model of a single linear neuron to include Hebbian infidelity, by introducing an error matrix which expresses the cross-talk between the Hebbian updating at different connections. The current linear model is sensitive only to pairwise statistics. We expect that a more powerful nonlinear model that takes into consideration higher correlations would show an error catastrophe under biologically plausible conditions. 11/11/2019 Anca Radulescu, CU Boulder Rocky Mountains Bioinformatics Conference, Snowmass, December 2006

Anca Radulescu, CU Boulder Linear neuron model: Input signals on a linear layer of n neurons, randomly drawn from a probability distribution P (x = (x1…xn)t) A unique output neuron: h = S xkwk Connections with plastic synaptic weights: w = (w1…wn)t which update following Hebb’s rule of learning: wk(s+1) = wk(s) + g h(s) xk(s) Normalize, expand in Taylor series w.r.t g and ignore the O(g2) terms for g small. Assume the process is stationary (slow) and that x(s) and w(s) are statistically independent: w(s+1) = w(s) + g [Cw - (wtCw)w] C = covariance matrix of the input distribution: C=<xtx> 11/11/2019 Anca Radulescu, CU Boulder

Anca Radulescu, CU Boulder Equivalent dynamical system to be studied: We study the stability of the synaptic vector w under iterations of the function: f : lRn → lRn f (w) = w + g [Cw - (wtCw)w] The result should depend on the distribution to be learnt (through the symmetric, positive definite matrix C) and on the size of g . w fixed point for f  w is a unit eigenvector of C w is a hyperbolic attractor for f  Dfw has only stable eigendirections (i.e. with eigenvalues |l|<1) 11/11/2019 Anca Radulescu, CU Boulder

Anca Radulescu, CU Boulder Dfw=I + g [C - 2w(Cw)t - (wtCw)I] Complete w to an orthonormal eigenbasis B of C. Then B is also an eigenbasis for Dfw: Dfwv = (1 – g [lw - lv])v Dfww = (1 - 2lw)w w is a hyperbolic fixed point of f if and only if: |1 – g (lw - lv) | < 1 |1 - 2lw| < 1 Conclusion: The fixed vector w is asymptotically stable if lw > lv , for all v  w and g < lw-1 11/11/2019 Anca Radulescu, CU Boulder

Anca Radulescu, CU Boulder To generalize for a system whose information transmission Is imperfect, we encode the errors in a matrix T  Mn(R), positive, symmetric, T = I for zero error: fT(w) = w + g [TCw - (wTCw)w] If the matrix TC has a positive maximal eigenvalue m with multiplicity one, then its corresponding eigenvector w normalized by wtCw =1 is the only asymptotically stable fixed vector of fT, provided that g is small. In the absence of error, the network learnt the principal eigenvector of the correlation matrix C. A reasonable measure of the output error for a given error matrix T is: cos(q) = wC,wTC||wC||-1||wTC||-1 11/11/2019 Anca Radulescu, CU Boulder

Anca Radulescu, University of Colorado at Boulder 11/11/2019 ◘ Analytic results in the particular case of uncorrelated inputs and neurons equivalently exposed to error ◘ Simulations for more general settings. 11/11/2019 Anca Radulescu, CU Boulder Rocky Mountains Bioinformatics Conference, Snowmass, December 2006