Adaptive Cooperative Systems Chapter 8 Synaptic Plasticity 8.11 ~ 8.13 Summary by Byoung-Hee Kim Biointelligence Lab School of Computer Sci. & Eng. Seoul.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Computational Neuroscience 03 Lecture 8
Introduction to Neural Networks Computing
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Dimension reduction (1)
PCA + SVD.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Un Supervised Learning & Self Organizing Maps. Un Supervised Competitive Learning In Hebbian networks, all neurons can fire at the same time Competitive.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Principal Component Analysis
Self Organization: Hebbian Learning CS/CMPE 333 – Neural Networks.
Dimensional reduction, PCA
Un Supervised Learning & Self Organizing Maps Learning From Examples
Independent Component Analysis (ICA) and Factor Analysis (FA)
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Continuous Latent Variables --Bishop
Techniques for studying correlation and covariance structure
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Understanding visual map formation through vortex dynamics of spin Hamiltonian models Myoung Won Cho BK21 Frontier Physics.
Empirical Modeling Dongsup Kim Department of Biosystems, KAIST Fall, 2004.
Summarized by Soo-Jin Kim
Principle Component Analysis (PCA) Networks (§ 5.8) PCA: a statistical procedure –Reduce dimensionality of input vectors Too many features, some of them.
Unsupervised learning
CSE554AlignmentSlide 1 CSE 554 Lecture 8: Alignment Fall 2014.
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Machine Vision for Robots
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
The BCM theory of synaptic plasticity.
CSE554AlignmentSlide 1 CSE 554 Lecture 5: Alignment Fall 2011.
Plasticity and learning Dayan and Abbot Chapter 8.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Unsupervised learning
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Yung-Kyun Noh and Joo-kyung Kim Biointelligence.
Ch 4. Linear Models for Classification (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized and revised by Hee-Woong Lim.
The BCM theory of synaptic plasticity. The BCM theory of cortical plasticity BCM stands for Bienestock Cooper and Munro, it dates back to It was.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent (If f(x) is more complex we usually cannot.
CSE554AlignmentSlide 1 CSE 554 Lecture 8: Alignment Fall 2013.
Unsupervised Learning Motivation: Given a set of training examples with no teacher or critic, why do we learn? Feature extraction Data compression Signal.
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Ch 5. The Patterning of Neural Connections 5.5 ~ 5.6 Adaptive Cooperative Systems, Martin Beckerman, Summarized by Kwonill, Kim Biointelligence Laboratory,
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Joo-kyung Kim Biointelligence Laboratory,
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Feature Extraction 主講人:虞台文.
Chapter 4. Analysis of Brain-Like Structures and Dynamics (2/2) Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans 09/25.
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
Independent Component Analysis features of Color & Stereo images Authors: Patrik O. Hoyer Aapo Hyvarinen CIS 526: Neural Computation Presented by: Ajay.
Jochen Triesch, UC San Diego, 1 Part 3: Hebbian Learning and the Development of Maps Outline: kinds of plasticity Hebbian.
1 (c) SNU CSE Biointelligence Lab, Chap 3.8 – 3.10 Joon Shik Kim BI study group.
Bayesian Brain - Chapter 11 Neural Models of Bayesian Belief Propagation Rajesh P.N. Rao Summary by B.-H. Kim Biointelligence Lab School of.
CSE 554 Lecture 8: Alignment
Biointelligence Laboratory, Seoul National University
Principle Component Analysis (PCA) Networks (§ 5.8)
Biointelligence Laboratory, Seoul National University
Biointelligence Laboratory, Seoul National University
Lecture 8:Eigenfaces and Shared Features
A principled way to principal components analysis
Presented by Rhee, Je-Keun
Introduction PCA (Principal Component Analysis) Characteristics:
Biointelligence Laboratory, Seoul National University
Adaptive Cooperative Systems Chapter 6 Markov Random Fields
Adaptive Cooperative Systems Chapter 3 Coperative Lattice Systems
Principal Component Analysis
Biointelligence Laboratory, Seoul National University
Presentation transcript:

Adaptive Cooperative Systems Chapter 8 Synaptic Plasticity 8.11 ~ 8.13 Summary by Byoung-Hee Kim Biointelligence Lab School of Computer Sci. & Eng. Seoul National University

(C) 2009 SNU CSE Biointelligence Lab Contents 8.11 Principal component neurons  Introductory remarks  Principal components and constrained optimization  Hebbian learning and synaptic constraints  Oja’s solution / Linsker’s model 8.12 Synaptic and phenomenological spin models  Phenomenological spin models  Synaptic models in the common input approximation 8.13 Objective function formulation of BCM theory  Projection pursuit  Objective function formulation of BCM theory 2

Goals and Contents Goal: the information-processing functions of model neurons in the visual system Contents  Principal component neurons  Special class of synaptic modification models  Relation to phenomenological spin models  Objective function formulation of BCM theory (C) 2009 SNU CSE Biointelligence Lab3

Introductory Remarks Images are highly organized spatial structures – some common statistical properties Development of the visual system is influenced by the statistical properties of the images  knowledge of the statistical properties of natural scenes ~ understanding the behavior of cells in the visual system (C) 2009 SNU CSE Biointelligence Lab4

Scale Invariance in Natural Images Studies of image statistics reveal non-preferrence of angular scale  Decimation procedure with the grey-valued pixels of the image assuming the role of the spins  p.d.f. of image constrasts and image gradients are unchanged  (Field 1987), (Ruderman and Bialek 1994), (Ruderman 1994) Representing the scale invariance through the covariance matrix  Gives a constraint on the form of the covariance matrix  Starting point for the PCA  (Hancock, et al. 1992), (Liu and Shouval 1995), (Liu and Shouval 1996) (C) 2009 SNU CSE Biointelligence Lab5

Principal Components We are rotating the coordinate system in order to find projections with desirable statistical properties Projections: maximally preserve information content while compressing the data into a few leading components (C) 2009 SNU CSE Biointelligence Lab6 Variance of the data projected onto the axis is maximal

Principal Components and Constrained Optimization (C) 2009 SNU CSE Biointelligence Lab7 n-component random vector correlation matrix If =0, then the covariance matrix Introducing a fixed vector that satisfies the normalization condition use this to help us find interesting projections Variance after operation: Optimization problem: find the vector a that satisfies the normalization condition, and maximizes the variance The variance is equal to the eigenvalue The maximum variance is given by the largest root

Hebbian Learning and Synaptic Constraints (C) 2009 SNU CSE Biointelligence Lab8 The simplest form Hebb’s rule for synaptic modification [Problem] Unstable. The synaptic weights would undergo unbounded growth c: output activity m: synaptic weight vector d: input activity vector On reaching a fixed point m is an eigenvector of the input correlation matrix with eigenvalue equal to zero

Solutions for the Unbounded Growth Problem Oja’s solution Linsker’s model (C) 2009 SNU CSE Biointelligence Lab9 On reaching a fixed point Results in a synaptic vector m for which the projection of the input activity has a maximum variance The synaptic system may be characterized as performing a principal component analysis of the input data constraint on the total synaptic strength Clipping - The sum of the synaptic weights are kept constant - each synaptic weight lies within a set range E Q : the variance in the input activity E k : constraint

Properties of the Linsker’s Model Stability corresponds to a global near minimum of the energy function Equivalent to the maximum in the input variance subject to the constraint Dynamics of the model system  In different regimes for the parameters k 1 and k 2, different receptive field structures dominate  As k 1 and k 2 are varied, particular eigenvectors other than the principal one gain in relative importance (C) 2009 SNU CSE Biointelligence Lab10

Synaptic and Phenomenological Spin Models Theory on synaptic modification  Model to explain the emergence of these highly ordered repeating structures (C) 2009 SNU CSE Biointelligence Lab11 Phenomena  Cells in the primate visual cortex self-organize onto ocular dominance columns and iso-orientation patches  The patterns observed experimentally are highly ordered

Phenomenological Spin Models 2D Ising lattice of eye-specificity encoding spints (Cowan and Friedman 1991)  Coupling strengths  If we take with, this type of coupling generates a short-range attraction plus a long-range repulsion between terminals from the same eye Hamiltonian for iso-orientation (C) 2009 SNU CSE Biointelligence Lab12

(C) 2009 SNU CSE Biointelligence Lab13

Synaptic Models in the Common Input Approximation Consider an LGN-cortico-cortico network with modifiable geniculocortico synapses and fixed cortico-cortico-connections Design of an energy function s.t. the fixed point of the network correspond to the minima of the energy function The common input model by Shouval and Cooper  hamiltonian in this model: (C) 2009 SNU CSE Biointelligence Lab14 general form correlational hamiltonian

Information-processing Activities by Common Input Neurons For exclusive excitatory connections  symmetry breaking does not occur  all receptive fields have the same orientation selectivity Inhibition  affects both the organization and structure of the receptive fields  If there is sufficient inhibition, the network will develop orientation selective receptive fields The cortical cells self-organize into iso-orientation patches with pinwheel singularities (C) 2009 SNU CSE Biointelligence Lab15

Objective Function Formulation of BCM Theory - Intro Distinguishment between information preservation (variance maximization) and classification (multimodality) (C) 2009 SNU CSE Biointelligence Lab16

Projection Pursuit Projection pursuit  a method for finding the most interesting low-dimensional features of high-dimensional data sets  The objective is to find orthogonal projections that reveal interesting structure in the data  PCA is a particular case with the proportion of total variance as the index of interestingness  Why is it needed? High-dimensional spaces are inherently sparse, or “curse of dimensionality” For classification purpose  Interesting projection is one that departs from normalcy (C) 2009 SNU CSE Biointelligence Lab17

Objective Function Formulation of BCM Theory (1/3) In the objective (energy) function formulation of BCM theory, a feature is associated with each projection direction A one-dimensional projection may be interpreted as a single feature extraction Goal: to find an objective (loss) function whose minimization produces a one-dimensional projection that is far from normal (C) 2009 SNU CSE Biointelligence Lab18

Objective Function Formulation of BCM Theory (2/3) (C) 2009 SNU CSE Biointelligence Lab19 Redefining the threshold function Synaptic modification functions How? Introduce a loss function With some assumptions

Objective Function Formulation of BCM Theory (3/3) (C) 2009 SNU CSE Biointelligence Lab20 The risk, or expected value of the loss, which is continously differentiable We are able to minimize the risk by means of gradient descent w.r.t. m i Slightly modified, deterministic version of the stochastic BCM modification equation A BCM neuron is extracting third-order statistical correlates of the data This would be a natural extension of principal component processing in the retina

Take-Home Message (Tomasso Poggio, NIPS 2007 tutorial) (C) 2009 SNU CSE Biointelligence Lab21