Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC 2001. Lecture 7: Coding and Representation 1 Computational Architectures in.

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

What is the neural code? Puchalla et al., What is the neural code? Encoding: how does a stimulus cause the pattern of responses? what are the responses.
What is the neural code?. Alan Litke, UCSD Reading out the neural code.
Face Recognition Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Dimensionality Reduction PCA -- SVD
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Reading population codes: a neural implementation of ideal observers Sophie Deneve, Peter Latham, and Alexandre Pouget.
How well can we learn what the stimulus is by looking at the neural responses? We will discuss two approaches: devise and evaluate explicit algorithms.
For stimulus s, have estimated s est Bias: Cramer-Rao bound: Mean square error: Variance: Fisher information How good is our estimate? (ML is unbiased:
Supervised and Unsupervised learning and application to Neuroscience Cours CA6b-4.
Principal Component Analysis
Application of Statistical Techniques to Neural Data Analysis Aniket Kaloti 03/07/2006.
Self Organization: Hebbian Learning CS/CMPE 333 – Neural Networks.
Dimensional reduction, PCA
Spike-triggering stimulus features stimulus X(t) multidimensional decision function spike output Y(t) x1x1 x2x2 x3x3 f1f1 f2f2 f3f3 Functional models of.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 6: Low-level features 1 Computational Architectures in Biological.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Multivariate Analysis A Unified Perspective
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks I PROF. DR. YUSUF OYSAL.
Radial Basis Function Networks
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
Summarized by Soo-Jin Kim
Multivariate Approaches to Analyze fMRI Data Yuanxin Hu.
Population Coding Alexandre Pouget Okinawa Computational Neuroscience Course Okinawa, Japan November 2004.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
Lecture 10: Mean Field theory with fluctuations and correlations Reference: A Lerchner et al, Response Variability in Balanced Cortical Networks, q-bio.NC/ ,
Basics of Neural Networks Neural Network Topologies.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
What is the neural code?. Alan Litke, UCSD What is the neural code?
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
BCS547 Neural Decoding.
Principal Manifolds and Probabilistic Subspaces for Visual Recognition Baback Moghaddam TPAMI, June John Galeotti Advanced Perception February 12,
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
Fast Learning in Networks of Locally-Tuned Processing Units John Moody and Christian J. Darken Yale Computer Science Neural Computation 1, (1989)
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Principle Component Analysis and its use in MA clustering Lecture 12.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
Information Processing by Neuronal Populations Chapter 6: Single-neuron and ensemble contributions to decoding simultaneously recoded spike trains Information.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Principal Components Analysis ( PCA)
Bayesian Perception.
Independent Component Analysis features of Color & Stereo images Authors: Patrik O. Hoyer Aapo Hyvarinen CIS 526: Neural Computation Presented by: Ajay.
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Data statistics and transformation revision Michael J. Watts
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
LECTURE 11: Advanced Discriminant Analysis
University of Ioannina
Principal Component Analysis (PCA)
Dimension Reduction via PCA (Principal Component Analysis)
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
spike-triggering stimulus features
Introduction to Radial Basis Function Networks
Parametric Methods Berlin Chen, 2005 References:
Volume 74, Issue 1, Pages (April 2012)
Lecture 16. Classification (II): Practical Considerations
Presentation transcript:

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in Biological Vision, USC, Spring 2001 Lecture 7. Coding and Representation Reading Assignments: None

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 2 Variability in neuronal responses Neuronal response f(.) can be represented as the sum of: - a mean response to given stimuli s, given by the neuron’s tuning t(.) - a noise term (that may also be stimulus-dependent) n(.) f(s) = t(s) + n(s) Examples of sources of variability (n(.) term): - transduction noise at the sensors (e.g., quantum light absorption noise) - synaptic noise - cross-talk noise due to activity in neighboring neurons - random fluctuations in ionic channel function

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 3 Mean response Mean response exhibits a tuning curve, typically well approximated by a Gaussian or cosine function.

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 4 Noise term but responses to exactly identical stimuli can vary greatly from trial to trial… In a first approximation, we consider the variability in spike count and timing, and discard any variability in the exact shape of each action potential.

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 5 Spike Variability

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 6 Estimating Firing Rates Basic idea: count spikes in a small time interval. Rate (frequency)= spike count / interval.

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 7 Population Code Each neuron is tuned to specific stimuli. Thus, a population of neurons with different preferred stimuli can represent all possible stimuli.

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 8 Population Vector - Represent each preferred stimulus by vector in stimulus space. - Population vector: weighted sum of preferred vectors (weighted by response)

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 9 Population Vector Works very well in many cases!

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 10 Noise limits discrimination performance a: population response to two stimuli with different directions b: population response to two stimuli with similar directions

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 11 Optimal Decoding Use results from statistical estimation theory to infer a theoretical bound on how well any decoder can perform. Key theoretical result: Cramer-Rao bound - Consider a random variable X which depends on a parameter p - Problem: try to recover p from the random observations of X - Result: no matter what (unbiased) function we use to estimate p from observations x 1 …x n of X, we will not be able to recover p perfectly… Rather, we will recover p up to a given accuracy bound (the Cramer-Rao bound) which we can compute formally for any unbiased estimator. Of all possible estimators, the Maximum Likelihood estimator comes very close to this ideal bound. Note: unbiased means that the estimator will on average yield p, not 0.9*p or 1.1*p

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 12 Population Code a: tuning curves of neural population b: responses of all neurons to one stimulus c: population vector d: maximum- likelihood estimate

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 13 Neural implementations A close approximation to the Maximum-Likelihood estimator can be built using a neural network (Pouget et al., Nature Reviews Neuroscience, 2000). Input: noisy hill (bottom) Output: clean hill whose mean is the ML estimate of stimulus orientation. Difficulty: choice of weights depends on width of input hill…

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 14 Case study: optimal estimation Problem: record psychophysical data but simulate neurons; how can we predict performance of the whole observer based on responses of a few neurons? Solution: train observers, and assume that they will approach an optimal decoder. Then do not worry about the implementation of the decoder itself. Rather, just use the result that it will perform close to the theoretical limit given by the Cramer-Rao bound.

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 15

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 16

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 17

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 18

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 19

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 20

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 21

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 22

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 23 How realistic is this optimal decoder? In some studies, single- neuron firing predicts very well the behavior of the animal!

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 24

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 25

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 26 Population Codes May Not be Obvious! Sometimes responses to different stimuli will have non-obvious characteristics: e.g., variations in phase of oscillatory response…

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 27 Time-Dependent Coding Neural responses often are not restricted to a single discharge event just following a stimulus, but may extend over time in complex fluctuations…

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 28 What more can we do? One good thing to do is to pre-process the data such as to eliminate redundancy and represent it in its most “natural” basis. How can we do that? Principal Component Analysis Independent Component Analysis Clustering techniques

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 29 Principal Component Analysis Goal: transforms a set of correlated variables into a smaller set of uncorrelated variables. Approach: find eigenvectors of covariance matrix and operate a change of basis along a subset of those vectors that correspond to non-negligible eigenvalues. Properties: - removes redundancy in data - finds “true dimensionality” of dataset - linear transformation, and hence preserves distance

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 30 PCA example Original data is skewed along one axis. Transformed data is unskewed.

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 31 Independent Components Analysis Goal: extract independent sources given only observed data as a mixture of the unknown sources. Approach: like PCA, ICA also decorrelates the signals and reduces dimensionality. The main difference is that the new coordinate system may not be orthogonal. Algorithm works by maximizing mutual information between assumed input sources and observed output data. Properties: same general properties as PCA, with the additional freedom of possibly non-orthogonal principal directions.

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 32 ICA outline

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 33 PCA vs. ICA: Example PCA Result

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 34 PCA vs. ICA: Example ICA Result

Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 35 Implications for Computer Architectures Population codes are omnipresent in cortex… Advantages: very robust to noisy signals very robust to isolated failing neurons Drawback: need a decoder! but that may not be a real drawback if the decoder has a clear physical meaning and is fast and easy to compute: e.g., population vector.