1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Brief introduction on Logistic Regression
INTRODUCTION TO MACHINE LEARNING Bayesian Estimation.
Pattern Recognition and Machine Learning
Biointelligence Laboratory, Seoul National University
Pattern Recognition and Machine Learning
Neural Computation Chapter 3. Neural Computation Outline Comparison of behavioral and neural response on a discrimination task –Bayes rule –ROC curves.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Chapter 4: Linear Models for Classification
Population Codes & Inference in Neurons
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Visual Recognition Tutorial
Reading population codes: a neural implementation of ideal observers Sophie Deneve, Peter Latham, and Alexandre Pouget.
How well can we learn what the stimulus is by looking at the neural responses? We will discuss two approaches: devise and evaluate explicit algorithms.
For stimulus s, have estimated s est Bias: Cramer-Rao bound: Mean square error: Variance: Fisher information How good is our estimate? (ML is unbiased:
Pattern Recognition and Machine Learning
Supervised and Unsupervised learning and application to Neuroscience Cours CA6b-4.
Application of Statistical Techniques to Neural Data Analysis Aniket Kaloti 03/07/2006.
Image processing. Image operations Operations on an image –Linear filtering –Non-linear filtering –Transformations –Noise removal –Segmentation.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Baysian Approaches Kun Guo, PhD Reader in Cognitive Neuroscience School of Psychology University of Lincoln Quantitative Methods 2011.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
PATTERN RECOGNITION AND MACHINE LEARNING
How do neurons deal with uncertainty?
Population Coding Alexandre Pouget Okinawa Computational Neuroscience Course Okinawa, Japan November 2004.
STUDY, MODEL & INTERFACE WITH MOTOR CORTEX Presented by - Waseem Khatri.
Neural coding (1) LECTURE 8. I.Introduction − Topographic Maps in Cortex − Synesthesia − Firing rates and tuning curves.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Geo597 Geostatistics Ch9 Random Function Models.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
CSC 2535 Lecture 8 Products of Experts Geoffrey Hinton.
Multifactor GPs Suppose now we wish to model different mappings for different styles. We will add a latent style vector s along with x, and define the.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
Population coding Population code formulation Methods for decoding: population vector Bayesian inference maximum a posteriori maximum likelihood Fisher.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Neural Modeling - Fall NEURAL TRANSFORMATION Strategy to discover the Brain Functionality Biomedical engineering Group School of Electrical Engineering.
BCS547 Neural Decoding.
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
An Introduction to Kalman Filtering by Arthur Pece
Chapter 7. Learning through Imitation and Exploration: Towards Humanoid Robots that Learn from Humans in Creating Brain-like Intelligence. Course: Robots.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Chapter 3. Stochastic Dynamics in the Brain and Probabilistic Decision-Making in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning.
Ch.9 Bayesian Models of Sensory Cue Integration (Mon) Summarized and Presented by J.W. Ha 1.
Machine Learning 5. Parametric Methods.
Tracking with dynamics
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Optimal Eye Movement Strategies In Visual Search.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Neural Codes. Neuronal codes Spiking models: Hodgkin Huxley Model (brief repetition) Reduction of the HH-Model to two dimensions (general) FitzHugh-Nagumo.
Bayesian Perception.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Bayesian Brain - Chapter 11 Neural Models of Bayesian Belief Propagation Rajesh P.N. Rao Summary by B.-H. Kim Biointelligence Lab School of.
1 Nonlinear models for Natural Image Statistics Urs Köster & Aapo Hyvärinen University of Helsinki.
Deep Feedforward Networks
Ch3: Model Building through Regression
Machine Learning Basics
Special Topics In Scientific Computing
Synapses Signal is carried chemically across the synaptic cleft.
Xaq Pitkow, Dora E. Angelaki  Neuron 
Confidence as Bayesian Probability: From Neural Origins to Behavior
Parametric Methods Berlin Chen, 2005 References:
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
Information Processing by Neuronal Populations Chapter 5 Measuring distributed properties of neural representations beyond the decoding of local variables:
NON-NEGATIVE COMPONENT PARTS OF SOUND FOR CLASSIFICATION Yong-Choon Cho, Seungjin Choi, Sung-Yang Bang Wen-Yi Chu Department of Computer Science &
Volume 74, Issue 1, Pages (April 2012)
Presentation transcript:

1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan, and Richard S. Zemel Annual review of neuroscience 2003 Presenter : Sangwook Hahn, Jisu Kim

2 / 41 Inference and Computation with Population Codes 13 November 2012 Outline 1.Introduction 2.The Standard Model ( First Part ) 1.Coding and Decoding 2.Computation with Population Codes 3.Discussion of Standard Model 3.Encoding Probability Distributions ( Second Part ) 1.Motivation 2.Psychophysical Evidence 3.Encoding and Decoding Probability Distributions 4.Examples in Neurophysiology 5.Computations Using Probabilistic Population Codes

3 / 41 Inference and Computation with Population Codes 13 November 2012 Introduction  Single aspects of the world –(induce)> activity in multiple neurons  For example –1. Air current is occurred by predator of cricket –2. Determine the direction of an air current –3. Evade with other direction from predicted predator’s move air current

4 / 41 Inference and Computation with Population Codes 13 November 2012 Introduction  Analyze the example at the view of neural activity –1. Air current is occurred by predator of cricket –2. Determine the direction of an air current ( i. population of neurons encode information about single variable ii. information decoded from population activity ) –3. Evade with other direction from predicted predator’s move air current

5 / 41 Inference and Computation with Population Codes 13 November 2012 Guiding Questions (At First Part)  Q1: How do populations of neurons encode information about single variables? How this information can be decoded from the population activity? How do neural populations realize function approximation?  Q2: How population codes support nonlinear computations over the information they represent?

6 / 41 Inference and Computation with Population Codes 13 November 2012 The Standard Model – Coding  Cricket cercal system has hair cells (a) as primary sensory neurons  Normalized mean firing rates of 4 low-velocity interneurons  s is the direction of an air current (induced by predator)

7 / 41 Inference and Computation with Population Codes 13 November 2012 The Standard Model – Encoding Model  Mean activity of cell a depends on s – : maximum firing rate – : preferred direction of cell a  Natural way of describing tuning curves –proportional to the threshold projection of v onto

8 / 41 Inference and Computation with Population Codes 13 November 2012 The Standard Model – Decoding  3 methods to decode homogeneous population codes –1. Population vector approach –2. Maximum likelihood decoding –3. Bayesian estimator  Population vector approach ( sum ) – : population vector – : preferred direction – : actual rates from the mean rates – : approximation of wind direction (r is noisy rates)

9 / 41 Inference and Computation with Population Codes 13 November 2012 The Standard Model – Decoding  Main problem of population vector method –It is not sensitive to the noise process that generates –However, it works quite well –Estimation of wind direction to within a few degrees is possible only with 4 noisy neurons

10 / 41 Inference and Computation with Population Codes 13 November 2012 The Standard Model – Decoding  Maximum likelihood decoding –This estimator starts from the full probabilistic encoding model by taking into account the noise corrupting neurons activities –A –If is high -> those s values are likely to the observed activities –If is low -> those s values are unlikely to the observed activities rms = root mean square deviation

11 / 41 Inference and Computation with Population Codes 13 November 2012 The Standard Model – Decoding  Bayesian estimators –Combine likelihood P[r|s] with any prior information about stimulus s to produce a posterior distribution P[s|r] : –If prior distribution P[s] is flat, there is no specific prior information of s and this is renormalization version of likelihood –Bayesian estimator does a little better than maximum likelihood and population vector

12 / 41 Inference and Computation with Population Codes 13 November 2012 The Standard Model – Decoding  In homogenous population –Bayesian & Maximum likelihood decoding >>> population vector –‘the greater the number of cells is, the greater the accuracy is’ since more cells can provide more information about stimulus

13 / 41 Inference and Computation with Population Codes 13 November 2012 Computation with Population Code  Discrimination –If there are and where is a small angle, we can use Bayesian poesterior (P[s|r]) in order to discriminate those –It is also possible to perform discrimination based directly on activities by computing a linear : – : usually 0 for a homogeneous population code – : Relative weight

14 / 41 Inference and Computation with Population Codes 13 November 2012 Computation with Population Code  Noise Removal –Maximum likelihood estimator is unclear about its neurobiological relevance. 1. finding a single scalar value seems unreasonable because population codes seem to be used throughout the brain 2. while finding maximum likelihood value is difficult in general –Solution : utilizing recurrent connection within population to make it behave like an autoassociative memory Autoassociative memories use nonlinear recurrent interactions to find the stored pattern that most closely matches a noisy input

15 / 41 Inference and Computation with Population Codes 13 November 2012 Computation with Population Code  Basis Function Computations –Function approximation compute the output of functions for the case of multiple stimulus dimensions. –For example, –s h : head-centered direction to a target s r : eye-centered direction s e : position of eyes in the head

16 / 41 Inference and Computation with Population Codes 13 November 2012 Computation with Population Code  Basis Function Computations

17 / 41 Inference and Computation with Population Codes 13 November 2012 Computation with Population Code  Basis Function Computations –linear solution for homogeneous population codes (mapping from one population code to another, ignoring noise )

18 / 41 Inference and Computation with Population Codes 13 November 2012 Guiding Questions (At First Part)  Q1: How do populations of neurons encode information about single variables? -> p.6~7 How this information can be decoded from the population activity? -> p.8~12 How do neural populations realize function approximation? -> p.13~14  Q2: How population codes support nonlinear computations over the information they represent? -> p.15~17

19 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding Probability Distributions

20 / 41 Inference and Computation with Population Codes 13 November 2012 Motivation  The standard model has two main restrictions :  We only consider uncertainty coming from noisy neural activities. (internal noise) : Uncertainty is inherent, independent of internal noise.  We do not consider anything other than estimating the single value. : Utilizing the full information contained in the posterior is crucial.

21 / 41 Inference and Computation with Population Codes 13 November 2012 Motivation  “ill-posed problems” : images do not contain enough information.  The aperture problem. : Images does not unambiguously specify the motion of the object.  Solution - probabilistic approach. : perception is conceived as statistical inference giving rise to probability distributions over the values.

22 / 41 Inference and Computation with Population Codes 13 November 2012 Motivation

23 / 41 Inference and Computation with Population Codes 13 November 2012 Psychophysical Evidence

24 / 41 Inference and Computation with Population Codes 13 November 2012 Psychophysical Evidence

25 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions

26 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions

27 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions

28 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions  Convolution encoding :  Can deal with non-Gaussian distributions that cannot be characterized by a few parameters, such as their means and variances.  Represent the distribution using a convolution code, obtained by convolving the distribution with a particular set of kernel functions.

29 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions

30 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions  Use large neuronal population of neurons to encode any function by devoting each neuron to the encoding of one particular coefficient.  The activity of neuron a is computed by taking the inner product between a kernel function assigned to that neuron and the function being encoded.

31 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions

32 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions

33 / 41 Inference and Computation with Population Codes 13 November 2012 Encoding and Decoding Probability Distributions

34 / 41 Inference and Computation with Population Codes 13 November 2012 Examples in Neurophysiology  Uncertainty in 2-AFC (2-alternative forced choice) : examples offer preliminary evidence that neurons represent probability distributions, or related quantities, such as log likelihood ratios.  There are also experiments supporting gain encoding, convolution codes, and DDPC, respectively.

35 / 41 Inference and Computation with Population Codes 13 November 2012 Computations Using Probabilistic Population Codes

36 / 41 Inference and Computation with Population Codes 13 November 2012 Computations Using Probabilistic Population Codes  If we use convolution code for all distributions –multiply all the population codes together term by term –requires neurons that can multiply or sum : achievable neural operation  If the probability distributions are encoded using the position and gain of population codes –Solution : Deneve et al. (2001) –Some limitations –Performs a Bayesian inference using noisy population codes

37 / 41 Inference and Computation with Population Codes 13 November 2012 Computations Using Probabilistic Population Codes

38 / 41 Inference and Computation with Population Codes 13 November 2012 Guiding Questions(At Second Part)  Q3: How may neural populations offer a rich representation of such things as uncertainty in the aspects of the stimuli they represent?  # 21 ~ # 24  Probabilistic approach : perception is conceived as statistical inference giving rise to probability distributions over the values.  Hence stimuli of neural populations represents probability distributions, which gives information of uncertainty.

39 / 41 Inference and Computation with Population Codes 13 November 2012 Guiding Questions(At Second Part)  Q4: How can populations of neurons represent probability distributions? How can they perform Bayesian probabilistic inference?  #25 ~ #31 (for first), #37 ~ #39 (for second)  Several schemes have been proposed for encoding probability distributions in populations of neurons : Log-likelihood method, Gain encoding for Gaussian distributions, Convolution encoding.  Bayesian probabilistic inference can be done by multiply all the population codes (convolution encoding), or using noisy population codes (gain encoding)

40 / 41 Inference and Computation with Population Codes 13 November 2012 Guiding Questions(At Second Part)  Q5: How multiple aspects of the world are represented in single populations? What computational advantages (or disadvantages) such schemes have?  # 25 ~ # 28 (first)  Log-likelihood : likelihood Gain encoding : mean and standard deviation Convolution encoding : probability distribution

41 / 41 Inference and Computation with Population Codes 13 November 2012 Guiding Questions(At Second Part)  Q5: How multiple aspects of the world are represented in single populations? What computational advantages (or disadvantages) such schemes have?  # 25 ~ # 28 (second)  Log-likelihood : decoding is simple, but some distribution limitation Gain encoding : strong distribution limitation. Convolution encoding : can work for complicated distribution.