Synapses Signal is carried chemically across the synaptic cleft.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Pattern Recognition and Machine Learning
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Probability and Statistics Basic concepts II (from a physicist point of view) Benoit CLEMENT – Université J. Fourier / LPSC
Neural Computation Chapter 3. Neural Computation Outline Comparison of behavioral and neural response on a discrimination task –Bayes rule –ROC curves.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 4: Linear Models for Classification
Shin Ishii Nara Institute of Science and Technology
Reading population codes: a neural implementation of ideal observers Sophie Deneve, Peter Latham, and Alexandre Pouget.
How well can we learn what the stimulus is by looking at the neural responses? We will discuss two approaches: devise and evaluate explicit algorithms.
For stimulus s, have estimated s est Bias: Cramer-Rao bound: Mean square error: Variance: Fisher information How good is our estimate? (ML is unbiased:
Classification and risk prediction
Maximum likelihood (ML)
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
Spike Train decoding Summary Decoding of stimulus from response –Two choice case Discrimination ROC curves –Population decoding MAP and ML estimators.
Computer vision: models, learning and inference
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Population Coding Alexandre Pouget Okinawa Computational Neuroscience Course Okinawa, Japan November 2004.
Model Inference and Averaging
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
MML Inference of RBFs Enes Makalic Lloyd Allison Andrew Paplinski.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Projects: 1.Predictive coding in balanced spiking networks (Erwan Ledoux). 2.Using Canonical Correlation Analysis (CCA) to analyse neural data (David Schulz).
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Population coding Population code formulation Methods for decoding: population vector Bayesian inference maximum a posteriori maximum likelihood Fisher.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
BCS547 Neural Decoding.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Machine Learning 5. Parametric Methods.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 04: GAUSSIAN CLASSIFIERS Objectives: Whitening.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Maximum likelihood estimators Example: Random data X i drawn from a Poisson distribution with unknown  We want to determine  For any assumed value of.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Digital Communications I: Modulation and Coding Course Spring Jeffrey N. Denenberg Lecture 3c: Signal Detection in AWGN.
Bayesian Perception.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Estimator Properties and Linear Least Squares
DEEP LEARNING BOOK CHAPTER to CHAPTER 6
12. Principles of Parameter Estimation
LECTURE 06: MAXIMUM LIKELIHOOD ESTIMATION
Probability Theory and Parameter Estimation I
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
ICS 280 Learning in Graphical Models
Model Inference and Averaging
Ch3: Model Building through Regression
CH 5: Multivariate Methods
Maximum Likelihood Estimation
Special Topics In Scientific Computing
Bias and Variance of the Estimator
Computing and Statistical Data Analysis / Stat 8
Summarizing Data by Statistics
spike-triggering stimulus features
Timing Rules for Synaptic Plasticity Matched to Behavioral Function
Pattern Recognition and Machine Learning
Computing and Statistical Data Analysis / Stat 7
Volume 66, Issue 4, Pages (May 2010)
Artificial neurons Nisheeth 10th January 2019.
Parametric Methods Berlin Chen, 2005 References:
Timescales of Inference in Visual Adaptation
12. Principles of Parameter Estimation
Volume 74, Issue 1, Pages (April 2012)
Volume 66, Issue 4, Pages (May 2010)
Presentation transcript:

Synapses Signal is carried chemically across the synaptic cleft

Post-synaptic conductances Requires pre- and post-synaptic depolarization Coincidence detection, Hebbian

Synaptic plasticity LTP, LTD Spike-timing dependent plasticity

Short-term synaptic plasticity Depression Facilitation

A simple model neuron: FitzHugh-Nagumo I = 0 phase portrait

Phase portrait of the FitzHugh-Nagumo neuron model W V

Reduced dynamical model for neurons

Population coding Population code formulation Methods for decoding: population vector Bayesian inference maximum a posteriori maximum likelihood Fisher information

Cricket cercal cells coding wind velocity

Population vector RMS error in estimate Theunissen & Miller, 1991

Population coding in M1 Cosine tuning: Pop. vector: For sufficiently large N, is parallel to the direction of arm movement

The population vector is neither general nor optimal. “Optimal”: Bayesian inference and MAP

Bayesian inference By Bayes’ law, Introduce a cost function, L(s,sBayes); minimise mean cost. For least squares, L(s,sBayes) = (s – sBayes)2 ; solution is the conditional mean.

MAP: s* which maximizes p[s|r] ML: s* which maximizes p[r|s] MAP and ML MAP: s* which maximizes p[s|r] ML: s* which maximizes p[r|s] Difference is the role of the prior: differ by factor p[s]/p[r] For cercal data:

Decoding an arbitrary continuous stimulus E.g. Gaussian tuning curves

Need to know full P[r|s] Assume Poisson: Assume independent: Population response of 11 cells with Gaussian tuning curves

Apply ML: maximise P[r|s] with respect to s Set derivative to zero, use sum = constant From Gaussianity of tuning curves, If all s same

Apply MAP: maximise p[s|r] with respect to s Set derivative to zero, use sum = constant From Gaussianity of tuning curves,

Given this data: Prior with mean -2, variance 1 Constant prior MAP:

How good is our estimate? For stimulus s, have estimated sest Bias: Variance: Mean square error: Cramer-Rao bound: Fisher information

Fisher information Alternatively: For the Gaussian tuning curves w/Poisson statistics:

Fisher information for Gaussian tuning curves Quantifies local stimulus discriminability

Do narrow or broad tuning curves produce better encodings? Approximate: Thus,  Narrow tuning curves are better But not in higher dimensions!

Fisher information and discrimination Recall d' = mean difference/standard deviation Can also decode and discriminate using decoded values. Trying to discriminate s and s+Ds: Difference in estimate is Ds (unbiased) variance in estimate is 1/IF(s). 

Comparison of Fisher information and human discrimination thresholds for orientation tuning Minimum STD of estimate of orientation angle from Cramer-Rao bound data from discrimination thresholds for orientation of objects as a function of size and eccentricity