Modeling Your Spiking Data with Generalized Linear Models.

Slides:



Advertisements
Similar presentations
NORMAL OR GAUSSIAN DISTRIBUTION Chapter 5. General Normal Distribution Two parameter distribution with a pdf given by:
Advertisements

Copula Regression By Rahul A. Parsa Drake University &
CHAPTER 13 M ODELING C ONSIDERATIONS AND S TATISTICAL I NFORMATION “All models are wrong; some are useful.”  George E. P. Box Organization of chapter.
General Classes of Lower Bounds on Outage Error Probability and MSE in Bayesian Parameter Estimation Tirza Routtenberg Dept. of ECE, Ben-Gurion University.
2011 COURSE IN NEUROINFORMATICS MARINE BIOLOGICAL LABORATORY WOODS HOLE, MA GENERALIZED LINEAR MODELS Uri Eden BU Department of Mathematics and Statistics.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Spike Train Statistics Sabri IPM. Review of spike train  Extracting information from spike trains  Noisy environment:  in vitro  in vivo  measurement.
Integration of sensory modalities
Model Assessment, Selection and Averaging
An Introductory to Statistical Models of Neural Data SCS-IPM به نام خالق ناشناخته ها.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Simulation Modeling and Analysis
Today Linear Regression Logistic Regression Bayesians v. Frequentists
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Machine Learning CMPT 726 Simon Fraser University
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Analysis of Simulation Input.. Simulation Machine n Simulation can be considered as an Engine with input and output as follows: Simulation Engine Input.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Bayesian kriging Instead of estimating the parameters, we put a prior distribution on them, and update the distribution using the data. Model: Prior: Posterior:
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
7. Nonparametric inference  Quantile function Q  Inference on F  Confidence bands for F  Goodness- of- fit tests 1.
Lecture II-2: Probability Review
Review of Lecture Two Linear Regression Normal Equation
GENERALIZED LINEAR MODELS
沈致远. Test error(generalization error): the expected prediction error over an independent test sample Training error: the average loss over the training.
On Model Validation Techniques Alex Karagrigoriou University of Cyprus "Quality - Theory and Practice”, ORT Braude College of Engineering, Karmiel, May.
Statistics for Engineer Week II and Week III: Random Variables and Probability Distribution.
CS433: Modeling and Simulation Dr. Anis Koubâa Al-Imam Mohammad bin Saud University 15 October 2010 Lecture 05: Statistical Analysis Tools.
Non-life insurance mathematics Nils F. Haavardsson, University of Oslo and DNB Skadeforsikring.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
Chapter 9 Input Modeling Banks, Carson, Nelson & Nicol Discrete-Event System Simulation.
Part II: Model Class Selection Given: Dynamic data from system and set of candidate model classes where each model class defines a set of possible predictive.
Machine Learning CUNY Graduate Center Lecture 4: Logistic Regression.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Generalized Linear Models (GLMs) and Their Applications.
Lecture 2: Statistical learning primer for biologists
Sampling and estimation Petter Mostad
Mutual Information Scheduling for Ranking Hamza Aftab Nevin Raj Paul Cuff Sanjeev Kulkarni Adam Finkelstein 1.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
Ch 3. Likelihood Based Approach to Modeling the Neural Code Bayesian Brain: Probabilistic Approaches to Neural Coding eds. K Doya, S Ishii, A Pouget, and.
CHARACTERIZATION OF NONLINEAR NEURON RESPONSES AMSC 664 Final Presentation Matt Whiteway Dr. Daniel A. Butts Neuroscience.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Computacion Inteligente Least-Square Methods for System Identification.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Estimating standard error using bootstrap
Modeling and Simulation CS 313
HST 583 fMRI DATA ANALYSIS AND ACQUISITION
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
Modeling and Simulation CS 313
2016 METHODS IN COMPUTATIONAL NEUROSCIENCE COURSE
ICS 280 Learning in Graphical Models
Ch3: Model Building through Regression
Statistical Learning Dong Liu Dept. EEIS, USTC.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
Integration of sensory modalities
Cross-validation for the selection of statistical models
Computing and Statistical Data Analysis / Stat 7
Parametric Methods Berlin Chen, 2005 References:
Mathematical Foundations of BME Reza Shadmehr
12. Principles of Parameter Estimation
Applied Statistics and Probability for Engineers
Introductory Statistics
Empirical Distributions
Presentation transcript:

Modeling Your Spiking Data with Generalized Linear Models

Outline I.Part One: Specifying models a.The generalized linear model form b.Fitting parameters II.Part Two: Evaluating model quality a.Parameter Uncertainty b.Relative: Akaike info. criterion (AIC) c.Absolute: Time rescaling theorem

“covariate” What are Spiking Models?

Neuron 1 Neuron 2 Neuron 3 Today’s Case Study: CA1 Hippocampal Place Cells

Challenges in Modeling Spikes Your neurons are in a vast network Each neuron has many biophysical parameters Biological noise Memory distributed across the network Solution: Stochastic Models

Data Likelihood & Maximum Likelihood (ML) Parameter Fitting

Introducing Generalized Linear Models System EEG, LFP covariates Linear Regression (Gaussian Model of Variability) System spikes covariates

Stochastic Models Linear Regression Properties of GLM: Convex likelihood surface Estimators asymptotically have minimum MSE Generalized Linear Models (GLM) Neural Spiking Models

Point Process Data Likelihood Intensity Model: Observed Spike Data: Data Likelihood, one neuron: The Conditional Intensity Function: In Discrete Time:

The Generalized Linear Model To establish a Generalized Linear Model set the natural parameter to be a linear function of the covariates In this case: The Exponential Family of Distributions This model form makes ML parameter estimation easy.

ML estimation is easier with this GLM: Convex Likelihood

Convex Likelihood Intensity Model

Fitting GLM In general, no closed form ML estimator Use numerical optimization, such as MATLAB’s glmfit

1.extends the concept of linear regression 2.allows for modeling spikes 3.easy to fit model parameters Summary: GLM, Part I

1: Plot raw data >> load glm_data.mat >> who

Example call to glmfit b = glmfit([xN yN],spikes_binned,'poisson');

Visualizing model from glmfit lambda = exp(b(1) + b(2)*x_new + b(3)*y_new);

Quadratic model with glmfit b = glmfit([xN yN xN.^2 yN.^2 xN.*yN],… spikes_binned,'poisson'); Visualizing this model from glmfit lambda = exp(b(1) + b(2)*x_new … + b(3)*y_new + b(4)*x_new.^2 … + b(5)*y_new.^2 … + b(6)*x_new.*y_new);

Fitting other covariates with glmfit b=glmfit([vxN vxN.^2],spikes_binned,'poisson'); Visualizing this model from glmfit plot(velocities,exp(b(1)+b(2)*velocities … +b(3)*velocities.^2),'r');

Part II: Model Quality 1.Parameter uncertainty 2.Relative: Akaike Info. Criterion (AIC) 3.Absolute: Time rescaling theorem

ML Parameter Uncertainty

>> load glm_data.mat >> [b,dev,stats] = glmfit( put inputs here); stats.se – standard error on parameter estimates stats.p – significance of parameter estimates errorbar(1:length(b), b, 2*stats.se );

Akaike Info. Criterion Approximates relative Kullback-Leibler divergence between real distribution and model Formula:

Computing AIC from glmfit >> [b,dev,stats] = glmfit( put inputs here); >> AIC = dev+2*length(b)

5. Time Rescaling Theorem Let’s return to the probability distribution for a single ISI: and make a change of variables, Then the distribution is exponential with parameter.

Illustration: Time Rescaling t1t1 z1z1 z2z2 z3z3 t3t3 t2t2 Rescaled Time Intensity (Hz)

Time Rescaling Statistic: KS Plots Graphical measure of goodness-of-fit, based on time rescaling, comparing an empirical and model cumulative distribution function. If the model is correct, then the rescaled ISIs are independent, identically distributed random variables whose ordered quantiles should produce a 45° line [Ogata, 1988]. Uniform CDF (Model Predicted ISI CDF) Empirical Rescaled ISI CDF

Construct KS Plot for Linear Model >> glm_ks

Construct KS Plot for Quadratic Model Modify glm_ks.m lambdaEst = exp( b_quad(1) + b_quad(2)*xN + b_quad(3)*yN + b_quad(4)*xN.^2 + b_quad(5)*yN.^2 + b_quad(6)*xN.*yN );

Questions to Ask by Modeling Spikes What stimulus features drive the spiking? How does the spiking vary with a covariate (stimulus, behavioral measure, cognitive state, or spike history,)? How well does the spiking represent a covariate ? How do neurons or brain regions cooperate in spiking? What physical processes determine the spiking?

Summary: GLM, Part II 1.extends the concept of linear regression 2.allows for modeling spikes 3.easy to fit model parameters 4.Fisher information and parameter uncertainty 5.Relative model quality: AIC 6.Absolute model quality: Time Rescaling