Presentation is loading. Please wait.

Presentation is loading. Please wait.

2016 METHODS IN COMPUTATIONAL NEUROSCIENCE COURSE

Similar presentations


Presentation on theme: "2016 METHODS IN COMPUTATIONAL NEUROSCIENCE COURSE"β€” Presentation transcript:

1 2016 METHODS IN COMPUTATIONAL NEUROSCIENCE COURSE
MARINE BIOLOGICAL LABORATORY WOODS HOLE, MA Brief introduction to probability theory, information measures, and latent variables Uri Eden BU Department of Mathematics and Statistics August 1, 2016

2 Probability and statistics
Probability theory uses probability models to describe and predict features of experimental data e.g. Assuming 55% of voters favor Clinton, what range of outcomes could we expect of a random poll of 1000 voters? Statistics uses data to estimate and infer features of a probability model that could have generated the data e.g. If a random poll of 1000 voters has 550 favor Clinton, what is a range of reasonable values for the fraction of voters in the population who favor her?

3 A case study Iygengar and Liu (1997) examined the activity of retinal neurons grown in culture under constant light and environmental conditions. Spontaneous spiking activity of these neurons was recorded. The objective is to develop a probability model which accurately describes the stochastic structure of this activity. In particular, we seek a model of the interspike interval (ISI) distribution.

4

5

6 Discrete vs continuous random variables
Discrete variables can only take on values in a discrete (finite or countably infinite) set. e.g. Result of coin flip, π‘‹βˆˆ H,T Number of spikes in an interval, π‘‹βˆˆ 0,1,2,… Continuous variables take on values in continuous intervals. e.g. Phase of a rhythm at which a spike occurs, π‘‹βˆˆ(βˆ’πœ‹,πœ‹] Interspike interval between two spikes, π‘‹βˆˆ[0,∞)

7 Discrete variables are defined by a pmf or CDF
probability mass function (pmf): 𝑝 π‘₯ = Pr 𝑋=π‘₯ =𝑃 π‘₯ 𝑝(π‘₯) π‘₯ 𝑝 π‘₯ =1 π‘₯ Cumulative Distribution Function (CDF): 𝐹(π‘₯) π‘₯ 𝐹 π‘₯ =Pr⁑(𝑋≀π‘₯) = π‘₯ 𝑖 ≀π‘₯ 𝑝 π‘₯ 𝑖

8 Continuous variables are defined by a pdf or CDF
probability density function (pdf): 𝑓(π‘₯)=𝑃 π‘₯ Pr 𝑋=π‘₯ =0 Pr π‘Ž<𝑋<𝑏 = π‘Ž 𝑏 𝑓 π‘₯ 𝑑π‘₯ 𝑓(π‘₯) π‘₯ 𝑓 π‘₯ 𝑑π‘₯ =1 π‘₯ Cumulative Distribution Function (CDF): 𝐹(π‘₯) π‘₯ 𝐹 π‘₯ = Pr 𝑋≀π‘₯ = π‘₯ 𝑖 ≀π‘₯ 𝑓 π‘₯ 𝑖 𝑑 π‘₯ 𝑖

9 Expected values Discrete variables: Continuous variables:
𝐸 𝑋 = π‘₯ π‘₯𝑓 π‘₯ 𝑑π‘₯ 𝐸 𝑋 = π‘₯ π‘₯𝑝 π‘₯ πœ‡= 𝑓(π‘₯) 𝐸 𝑔(𝑋) = π‘₯ 𝑔(π‘₯)𝑓 π‘₯ 𝑑π‘₯ 𝐸 𝑔(𝑋) = π‘₯ 𝑔(π‘₯)𝑝 π‘₯ Var(𝑋)=𝐸 π‘‹βˆ’πœ‡ 2

10 Question What probability model should we use to describe the retinal spiking data?

11 Nonparametric model: the empirical CDF
The empirical CDF (eCDF) is a function of π‘₯ that computes the fraction of data less than or equal to π‘₯. 𝐹 π‘₯ = 1 𝑛 𝑖=1 𝑛 1 π‘₯ 𝑖 ≀π‘₯ 𝐹 (π‘₯) π‘₯ Dvoretzky–Kiefer–Wolfowitz (DKW) inequality: Pr⁑ sup π‘₯ 𝐹 π‘₯ βˆ’ 𝐹 π‘₯ >πœ€ ≀2 𝑒 βˆ’2𝑛 πœ€ 2 Pr⁑ sup π‘₯ 𝐹 π‘₯ βˆ’ 𝐹 π‘₯ > 𝑛 ≀0.95

12 Expected values of eCDF
The eCDF defines a discrete probability model that places probability mass 1 𝑛 at each observed data point. 𝐸 𝑋 = π‘₯ π‘₯ 𝑝 π‘₯ = 1 𝑛 𝑖=1 𝑛 π‘₯ 𝑖 = π‘₯ 𝐸 𝑔(𝑋) = 1 𝑛 𝑖=1 𝑛 𝑔( π‘₯ 𝑖 ) Var(𝑋)= π‘₯ π‘₯βˆ’πΈ[𝑋] 2 𝑝 (π‘₯) = 1 𝑛 𝑖=1 𝑛 π‘₯ 𝑖 βˆ’ π‘₯ 2

13 Parametric models: common discrete distributions
𝑝 π‘₯ = 1 𝑛 Discrete uniform Bernoulli Binomial Poisson π‘₯∈ 1,…,𝑛 𝑝 π‘₯ = 𝑝 π‘₯ 1βˆ’π‘ 1βˆ’π‘₯ π‘₯∈ 0,1 𝑝 π‘₯ = 𝑛 π‘₯ 𝑝 π‘₯ 1βˆ’π‘ π‘›βˆ’π‘₯ π‘₯∈ 0,…,𝑛 𝑝 π‘₯ = πœ† π‘₯ 𝑒 βˆ’πœ† π‘₯! π‘₯∈ 0,…,∞

14 Parametric models: common continuous distributions
𝑓 π‘₯ = 1 π‘βˆ’π‘Ž Uniform Normal Exponential Gamma π‘₯∈[π‘Ž,𝑏] 𝑓 π‘₯ = πœ‹ 𝜎 2 𝑒 βˆ’ π‘₯βˆ’πœ‡ 𝜎 2 π‘₯∈(βˆ’βˆž,∞) 𝑓 π‘₯ = 1 πœƒ 𝑒 βˆ’ π‘₯ πœƒ π‘₯∈[0,∞) 𝑓 π‘₯ ∝ π‘₯ π›Όβˆ’1 𝑒 βˆ’π›½π‘₯ π‘₯∈[0,∞)

15 Nonparametric vs parametric models
Nonparametric models and methods require fewer assumptions about data typically involve less complicated estimation methods Parametric models and methods tend to have more statistical power often have well-studied, well-behaved properties involve parameters that often have useful interpretations tend to scale better with large, high-dimensional datasets

16 Parametric probability models for ISI data
𝑓 π‘₯ = 1 πœƒ 𝑒 βˆ’ π‘₯ πœƒ Exponential model: Which model (which value of πœƒ) should we pick? 𝑓 π‘₯ ISI (ms)

17 Model fitting – parameter estimation
Two related perspectives on parameter estimation Minimize some cost function e.g. πœƒ = min πœƒ 𝐸 πœƒ βˆ’πœƒ 2 Select estimators with known properties and distributions. e.g. unbiasedness, minimum variance, asymptotic normality Best of both worlds: Maximum Likelihood

18 Maximum likelihood The likelihood is the joint probability of the data, as a function of the model parameters 𝑓 π‘₯ = 1 πœƒ 𝑒 βˆ’ π‘₯ πœƒ Exponential model: Likelihood: 𝐿 πœƒ = 𝑖=1 𝑛 1 πœƒ 𝑒 βˆ’ π‘₯ 𝑖 πœƒ x 𝐿(πœƒ) πœƒ

19 Maximum log likelihood
The log likelihood is maximized at the same point and is often easier to work with. 𝐿 πœƒ = 𝑖=1 𝑛 1 πœƒ 𝑒 βˆ’ π‘₯ 𝑖 πœƒ Likelihood: Log likelihood: log 𝐿 πœƒ =βˆ’ 𝑖=1 𝑛 log πœƒ + π‘₯ 𝑖 πœƒ log 𝐿(πœƒ) πœƒ

20 Maximum likelihood The likelihood is the joint probability of the data, as a function of the model parameters log 𝐿 πœƒ =βˆ’ 𝑖=1 𝑛 log πœƒ + π‘₯ 𝑖 πœƒ Log likelihood: 𝑑log 𝐿 πœƒ π‘‘πœƒ =βˆ’ 𝑖=1 𝑛 1 πœƒ βˆ’ π‘₯ 𝑖 πœƒ 2 βˆ’ 𝑛 πœƒ πœƒ 2 𝑖=1 𝑛 π‘₯ 𝑖 =0 β‡’ πœƒ = π‘₯ 𝑖 𝑛 =30.8 ms

21 Goodness-of-fit Goodness-of-fit is assessed across multiple measures
One intuitive approach – compare model to data Kolmogorov-Smirnov (KS) Plot 𝐹 π‘₯ 𝐹 π‘₯ 𝑓 π‘₯ 𝐹 π‘₯ ISI (ms)

22 Interval estimation Any point estimate for a parameter is almost certainly incorrect and hard to express confidence about Instead we want a statement about a reasonable range of parameter values and our confidence in that range. The likelihood provides an intuitive approach: Examine curvature of log 𝐿(πœƒ) at πœƒ ML log 𝐿(πœƒ) πœƒ ML log 𝐿(πœƒ)

23 Fisher Information Information contained in data about model parameter
Expected Fisher Information: Observed Fisher Information: 𝐼 πœƒ =βˆ’πΈ 𝑑 2 log 𝐿(πœƒ) 𝑑 πœƒ 2 𝐼 πœƒ =βˆ’ 𝑑 2 log 𝐿(πœƒ) 𝑑 πœƒ πœƒ ML Cramer-Rao bound for any unbiased estimator: ML estimators achieve this bound and have a normal distribution asymptotically var πœƒ β‰₯ 𝐼(πœƒ) βˆ’1 95% CI for πœƒ: πœƒ ML Β±2 𝐼(πœƒ) βˆ’ 1 2

24 Fisher Information For our exponential model of the retinal ISI data
𝑑log 𝐿 πœƒ π‘‘πœƒ =βˆ’ 𝑖=1 𝑛 1 πœƒ βˆ’ π‘₯ 𝑖 πœƒ 2 𝑑 2 log 𝐿 πœƒ 𝑑 πœƒ 2 = 𝑖=1 𝑛 1 πœƒ 2 βˆ’ 2 π‘₯ 𝑖 πœƒ 3 𝐼 πœƒ =βˆ’πΈ 𝑑 2 log 𝐿 πœƒ 𝑑 πœƒ 2 =βˆ’ 𝑛 πœƒ π‘›πœƒ πœƒ 3 = 𝑛 πœƒ 2 𝐼 πœƒ =βˆ’ 𝑑 2 log 𝐿 πœƒ 𝑑 πœƒ πœƒ ML =βˆ’ 𝑛 π‘₯ 𝑛 π‘₯ π‘₯ 3 = 𝑛 π‘₯ 2 var πœƒ β‰₯ πœƒ 2 𝑛 95% CI for πœƒ: π‘₯ Β±2 π‘₯ 𝑛 =[28.9, 32.8]

25 Information theory Claude Shannon – A mathematical theory of communication (1948) Formulated to quantify communication across a noisy channel Defines particular properties of probability models Goodness-of-fit often neglected Information theoretic statistics are rarely most powerful statistically These statistics can be useful when they have intuitive interpretations.

26 Information theory: entropy
For discrete probability models, entropy is one of many measures to quantify variability/dispersion Self information/surprisal of an outcome: βˆ’log 𝑝(π‘₯) Larger for less likely outcomes Information of independent events is additive Entropy is the expected surprisal: Maximized when each event is equally surprising Almost always expressed using log 2 (bits) 𝐻 𝑋 =βˆ’πΈ log 𝑝 𝑋 =βˆ’ π‘₯ 𝑝 π‘₯ log 𝑝(π‘₯)

27 Information theory: differential entropy
Not quite analogous measure of dispersion for continuous probability models Typically computed using log 2 (in bits) Depends on units of 𝑋 Can take on negative values Harder to interpret, but smaller values suggest probability mass confined to smaller volumes 𝐻 𝑋 =βˆ’πΈ log 𝑓 𝑋 =βˆ’ π‘₯ 𝑓 π‘₯ log 𝑓(π‘₯) 𝑑π‘₯

28 Differential entropy of our ISI model
𝐻 𝑋 =βˆ’ 0 ∞ 1 πœƒ 𝑒 βˆ’ π‘₯ πœƒ log 1 πœƒ 𝑒 βˆ’ π‘₯ πœƒ 𝑑π‘₯ = log πœƒ + 1 πœƒ 0 ∞ π‘₯ πœƒ 𝑒 βˆ’ π‘₯ πœƒ 𝑑π‘₯ = log πœƒ + 1 πœƒ 𝐸 𝑋 = log πœƒ +1 =4.43 nats=6.39 bits This does NOT mean that retinal spiking contains 6.4 bits of information. 𝐻 𝑋 = -3.6 bits if πœƒ is in seconds. Watch out for many nonsense statements related to information theory!

29 Information theory: KL divergence
Distance measure between two distributions Some properties: Asymmetric, 𝐷 KL (𝑋||π‘Œ)β‰₯0. Interpretation (discrete): If 𝑃 𝑋 𝑋 is correct, information gain using 𝑃 𝑋 𝑋 instead of incorrect 𝑃 π‘Œ 𝑋 . Many relations to other information measures. 𝐷 KL 𝑋||π‘Œ =βˆ’ 𝐸 𝑋 log 𝑃 π‘Œ 𝑋 𝑃 𝑋 𝑋 Discrete variables: Continuous variables: 𝐷 KL 𝑋||π‘Œ =βˆ’ π‘₯ 𝑝 𝑋 π‘₯ log 𝑝 π‘Œ π‘₯ 𝑝 𝑋 π‘₯ 𝐷 KL 𝑋||π‘Œ =βˆ’ π‘₯ 𝑓 𝑋 𝑋 log 𝑓 π‘Œ π‘₯ 𝑓 𝑋 π‘₯ 𝑑π‘₯

30 Joint distributions In most experiments, we measure multiple variables. If variables are random, their relationship to each other is described by their joint distribution. Most statistical analyses relate to the estimation of some form of joint or conditional distribution. Some similar and some new challenges: Discrete vs continuous variables can lead to additional issues that require measure theoretic approaches More variables require high dimensional analysis methods

31 Joint distributions Imagine for retinal data we find that light intensity (X) varied in time and influenced ISIs (Y). ISIs (ms) Light intensity (fractional change)

32 Joint distributions Since both variables are continuous, we can describe distribution of probability mass using a joint density. 𝑓(π‘₯,𝑦) 𝑓(π‘₯,𝑦)β‰₯0 π‘₯ 𝑦 𝑓 π‘₯,𝑦 𝑑𝑦 𝑑π‘₯ =1 𝑦 Joint CDF: 𝐹 π‘Ž,𝑏 = π‘₯<π‘Ž 𝑦<𝑏 𝑓 π‘₯,𝑦 𝑑𝑦 𝑑π‘₯ π‘₯

33 Marginalization The marginal distribution of a set of variables is computed by integrating out any additional variables over their full support 𝑓 𝑦 = π‘₯ 𝑓 π‘₯,𝑦 𝑑π‘₯ 𝑓 π‘₯ = 𝑦 𝑓 π‘₯,𝑦 𝑑𝑦

34 Conditional distributions
Conditional distributions characterize the distribution of one set of variables at specific values of other variables 𝑃 π‘Œ 𝑋=π‘₯ = 𝑃(𝑋=π‘₯,π‘Œ) 𝑃(𝑋=π‘₯)

35 Conditional distributions
𝑃 π‘Œ 𝑋 , written as a function of the variable 𝑋, expresses the conditional distribution P(Y|𝑋=π‘₯) at each value 𝑋=π‘₯. 𝑃 π‘Œ 𝑋 = 𝑃(𝑋,π‘Œ) 𝑃(𝑋) 𝑃 𝑋,π‘Œ =𝑃 π‘Œ 𝑋 𝑃(𝑋) 𝑦 π‘₯

36 Independence Variables are independent when knowledge of one can provide no information about the other. 𝑃 π‘Œ 𝑋 =𝑃(π‘Œ) 𝑃(𝑋,π‘Œ) 𝑃(𝑋) =𝑃(π‘Œ) 𝑃(𝑋,π‘Œ)=𝑃 𝑋 𝑃(π‘Œ) 𝑦 π‘₯

37 Linear models Linear models define a conditional distribution of a response variable as a function of one or more predictor variables. Typically, we assume a Gaussian noise model Note that this is linear with respect to the parameters, but 𝑋 and π‘Œ can be any nonlinear functions of measured variables. Estimate parameters by maximum likelihood. π‘Œ= 𝛽 0 + 𝛽 1 𝑋+πœ€ π‘Œ|𝑋~𝑁( 𝛽 0 + 𝛽 1 𝑋,𝜎) 𝛽 1 = π‘₯ 𝑖 βˆ’ π‘₯ 𝑦 𝑖 βˆ’ 𝑦 π‘₯ 𝑖 βˆ’ π‘₯ 2 𝛽 0 = 𝑦 βˆ’ 𝛽 1 π‘₯

38 Linear fit of light intensity – ISI relationship
𝑦 π‘₯

39 Linear fit of light intensity – log(ISI) relationship
π‘₯ 𝑦 log⁑(𝑦) π‘₯

40 Linear models Basic principles of linear modeling contain most basic statistical methods as special cases: z-tests, t-tests – linear model with binary predictor ANOVA – linear model with categorical predictors multiple regression – linear model with multiple predictors Quadratic /polynomial/spline/wavelet/etc. regression – linear model with predictors that are functions of select variables. Spectral estimation – linear model with sinusoid predictors Autocorrelation analysis/AR models – linear models with past signal values as predictors

41 More information theory
Conditional entropy – expected remaining entropy in Y once X is known 𝐻 π‘Œ|𝑋 =βˆ’πΈ log 𝑃(π‘Œ|𝑋) =βˆ’ π‘₯,𝑦 𝑝 π‘₯,𝑦 log 𝑝(π‘₯,𝑦) 𝑝(π‘₯) 𝐼 𝑋;π‘Œ =βˆ’πΈ log 𝑃 𝑋 𝑃 π‘Œ 𝑃 𝑋,π‘Œ =βˆ’ π‘₯,𝑦 𝑝 π‘₯,𝑦 log 𝑝 π‘₯ 𝑝 𝑦 𝑝 π‘₯,𝑦 =𝐻 π‘Œ βˆ’π» π‘Œ 𝑋 = 𝐷 𝐾𝐿 (𝑃(𝑋,π‘Œ)||𝑃 𝑋 𝑃(π‘Œ)) Mutual information – Expected change in entropy of Y by learning value of X. 𝐻 π‘Œ =3.15 nats. 𝐻 π‘Œ 𝑋 =2.97 nats. 𝐼 𝑋;π‘Œ =0.18 nats=0.27 bits. For retinal dataset:

42 Decoding problem Imagine in our retinal ISI data that we neglected to record the light intensity level through time for part of the experiment. Can we estimate it from the ISIs?

43 Decoding using the likelihood
Our linear model describes the conditional distribution of the log ISI given the light intensity: Without any probability model for the intensity we could estimate its value for each ISI using ML log⁑(π‘Œ)|𝑋~𝑁( 𝛽 0 + 𝛽 1 𝑋,𝜎) 𝐿 π‘₯ 𝑑 = 1 2πœ‹ 𝜎 2 𝑒 βˆ’ log 𝑦 𝑑 βˆ’ 𝛽 0 βˆ’ 𝛽 1 π‘₯ 𝑑 𝜎 2 𝑑log⁑(𝐿 π‘₯ 𝑑 ) 𝑑 π‘₯ 𝑑 =βˆ’ 2 log 𝑦 𝑑 βˆ’ 𝛽 0 βˆ’ 𝛽 1 π‘₯ 𝑑 2 𝜎 2 β‡’ π‘₯ 𝑑 = log 𝑦 𝑑 βˆ’ 𝛽 0 𝛽 1

44 Bayes’ Rule Consequence of conditional distribution definition
𝑃 𝑋,π‘Œ =𝑃 π‘Œ 𝑋 𝑃(𝑋) 𝑃 𝑋,π‘Œ =𝑃 𝑋 π‘Œ 𝑃(π‘Œ) 𝑃 𝑋|π‘Œ 𝑃 π‘Œ =𝑃 π‘Œ 𝑋 𝑃(𝑋) The posterior is proportional to the prior times the likelihood 𝑃 𝑋 π‘Œ = 𝑃 π‘Œ|𝑋 𝑃(𝑋) 𝑃(π‘Œ)

45 Bayesian Decoder Model 𝑋 as an independent normal process:
Prior: 𝑋~𝑁 πœ‡ π‘₯ , 𝜎 π‘₯ Likelihood: log⁑(π‘Œ)|𝑋~𝑁( 𝛽 0 + 𝛽 1 𝑋,𝜎) Posterior: 𝑓 𝑋 log π‘Œ βˆπ‘“ log π‘Œ 𝑋 𝑓 𝑋 . ∝ exp βˆ’ log 𝑦 βˆ’ 𝛽 0 βˆ’ 𝛽 1 π‘₯ 𝜎 2 βˆ’ π‘₯βˆ’ πœ‡ π‘₯ 𝜎 π‘₯ 2 𝑋|log π‘Œ ~𝑁 𝛽 1 𝜎 π‘₯ 2 log 𝑦 βˆ’ 𝛽 0 + 𝜎 2 πœ‡ π‘₯ 𝛽 1 2 𝜎 π‘₯ 2 + 𝜎 2 , 𝜎 π‘₯ 2 𝜎 2 𝛽 1 2 𝜎 π‘₯ 2 + 𝜎 2

46 Decoding ISI Data ISI (ms) Likelihood Decoder Intensity
Bayesian Decoder Intensity Spike index

47 𝑃 𝑦 1:𝑛 π‘₯ 1:π‘š =𝑃( 𝑦 1 | π‘₯ 1:π‘š )βˆ™βˆ™βˆ™π‘ƒ( 𝑦 𝑛 | π‘₯ 1:π‘š )
Latent variables Informal definition: any variable that is not directly observed, but which provides additional insight about the data is called latent. One formal definition: Latent variables are ones that makes the data conditionally independent: 𝑃( 𝑦 1:𝑛 )≠𝑃( 𝑦 1 )βˆ™βˆ™βˆ™π‘ƒ( 𝑦 𝑛 ) 𝑃 𝑦 1:𝑛 π‘₯ 1:π‘š =𝑃( 𝑦 1 | π‘₯ 1:π‘š )βˆ™βˆ™βˆ™π‘ƒ( 𝑦 𝑛 | π‘₯ 1:π‘š ) Latent variables are often used to model unobserved factors influencing data, low-dimensional representations of high-dimensional data, unobserved dynamics underlying data, …

48 The state-space paradigm
An unobserved (latent) state process (xt) undergoes stochastic dynamics that influence an observation process (yt) 𝑓 𝑦 π‘˜ | π‘₯ π‘˜ 𝑓 π‘₯ π‘˜ | π‘₯ π‘˜βˆ’1 Filtering problem: Estimate state, π‘₯ π‘˜ , given observations up to current time, 𝑦 0:π‘˜ . Smoothing problem: Estimate state, π‘₯ π‘˜ , given all observations, 𝑦 0:𝑇 . In our example, we may assume the light intensity varies smoothly in time: π‘₯ π‘˜ | π‘₯ π‘˜βˆ’1 ~𝑁(0,πœ–).

49 Filters and Smoothers Using basic probability theory, we can compute posterior of latent state, given all the observations up to current time or given all observations in the experiment 𝑓 π‘₯ π‘˜ | 𝑦 1:π‘˜ βˆπ‘“ 𝑦 π‘˜ | π‘₯ π‘˜ , 𝑦 1:π‘˜βˆ’1 𝑓( π‘₯ π‘˜ | 𝑦 1:π‘˜βˆ’1 ) 𝑓 π‘₯ π‘˜ 𝑦 1:π‘˜βˆ’1 = βˆ’βˆž ∞ 𝑓 π‘₯ π‘˜ | π‘₯ π‘˜βˆ’1 , 𝑦 1:π‘˜βˆ’1 𝑓( π‘₯ π‘˜βˆ’1 | 𝑦 1:π‘˜βˆ’1 )𝑑 π‘₯ π‘˜βˆ’1 𝑓 π‘₯ π‘˜ | 𝑦 1:π‘˜ βˆπ‘“ 𝑦 π‘˜ | π‘₯ π‘˜ βˆ’βˆž ∞ 𝑓 π‘₯ π‘˜ | π‘₯ π‘˜βˆ’1 𝑓( π‘₯ π‘˜βˆ’1 | 𝑦 1:π‘˜βˆ’1 )𝑑 π‘₯ π‘˜βˆ’1 𝑓 π‘₯ π‘˜ | 𝑦 1:𝐾 =𝑓 π‘₯ π‘˜ | 𝑦 1:π‘˜ βˆ’βˆž ∞ 𝑓 π‘₯ π‘˜+1 | 𝑦 1:𝐾 𝑓 π‘₯ π‘˜ | π‘₯ π‘˜βˆ’1 𝑓 π‘₯ π‘˜+1 | 𝑦 0:π‘˜ 𝑑 π‘₯ π‘˜+1

50 Latent variables ISI Data ISI (ms) Filter Estimate Intensity
Smoother Estimate Intensity Spike index

51 Take home points Most data analysis techniques are fundamentally concerned with expressing a (joint) probability model for the data. Nonparametric and parametric modeling methods are complementary and can provide robust and interpretable properties of the data. Maximum likelihood provides an often statistically optimal approach for estimating parameters of models. Information measures describe properties of a probability model, which can in turn provide insight into data. Latent variable methods use knowledge or assumptions about unobserved variables to improve understanding of observed data.


Download ppt "2016 METHODS IN COMPUTATIONAL NEUROSCIENCE COURSE"

Similar presentations


Ads by Google