Bayesian Inference Chris Mathys Wellcome Trust Centre for Neuroimaging UCL SPM Course London, May 12, 2014 Thanks to Jean Daunizeau and Jérémie Mattout.

Slides:



Advertisements
Similar presentations
The General Linear Model (GLM)
Advertisements

J. Daunizeau Institute of Empirical Research in Economics, Zurich, Switzerland Brain and Spine Institute, Paris, France Bayesian inference.
Bayesian inference Jean Daunizeau Wellcome Trust Centre for Neuroimaging 16 / 05 / 2008.
Bayesian Inference Lee Harrison York Neuroimaging Centre 14 / 05 / 2010.
Tests of Hypotheses Based on a Single Sample
Wellcome Trust Centre for Neuroimaging
SPM Software & Resources Wellcome Trust Centre for Neuroimaging University College London SPM Course London, May 2011.
Bayesian models for fMRI data
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
A Brief Introduction to Bayesian Inference Robert Van Dine 1.
Lectures on Stochastic System Analysis and Bayesian Updating June 29-July James L. Beck, California Institute of Technology Jianye Ching, National.
MEG/EEG Inverse problem and solutions In a Bayesian Framework EEG/MEG SPM course, Bruxelles, 2011 Jérémie Mattout Lyon Neuroscience Research Centre ? ?
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
Bayes for beginners Methods for dummies 27 February 2013 Claire Berna
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Introduction  Bayesian methods are becoming very important in the cognitive sciences  Bayesian statistics is a framework for doing inference, in a principled.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
I The meaning of chance Axiomatization. E Plurbus Unum.
Inference about a Mean Part II
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
A Practical Course in Graphical Bayesian Modeling; Class 1 Eric-Jan Wagenmakers.
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 9. Hypothesis Testing I: The Six Steps of Statistical Inference.
Hypothesis Testing. Distribution of Estimator To see the impact of the sample on estimates, try different samples Plot histogram of answers –Is it “normal”
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Statistical Decision Theory
Model Inference and Averaging
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Methods for Dummies 2009 Bayes for Beginners Georgina Torbet & Raphael Kaplan.
Chapter 9 Power. Decisions A null hypothesis significance test tells us the probability of obtaining our results when the null hypothesis is true p(Results|H.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Bayesian Methods I: Parameter Estimation “A statistician is a person who draws a mathematically precise line from an unwarranted assumption to a foregone.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
Therapeutic Equivalence & Active Control Clinical Trials Richard Simon, D.Sc. Chief, Biometric Research Branch National Cancer Institute.
Week 71 Hypothesis Testing Suppose that we want to assess the evidence in the observed data, concerning the hypothesis. There are two approaches to assessing.
Bayesian statistics Probabilities for everything.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK UZH – Foundations of Human Social Behaviour, Zurich, Switzerland Dynamic Causal Modelling:
Not in FPP Bayesian Statistics. The Frequentist paradigm Defines probability as a long-run frequency independent, identical trials Looks at parameters.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Reasoning Under Uncertainty. 2 Objectives Learn the meaning of uncertainty and explore some theories designed to deal with it Find out what types of errors.
Simple examples of the Bayesian approach For proportions and means.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
MACHINE LEARNING 3. Supervised Learning. Learning a Class from Examples Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
Mixture Models with Adaptive Spatial Priors Will Penny Karl Friston Acknowledgments: Stefan Kiebel and John Ashburner The Wellcome Department of Imaging.
Bayes for Beginners Anne-Catherine Huys M. Berk Mirza Methods for Dummies 20 th January 2016.
Contact Info: Improving Decision Making: The use of simple heuristics Dr. Guillermo Campitelli Cognition Research Group Edith.
Model Comparison.
J. Daunizeau ICM, Paris, France TNU, Zurich, Switzerland
Group Analyses Guillaume Flandin SPM Course London, October 2016
Unit 3 Hypothesis.
Model Inference and Averaging
Wellcome Trust Centre for Neuroimaging
PSY 626: Bayesian Statistics for Psychological Science
Bayes for Beginners Luca Chech and Jolanda Malamud
Bayesian inference J. Daunizeau
Wellcome Centre for Neuroimaging, UCL, UK.
Mixture Models with Adaptive Spatial Priors
CS639: Data Management for Data Science
Presentation transcript:

Bayesian Inference Chris Mathys Wellcome Trust Centre for Neuroimaging UCL SPM Course London, May 12, 2014 Thanks to Jean Daunizeau and Jérémie Mattout for previous versions of this talk

A spectacular piece of information 2May 12, 2014

A spectacular piece of information Messerli, F. H. (2012). Chocolate Consumption, Cognitive Function, and Nobel Laureates. New England Journal of Medicine, 367(16), 1562– May 12, 2014

So will I win the Nobel prize if I eat lots of chocolate? 4May 12, 2014

«Bayesian» = logical and logical = probabilistic «The actual science of logic is conversant at present only with things either certain, impossible, or entirely doubtful, none of which (fortunately) we have to reason on. Therefore the true logic for this world is the calculus of probabilities, which takes account of the magnitude of the probability which is, or ought to be, in a reasonable man's mind.» — James Clerk Maxwell, May 12, 2014

But in what sense is probabilistic reasoning (i.e., reasoning about uncertain quantities according to the rules of probability theory) «logical»? R. T. Cox showed in 1946 that the rules of probability theory can be derived from three basic desiderata: 1.Representation of degrees of plausibility by real numbers 2.Qualitative correspondence with common sense (in a well-defined sense) 3.Consistency «Bayesian» = logical and logical = probabilistic 6May 12, 2014

The rules of probability 7May 12, 2014

Conditional probabilities 8May 12, 2014

The chocolate example 9May 12, 2014 prior posterior likelihood evidence model

forward problem likelihood inverse problem posterior distribution Inference in SPM 10May 12, 2014

Likelihood: Prior: Bayes’ theorem: Inference in SPM 11May 12, 2014

A simple example of Bayesian inference (adapted from Jaynes (1976)) Assuming prices are comparable, from which manufacturer would you buy? A: B: Two manufacturers, A and B, deliver the same kind of components that turn out to have the following lifetimes (in hours): May 12,

A simple example of Bayesian inference How do we compare such samples? May 12,

What next? A simple example of Bayesian inference May 12,

A simple example of Bayesian inference The procedure in brief: Determine your question of interest («What is the probability that...?») Specify your model (likelihood and prior) Calculate the full posterior using Bayes’ theorem [Pass to the uninformative limit in the parameters of your prior] Integrate out any nuisance parameters Ask your question of interest of the posterior All you need is the rules of probability theory. (Ok, sometimes you’ll encounter a nasty integral – but that’s a technical difficulty, not a conceptual one). May 12,

A simple example of Bayesian inference The question: What is the probability that the components from manufacturer B have a longer lifetime than those from manufacturer A? More specifically: given how much more expensive they are, how much longer do I require the components from B to live. Example of a decision rule: if the components from B live 3 hours longer than those from A with a probability of at least 80%, I will choose those from B. May 12,

A simple example of Bayesian inference May 12,

A simple example of Bayesian inference May 12,

A simple example of Bayesian inference May 12,

A simple example of Bayesian inference May 12,

A simple example of Bayesian inference May 12,

Bayesian inference The procedure in brief: Determine your question of interest («What is the probability that...?») Specify your model (likelihood and prior) Calculate the full posterior using Bayes’ theorem [Pass to the uninformative limit in the parameters of your prior] Integrate out any nuisance parameters Ask your question of interest of the posterior All you need is the rules of probability theory. May 12,

Frequentist (or: orthodox, classical) versus Bayesian inference: hypothesis testing if then reject H 0 estimate parameters (obtain test stat.) define the null, e.g.: apply decision rule, i.e.: Classical 23May 12, 2014 ifthen accept H 0 invert model (obtain posterior pdf) define the null, e.g.: apply decision rule, i.e.: Bayesian

Principle of parsimony: «plurality should not be assumed without necessity» Automatically enforced by Bayesian model comparison y=f(x) x Model comparison: general principles model evidence p(y|m) space of all data sets Model evidence: “Occam’s razor” : 24May 12, 2014

Model comparison: negative variational free energy F 25May 12, 2014 Jensen’s inequality sum rule product rule Kullback-Leibler divergence a lower bound on the log-model evidence

Model comparison: F in relation to Bayes factors, AIC, BIC 26May 12, 2014 Posterior odds Prior odds Bayes factor Number of parameters Number of data points

A note on informative priors 27May 12, 2014 Any model consists of two parts: likelihood and prior. The choice of likelihood requires as much justification as the choice of prior because it is just as «subjective» as that of the prior. The data never speak for themselves. They only acquire meaning when seen through the lens of a model. However, this does not mean that all is subjective because models differ in their validity. In this light, the widespread concern that informative priors might bias results (while the form of the likelihood is taken as a matter of course requiring no justification) is misplaced. Informative priors are an important tool and their use can be justified by establishing the validity (face, construct, and predictive) of the resulting model as well as by model comparison.

Applications of Bayesian inference 28May 12, 2014

realignmentsmoothing normalisation general linear model template Gaussian field theory p <0.05 statisticalinference segmentation and normalisation segmentation and normalisation dynamic causal modelling dynamic causal modelling posterior probability maps (PPMs) posterior probability maps (PPMs) multivariate decoding multivariate decoding 29May 12, 2014

grey matterCSFwhite matter … … class variances class means i th voxel value i th voxel label class frequencies Segmentation (mixture of Gaussians-model) 30May 12, 2014

PPM: regions best explained by short-term memory model PPM: regions best explained by long-term memory model fMRI time series GLM coeff prior variance of GLM coeff prior variance of data noise AR coeff (correlated noise) short-term memory design matrix (X) long-term memory design matrix (X) fMRI time series analysis 31May 12, 2014

m2m2 m1m1 m3m3 m4m4 V1 V5 stim PPC attention V1 V5 stim PPC attention V1 V5 stim PPC attention V1 V5 stim PPC attention m1m1 m2m2 m3m3 m4m V1 V5 stim PPC attention estimated effective synaptic strengths for best model (m 4 ) models marginal likelihood Dynamic causal modeling (DCM) 32May 12, 2014

m1m1 m2m2 differences in log- model evidences subjects Fixed effect Random effect Assume all subjects correspond to the same model Assume different subjects might correspond to different models Model comparison for group studies 33May 12, 2014

Thanks 34May 12, 2014