Bayes for Beginners Reverend Thomas Bayes (1702-61) Velia Cardin Marta Garrido.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Bayesian models for fMRI data
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Comments on Hierarchical models, and the need for Bayes Peter Green, University of Bristol, UK IWSM, Chania, July 2002.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
A Brief Introduction to Bayesian Inference Robert Van Dine 1.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Bayesian models for fMRI data
Bayesian models for fMRI data Methods & models for fMRI data analysis 06 May 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
Chapter Seventeen HYPOTHESIS TESTING
A/Prof Geraint Lewis A/Prof Peter Tuthill
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
Applied Bayesian Analysis for the Social Sciences Philip Pendergast Computing and Research Services Department of Sociology
Estimating the Transfer Function from Neuronal Activity to BOLD Maria Joao Rosa SPM Homecoming 2008 Wellcome Trust Centre for Neuroimaging.
Machine Learning Saarland University, SS 2007 Holger Bast [with input from Ingmar Weber] Max-Planck-Institut für Informatik Saarbrücken, Germany Lecture.
Bayes for Beginners Methods for Dummies FIL, UCL, Caroline Catmur, Psychology Department, UCL & Robert Adam, ICN, UCL 31 st October 2007.
Statistical Decision Theory
Bayes for Beginners Presenters: Shuman ji & Nick Todd.
METHODSDUMMIES BAYES FOR BEGINNERS. Any given Monday at pm “I’m sure this makes sense, but you lost me about here…”
Methods for Dummies 2009 Bayes for Beginners Georgina Torbet & Raphael Kaplan.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Significance Tests: THE BASICS Could it happen by chance alone?
Hypothesis Testing Hypothesis Testing Topic 11. Hypothesis Testing Another way of looking at statistical inference in which we want to ask a question.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
Bayesian Inference and Posterior Probability Maps Guillaume Flandin Wellcome Department of Imaging Neuroscience, University College London, UK SPM Course,
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Bayesian statistics Probabilities for everything.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Introduction to Inferential Statistics Statistical analyses are initially divided into: Descriptive Statistics or Inferential Statistics. Descriptive Statistics.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
DCM – the theory. Bayseian inference DCM examples Choosing the best model Group analysis.
Not in FPP Bayesian Statistics. The Frequentist paradigm Defines probability as a long-run frequency independent, identical trials Looks at parameters.
Bayesian Inference, Review 4/25/12 Frequentist inference Bayesian inference Review The Bayesian Heresy (pdf)pdf Professor Kari Lock Morgan Duke University.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Populations III: evidence, uncertainty, and decisions Bio 415/615.
Bayesian models for fMRI data Methods & models for fMRI data analysis November 2011 With many thanks for slides & images to: FIL Methods group, particularly.
Simple examples of the Bayesian approach For proportions and means.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Logic and Vocabulary of Hypothesis Tests Chapter 13.
1 URBDP 591 A Lecture 12: Statistical Inference Objectives Sampling Distribution Principles of Hypothesis Testing Statistical Significance.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
- 1 - Outline Introduction to the Bayesian theory –Bayesian Probability –Bayes’ Rule –Bayesian Inference –Historical Note Coin trials example Bayes rule.
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
Uncertainty and confidence Although the sample mean,, is a unique number for any particular sample, if you pick a different sample you will probably get.
Tests of Significance We use test to determine whether a “prediction” is “true” or “false”. More precisely, a test of significance gets at the question.
Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Zurich, February 2008 Bayesian Inference.
Bayes for Beginners Anne-Catherine Huys M. Berk Mirza Methods for Dummies 20 th January 2016.
Bayesian Inference in SPM2 Will Penny K. Friston, J. Ashburner, J.-B. Poline, R. Henson, S. Kiebel, D. Glaser Wellcome Department of Imaging Neuroscience,
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Bayes Net Learning: Bayesian Approaches
Special Topics In Scientific Computing
Bayes for Beginners Methods for Dummies FIL, UCL,
SPM2: Modelling and Inference
Wellcome Trust Centre for Neuroimaging
LECTURE 07: BAYESIAN ESTIMATION
Bayes for Beginners Luca Chech and Jolanda Malamud
Bayesian Inference in SPM2
Wellcome Centre for Neuroimaging, UCL, UK.
Mixture Models with Adaptive Spatial Priors
CS639: Data Management for Data Science
Mathematical Foundations of BME Reza Shadmehr
Presentation transcript:

Bayes for Beginners Reverend Thomas Bayes ( ) Velia Cardin Marta Garrido

“In addition to WLS estimators and classical inference, SPM2 also supports Bayesian estimation and inference. In this instance the statistical parametric maps become posterior probability maps Posterior Probability Maps (PPMs), where the posterior probability is a probability of an effect given the data. There is no multiple comparison problem in Bayesian inference and the posterior probabilities do not require adjustment.”BayesianPPM

Overview Velia Bayes vs Frequentist approach An example Bayes theorem Marta Bayesian Inference Posterior Probability Distribution Bayes in SPM Summary

vs Bayesian statistics Frequentist Statistics

With a frequentist approach…. The frequentist conclusion is restricted to the data at hand, it doesn’t take into account previous, valuable information. In general, we want to relate an event (E) to a hypothesis (H) …and the probability of E given H If the p-value is sufficiently small, you reject the null hypothesis, but… it doesn’t say anything about the probability of H i. We obtain a p-value that is the probability, given a true H 0, for the outcome to be more or equally extreme as the observed outcome.

Conclusions depends on previous evidence. Bayesian approach is not data analysis per se, it brings different types of evidence to answer the questions of importance. In general, we want to relate an event (E) to a hypothesis (H) and the probability of E given H Given a prior state of knowledge or belief, it tells how to update beliefs based upon observations (current data). The probability of a H being true is determined. With a Bayesian approach… You can compare the probabilities of different H for a same E A probability distribution of the parameter or hypothesis is obtained

Macaulay Culkin Busted for Drugs! Our observations…. DREW BARRYMORE REVEALS ALCOHOL AND DRUG PROBLEMS STARTED AGED EIGHT Feldman, arrested and charged with heroin possession Corey Haim in a spiral of prescription drug abuse! Dana Plato died of a drug overdose at age 34 Todd Bridges on suspicion of shooting and stabbing alleged drug dealer in a crack house....

We took a random sample of 40 people, 10 of them were young stars, being 3 of them addicted to drugs. From the other 30, just one. Our hypothesis is: “Young actors have more probability of becoming drug-addicts” Drug-addicted young actors control YA+ YA- D+D-

With a frequentist approach we will test: H i :’Conditions A and B have different effects’ Young actors have a different probability of becoming drug addicts than the rest of the people H 0 :’There is no difference in the effect of conditions A and B’ This is not what we want to know!!! …and we have strong believes that young actors have more probability of becoming drug addicts!!! The statistical test of choice is  2 and Yates’ correction:  2 = 3.33 p=0.07 We can’t reject the null hypothesis, and the information the p is giving us is basically that if we “do this experiment” many times, 7% of the times we will obtain this result if there is no difference between both conditions.

We want to know if 1 (0.025)29 (0.725) 10 (0.25) 30 (0.75) 3 (0.075)7 (0.175) 4 (0.1)36 (0.9) 40 (1) p(D+  YA+) p (D+  YA+) = p (D+ and YA+) / p (YA+) p (D+  YA+)  p (YA+  D+) Reformulating This is Bayes’ Theorem !!! YA+ YA- D-D+ total p (YA+  D+) = p (D+ and YA+) / p (D+) p (D+  YA+) = / 0.25 = 0.3 p (YA+  D+) = / 0.1 =  0.75 p (D+ and YA+) = p (YA+  D+) * p (D+) p (D+  YA-) = p (D+ and YA-) / p (YA-) p (D+  YA+) > p (D+  YA-) p (D+  YA-) = / 0.75 = p (D+  YA+) > p (D+  YA-) 0.3 > With a Bayesian approach… p (D+  YA-) p (YA+  D+) p (YA+  D+) * p (D+) p (YA+) p (D+  YA+) Substituting p (D+ and YA+) on p (D+  YA+) = p (D+ and YA+) / p (YA+)

It relates the conditional density of a parameter (posterior probability) with its unconditional density (prior, since depends on information present before the experiment). The likelihood is the probability of the data given the parameter and represents the data now available. Bayes’ Theorem for a given parameter  p (  data) = p (data  ) p (  ) / p (data) 1/P (data) is basically a normalizing constant Posterior  likelihood x prior The prior is the probability of the parameter and represents what was thought before seeing the data. The posterior represents what is thought given both prior information and the data just seen.

In fMRI…. Classical –‘What is the likelihood of getting these data given no activation occurred?’ Bayesian (SPM2) –‘What is the chance of getting these parameters, given these data?

What you know about the model after the data arrive,, is what you knew before,, and what the data told you,. In order to make probability statements about  given y we begin with a model Bayesian Inference joint prob. distribution where discrete case continuous case or likelihood prior posterior

Likelihood:p(y|  ) =  (M d, d -1 ) Prior:p(  ) =  (M p, p -1 ) Posterior: p(  y) ∝ p(y|  )*  p(  ) =  (M post, post -1 ) MpMp p -1 M post post -1 d -1 MdMd post = d + p M post = d M d + p M p  post Posterior Probability Distribution precision = 1/  2

The effects of different precisions p = d p < d p > d p ≈ 0

Multivariate Distributions

spatial normalization segmentation and Bayesian inference in… Posterior Probability Maps (PPM) Dynamic Causal Modelling (DCM) SPM uses priors for estimation in… Bayes in SPM

Shrinkage Priors Small, variable effect Large, variable effect Small, consistent effect Large, consistent effect

Thresholding p(  | y) = 0.95 

Summary In Bayesian estimation we… 1.…start with the formulation of a model that we hope is adequate to describe the situation of interest. 2.…observe the data and when the information available changes it is necessary to update the degrees of belief (probability). 3.…evaluate the fit of the model. If necessary we compute predictive distributions for future observations. priors over the parameters posterior distributions new priors over the parameters Prejudices or scientific judgment? The selection of a prior is subjective and arbitrary. It is reasonable to draw conclusions in the light of some reason. Bayesian methods use probability models for quantifying uncertainty in inferences based on statistical data analysis.

(Bayes’ original essay!!!) A. Gelman, J.B. Carlin, H.S. Stern and D.B. Rubin, 2 nd ed. Bayesian Data Analysis. Chapman & Hall/CRC. Mackay D. Information Theory, Inference and Learning Algorithms. Chapter 37: Bayesian inference and sampling theory. Cambridge University Press, Berry D, Stangl D. Bayesian Methods in Health-Realated Research. In: Bayesian Biostatistics. Berry D and Stangl K (eds). Marcel Dekker Inc, Friston KJ, Penny W, Phillips C, Kiebel S, Hinton G, Ashburner J. Classical and Bayesian inference in neuroimaging: theory. Neuroimage Jun;16(2): References

Bayes for Beginners Reverend Thomas Bayes ( ) …good-bayes!!! “We don’t see what we don’t seek.” E. M. Forster