Statistics for fMRI.

Slides:



Advertisements
Similar presentations
On Comparing Classifiers : Pitfalls to Avoid and Recommended Approach
Advertisements

VBM Voxel-based morphometry
Mkael Symmonds, Bahador Bahrami
ANOVA & Pairwise Comparisons
Thursday, September 12, 2013 Effect Size, Power, and Exam Review.

Empirical Analysis Doing and interpreting empirical work.
07/01/15 MfD 2014 Xin You Tai & Misun Kim
Multiple comparison correction Methods & models for fMRI data analysis 18 March 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
Statistics for the Social Sciences
Circular analysis in systems neuroscience – with particular attention to cross-subject correlation mapping Nikolaus Kriegeskorte Laboratory of Brain and.
Comparison of Parametric and Nonparametric Thresholding Methods for Small Group Analyses Thomas Nichols & Satoru Hayasaka Department of Biostatistics U.
Statistics Micro Mini Threats to Your Experiment!
Multiple comparison correction Methods & models for fMRI data analysis 29 October 2008 Klaas Enno Stephan Branco Weiss Laboratory (BWL) Institute for Empirical.
Lecture 9: One Way ANOVA Between Subjects
Challenger Disaster click here for video. Fig. 1-1, p. 4.
On Comparing Classifiers: Pitfalls to Avoid and Recommended Approach Published by Steven L. Salzberg Presented by Prakash Tilwani MACS 598 April 25 th.
Issues with analysis and interpretation - Type I/ Type II errors & double dipping - Madeline Grade & Suz Prejawa Methods for Dummies 2013.
Multiple Comparison Correction in SPMs Will Penny SPM short course, Zurich, Feb 2008 Will Penny SPM short course, Zurich, Feb 2008.
Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 1 False discovery rate: setting the probability.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005 David Carmichael MfD 2006 David Carmichael.
Chapter 9 Comparing More than Two Means. Review of Simulation-Based Tests  One proportion:  We created a null distribution by flipping a coin, rolling.
Introductory Statistics for Laboratorians dealing with High Throughput Data sets Centers for Disease Control.
PSY2004 Research Methods PSY2005 Applied Research Methods Week Eleven Stephen Nunn.
Step 3 of the Data Analysis Plan Confirm what the data reveal: Inferential statistics All this information is in Chapters 11 & 12 of text.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
Methods for Dummies Random Field Theory Annika Lübbert & Marian Schneider.
PSY2004 Research Methods PSY2005 Applied Research Methods Week Five.
Introduction to Microarrays Dr. Özlem İLK & İbrahim ERKAN 2011, Ankara.
Multiple Testing Matthew Kowgier. Multiple Testing In statistics, the multiple comparisons/testing problem occurs when one considers a set of statistical.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005.
Stats Lunch: Day 3 The Basis of Hypothesis Testing w/ Parametric Statistics.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 6 –Multiple hypothesis testing Marshall University Genomics.
King Faisal University جامعة الملك فيصل Deanship of E-Learning and Distance Education عمادة التعلم الإلكتروني والتعليم عن بعد [ ] 1 جامعة الملك فيصل عمادة.
Kriegeskorte N, Simmons WK, Bellgowan PSF, Baker CI. Circular analysis in systems neuroscience – the dangers of double dipping slides by Nikolaus Kriegeskorte,
Introduction to Power and Effect Size  More to life than statistical significance  Reporting effect size  Assessing power.
Multivariate Pattern Analysis of fMRI data. Goal of this lecture Introduction of basic concepts & a few commonly used approaches to multivariate pattern.
What a Cluster F… ailure!
Chi-Square Test A fundamental problem is genetics is determining whether the experimentally determined data fits the results expected from theory (i.e.
Analysis of variance ANOVA.
Psychology Unit Research Methods - Statistics
Statistical Core Didactic
Statistics for the Social Sciences
Inference and Tests of Hypotheses
Chapter 21 More About Tests.
Sample Power.
Central Limit Theorem, z-tests, & t-tests
Issues with analysis and interpretation
Inferential Statistics
Object Oriented Programming (OOP) LAB # 8
Methods for Dummies Random Field Theory
Making Data-Based Decisions
Sequence comparison: Multiple testing correction
Recap Questions What is interactionism?
When will we see people of negative height?
Statistical Tests P Values.
Contrasts & Statistical Inference
Two Categorical Variables: The Chi-Square Test
One Way ANOVAs One Way ANOVAs
This is a hot topic!! Science uptake figures are 'science fiction', says report Labour has been accused of fiddling the figures on the number of students.
Chapter 12 Power Analysis.
Sequence comparison: Multiple testing correction
Inferential Statistics
Psych 231: Research Methods in Psychology
Contrasts & Statistical Inference
Psych 231: Research Methods in Psychology
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
Calculating t X -  t = s x X1 – X2 t = s x1 – x2 s d One sample test
Contrasts & Statistical Inference
Presentation transcript:

Statistics for fMRI

Statistics and science Quick poll: When was statistics (p-values for rejecting a hypothesis, t-tests etc.) invented? A: Early 1700s B: Early 1800s C: Early 1900s

Statistics and science Answer: Early 1900s t-test: 1908 (Gosset / “Student” ) General principles: 1925 “Statistical Methods for Research Workers” Ronald A. Fisher

Statistics and science A puzzle: Lots of great science was done before statistics came along! t-test: 1908 (Gosset / “Student” ) General principles: 1925 “Statistical Methods for Research Workers” Ronald A. Fisher Some discoveries which (today) seem unthinkable without statistics: Development of vaccines (Jenner, 1796) Genetics and inheritance (Mendel, 1866) Almost all of the rest of science!

A metaphor: doing stats right is like putting on your seatbelt What really matters in science is finding a true underlying regularity in nature Gold standard test: independent replication Neither necessary nor sufficient, but helpful: doing the stats right What really matters in driving is getting home safely Neither necessary nor sufficient, but helpful: wearing your seatbelt For a long time, cars didn’t even have seatbelts. But people still got home safely (usually)

Multiple comparisons test There are lots of voxels in the brain (~30,000) You can test each one separately to see whether it is active If you run lots of tests, some of them will give positive results even if nothing is there

http://xkcd.com/882/

http://xkcd.com/882/

http://xkcd.com/882/

http://xkcd.com/882/

Mouse-float-over title text: http://www.explainxkcd.com/wiki/index.php/882:_Significant Mouse-float-over title text: So, uh, we did the green study again and got no link. It was probably a-- "RESEARCH CONFLICTED ON GREEN JELLY BEAN/ACNE LINK; MORE STUDY RECOMMENDED!"

http://xkcd.com/552/ Correlation does not imply causation, but it does waggle its eyebrows suggestively and gesture furtively while mouthing "look over there."

What does “Result X is statistically significant (p<0 What does “Result X is statistically significant (p<0.05)” actually mean? Take a vote! A: Given the data that we found, there is only a 5% chance that Result X is false B: If Result X were false, then there would only be a 5% chance of us finding the data that we did find C: If some other people now go off and look for Result X, there is only a 5% chance that they will fail to find it D: All of the above answers all mean basically the same thing

Base rates and false positives Sensitivity: probability of detecting a real effect Specificity: prob. of not getting a false positive Jacob Cohen “The Earth is round (p<0.05)”

What does “Result X is statistically significant (p<0 What does “Result X is statistically significant (p<0.05)” actually mean? A: Given the data that we found, there is only a 5% chance that Result X is false p( Hypothesis | Data ) B: If Result X were false, then there would only be a 5% chance of us finding the data that we did find p( Data | Hypothesis ) C: If some other people now go off and look for Result X, there is only a 5% chance that they will fail to find it D: All of the above answers all mean basically the same thing

Example: testing for a rare disease Suppose 0.1% of population has bubonic plague Suppose test for plague has 99% specificity (false positive rate) Test 1000 people: 1 person has plague 10 people test positive 9 of those are false positives!

What does “probability” actually mean? True or false? There is a single widely-agreed upon answer to the question of what exactly it means to say “The probability that it will rain tomorrow is 0.1” A1: True, there is an agreed upon answer A2: False, lots of sensible people disagree on this

Different ways of correcting for multiple comparisons Bonferroni correction Family-Wise Error False Discovery Rate

The infamous “dead salmon” study Bennett et al. (2009) If you scan a dead salmon and don’t do any multiple comparison correction, then you get some active voxels This is not exactly news, but it is at least memorable Like saying “Big discovery in traffic safety: remember to wear your seatbelt!”

Different ways of correcting for multiple comparisons Raw signal intensity Uncorrected results Family-Wise Error correction False Discovery Rate correction

Voodoo correlations – Vul et al (2009) ROI analysis Correlation with behaviour within that ROI Problem: It is tempting (and easy) to select the ROI, and to look for correlations within the ROI, using the same data: non-independent

An example highlighted in the voodoo correlations paper Eisenberger, Naomi I., Matthew D. Lieberman, and Kipling D. Williams. "Does rejection hurt? An fMRI study of social exclusion." Science 302.5643 (2003): 290-292.

Problem: correlations become inflated Vul, E. et al. (2009) "Puzzlingly high correlations in fMRI studies of emotion, personality, and social cognition." Perspectives on Psychological Science, 4(3): 274-290. Similar argument made simultaneously in: Kriegeskorte, N. et al. (2009). Circular analysis in systems neuroscience: the dangers of double dipping. Nature Neuroscience, 12(5), 535-540.

The solution: Cross-validation The solution: Cross-validation. Splitting up your data into independent parts Training set and testing set Independent parts of the data Just like in using classifiers

This Thursday Computer Lab: Meliora Rm 178. Registered students only (sorry) Explore hands-on the preprocessing steps involved in fMRI, using SPM: Motion correction Spatial normalisation to a template Smoothing You are welcome to bring your own laptop, but you’ll need to pre-load the Haxby data, Matlab and SPM8