False Discovery Rate Methods for Functional Neuroimaging Thomas Nichols Department of Biostatistics University of Michigan.

Slides:



Advertisements
Similar presentations
Inference on SPMs: Random Field Theory & Alternatives
Advertisements

Concepts of SPM data analysis Marieke Schölvinck.
Multiple testing and false discovery rate in feature selection
Statistical Modeling and Data Analysis Given a data set, first question a statistician ask is, “What is the statistical model to this data?” We then characterize.
FMRI Data Analysis: I. Basic Analyses and the General Linear Model
Multiple comparison correction
Topological Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course London, May 2014 Many thanks to Justin.
Inference on SPMs: Random Field Theory & Alternatives
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
07/01/15 MfD 2014 Xin You Tai & Misun Kim
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
Multiple comparison correction Methods & models for fMRI data analysis 18 March 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
Differentially expressed genes
Comparison of Parametric and Nonparametric Thresholding Methods for Small Group Analyses Thomas Nichols & Satoru Hayasaka Department of Biostatistics U.
Multiple comparison correction Methods & models for fMRI data analysis 29 October 2008 Klaas Enno Stephan Branco Weiss Laboratory (BWL) Institute for Empirical.
1 Overview of Hierarchical Modeling Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics Mixed Effects.
Significance Tests P-values and Q-values. Outline Statistical significance in multiple testing Statistical significance in multiple testing Empirical.
FMRI: Biological Basis and Experiment Design Lecture 26: Significance Review of GLM results Baseline trends Block designs; Fourier analysis (correlation)
Lorelei Howard and Nick Wright MfD 2008
FMRI – Week 9 – Analysis I Scott Huettel, Duke University FMRI Data Analysis: I. Basic Analyses and the General Linear Model FMRI Undergraduate Course.
False Discovery Rate (FDR) = proportion of false positive results out of all positive results (positive result = statistically significant result) Ladislav.
General Linear Model & Classical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM M/EEGCourse London, May.
Multiple Comparison Correction in SPMs Will Penny SPM short course, Zurich, Feb 2008 Will Penny SPM short course, Zurich, Feb 2008.
Multiple testing in high- throughput biology Petter Mostad.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005 David Carmichael MfD 2006 David Carmichael.
Basics of fMRI Inference Douglas N. Greve. Overview Inference False Positives and False Negatives Problem of Multiple Comparisons Bonferroni Correction.
Random field theory Rumana Chowdhury and Nagako Murase Methods for Dummies November 2010.
Differential Expression II Adding power by modeling all the genes Oct 06.
Computational Biology Jianfeng Feng Warwick University.
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
Multiple comparison correction Methods & models for fMRI data analysis October 2013 With many thanks for slides & images to: FIL Methods group & Tom Nichols.
Multiple comparisons in M/EEG analysis Gareth Barnes Wellcome Trust Centre for Neuroimaging University College London SPM M/EEG Course London, May 2013.
1 Inference on SPMs: Random Field Theory & Alternatives Thomas Nichols, Ph.D. Department of Statistics & Warwick Manufacturing Group University of Warwick.
Methods for Dummies Random Field Theory Annika Lübbert & Marian Schneider.
1 Data Modeling General Linear Model & Statistical Inference Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics
Thresholding and multiple comparisons
1 Nonparametric Thresholding Methods (FWE inference w/ SnPM) Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics University of Michigan.
**please note** Many slides in part 1 are corrupt and have lost images and/or text. Part 2 is fine. Unfortunately, the original is not available, so please.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005.
The False Discovery Rate A New Approach to the Multiple Comparisons Problem Thomas Nichols Department of Biostatistics University of Michigan.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
1 Identifying Robust Activation in fMRI Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics University of Michigan
Statistical Analysis An Introduction to MRI Physics and Analysis Michael Jay Schillaci, PhD Monday, April 7 th, 2007.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 6 –Multiple hypothesis testing Marshall University Genomics.
Multiple comparisons problem and solutions James M. Kilner
Topological Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course London, May 2015 With thanks to Justin.
Multiple comparison correction
1 Massively Univariate Inference for fMRI Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics University of Michigan
1 השוואות מרובות מדדי טעות, עוצמה, רווחי סמך סימולטניים ד"ר מרינה בוגומולוב מבוסס על ההרצאות של פרופ' יואב בנימיני ופרופ' מלכה גורפיין.
False Discovery Rate for Functional Neuroimaging Thomas Nichols Department of Biostatistics University of Michigan Christopher Genovese & Nicole Lazar.
Statistics Part II John VanMeter, Ph.D. Center for Functional and Molecular Imaging Georgetown University Medical Center.
Estimating the False Discovery Rate in Genome-wide Studies BMI/CS 576 Colin Dewey Fall 2008.
What a Cluster F… ailure!
Topological Inference
Nonparametric Inference with SnPM
Inference on SPMs: Random Field Theory & Alternatives
Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics
Methods for Dummies Random Field Theory
Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics
Topological Inference
Contrasts & Statistical Inference
Inference on SPMs: Random Field Theory & Alternatives
Inference on SPMs: Random Field Theory & Alternatives
Inference on SPMs: Random Field Theory & Alternatives
Statistical Parametric Mapping
The general linear model and Statistical Parametric Mapping
Statistical Challenges in “Big Data” Human Neuroimaging
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
Contrasts & Statistical Inference
Presentation transcript:

False Discovery Rate Methods for Functional Neuroimaging Thomas Nichols Department of Biostatistics University of Michigan

Outline Functional MRI A Multiple Comparison Solution: False Discovery Rate (FDR) FDR Properties FDR Example

fMRI Models & Multiple Comparisons Massively Univariate Modeling Fit model at each volume element or “voxel” Create statistic images of effect Which of 100,000 voxels are significant? =0.05  5,000 false positives! t > 0.5 t > 1.5 t > 2.5 t > 3.5 t > 4.5 t > 5.5 t > 6.5

Solutions for the Multiple Comparison Problem A MCP Solution Must Control False Positives How to measure multiple false positives? Familywise Error Rate (FWER) Chance of any false positives Controlled by Bonferroni & Random Field Methods False Discovery Rate (FDR) Proportion of false positives among rejected tests

obsFDR = V0R/(V1R+V0R) = V0R/NR False Discovery Rate Accept Reject Null True V0A V0R m0 Null False V1A V1R m1 NA NR V Observed FDR obsFDR = V0R/(V1R+V0R) = V0R/NR If NR = 0, obsFDR = 0 Only know NR, not how many are true or false Control is on the expected FDR FDR = E(obsFDR)

False Discovery Rate Illustration: Noise Signal Signal+Noise

Control of Per Comparison Rate at 10% 11.3% 12.5% 10.8% 11.5% 10.0% 10.7% 11.2% 10.2% 9.5% Percentage of Null Pixels that are False Positives Control of Familywise Error Rate at 10% FWE Occurrence of Familywise Error Control of False Discovery Rate at 10% 6.7% 10.4% 14.9% 9.3% 16.2% 13.8% 14.0% 10.5% 12.2% 8.7% Percentage of Activated Pixels that are False Positives

Benjamini & Hochberg Procedure Select desired limit q on FDR Order p-values, p(1)  p(2)  ...  p(V) Let r be largest i such that Reject all hypotheses corresponding to p(1), ... , p(r). JRSS-B (1995) 57:289-300 1 p(i)  i/V  q/c(V) p(i) p-value i/V  q/c(V) i/V 1

Benjamini & Hochberg Procedure c(V) = 1 Positive Regression Dependency on Subsets P(X1c1, X2c2, ..., Xkck | Xi=xi) is non-decreasing in xi Only required of test statistics for which null true Special cases include Independence Multivariate Normal with all positive correlations Same, but studentized with common std. err. c(V) = i=1,...,V 1/i  log(V)+0.5772 Arbitrary covariance structure Benjamini & Yekutieli (2001). Ann. Stat. 29:1165-1188

Other FDR Methods John Storey JRSS-B (2002) 64:479-498 pFDR “Positive FDR” FDR conditional on one or more rejections Critical threshold is fixed, not estimated pFDR and Emperical Bayes Asymptotically valid under “clumpy” dependence James Troendle JSPI (2000) 84:139-158 Normal theory FDR More powerful than BH FDR Requires numerical integration to obtain thresholds Exactly valid if whole correlation matrix known

Benjamini & Hochberg: Key Properties FDR is controlled E(obsFDR)  q m0/V Conservative, if large fraction of nulls false Adaptive Threshold depends on amount of signal More signal, More small p-values, More p(i) less than i/V  q/c(V)

Controlling FDR: Varying Signal Extent p = z = Signal Intensity 3.0 Signal Extent 1.0 Noise Smoothness 3.0 1

Controlling FDR: Varying Signal Extent p = z = Signal Intensity 3.0 Signal Extent 2.0 Noise Smoothness 3.0 2

Controlling FDR: Varying Signal Extent p = z = Signal Intensity 3.0 Signal Extent 3.0 Noise Smoothness 3.0 3

Controlling FDR: Varying Signal Extent p = 0.000252 z = 3.48 Signal Intensity 3.0 Signal Extent 5.0 Noise Smoothness 3.0 4

Controlling FDR: Varying Signal Extent p = 0.001628 z = 2.94 Signal Intensity 3.0 Signal Extent 9.5 Noise Smoothness 3.0 5

Controlling FDR: Varying Signal Extent p = 0.007157 z = 2.45 Signal Intensity 3.0 Signal Extent 16.5 Noise Smoothness 3.0 6

Controlling FDR: Varying Signal Extent p = 0.019274 z = 2.07 Signal Intensity 3.0 Signal Extent 25.0 Noise Smoothness 3.0 7

Controlling FDR: Benjamini & Hochberg Illustrating BH under dependence Extreme example of positive dependence 8 voxel image 1 32 voxel image (interpolated from 8 voxel image) p(i) p-value i/V  q/c(V) i/V 1

Controlling FDR: Varying Noise Smoothness p = 0.000132 z = 3.65 Signal Intensity 3.0 Signal Extent 5.0 Noise Smoothness 0.0 1

Controlling FDR: Varying Noise Smoothness p = 0.000169 z = 3.58 Signal Intensity 3.0 Signal Extent 5.0 Noise Smoothness 1.5 2

Controlling FDR: Varying Noise Smoothness p = 0.000167 z = 3.59 Signal Intensity 3.0 Signal Extent 5.0 Noise Smoothness 2.0 3

Controlling FDR: Varying Noise Smoothness p = 0.000252 z = 3.48 Signal Intensity 3.0 Signal Extent 5.0 Noise Smoothness 3.0 4

Controlling FDR: Varying Noise Smoothness p = 0.000253 z = 3.48 Signal Intensity 3.0 Signal Extent 5.0 Noise Smoothness 4.0 5

Controlling FDR: Varying Noise Smoothness p = 0.000271 z = 3.46 Signal Intensity 3.0 Signal Extent 5.0 Noise Smoothness 5.5 6

Controlling FDR: Varying Noise Smoothness p = 0.000274 z = 3.46 Signal Intensity 3.0 Signal Extent 5.0 Noise Smoothness 7.5 7

Benjamini & Hochberg: Properties Adaptive Larger the signal, the lower the threshold Larger the signal, the more false positives False positives constant as fraction of rejected tests Not such a problem with imaging’s sparse signals Smoothness OK Smoothing introduces positive correlations

Controlling FDR Under Dependence FDR under low df, smooth t images Validity PRDS only shown for studentization by common std. err. Sensitivity If valid, is control tight? Null hypothesis simulation of t images 3000, 323232 voxel images simulated df: 8, 18, 28 (Two groups of 5, 10 & 15) Smoothness: 0, 1.5, 3, 6, 12 FWHM (Gaussian, 0~5 ) Painful t simulations

Dependence Simulation Results Observed FDR For very smooth cases, rejects too infrequently Suggests conservativeness in ultrasmooth data OK for typical smoothnesses

Dependence Simulation FDR controlled under complete null, under various dependency Under strong dependency, probably too conservative

Positive Regression Dependency Does fMRI data exhibit total positive correlation? Initial Exploration 160 scan experiment Simple finger tapping paradigm No smoothing Linear model fit, residuals computed Voxels selected at random Only one negative correlation...

Positive Regression Dependency Negative correlation between ventricle and brain

Positive Regression Dependency More data needed Positive dependency assumption probably OK Users usually smooth data with nonnegative kernel Subtle negative dependencies swamped

Example Data fMRI Study of Working Memory ... D yes UBKDA Active fMRI Study of Working Memory 12 subjects, block design Marshuetz et al (2000) Item Recognition Active:View five letters, 2s pause, view probe letter, respond Baseline: View XXXXX, 2s pause, view Y or N, respond Random/Mixed Effects Modeling Model each subject, create contrast of interest One sample t test on contrast images yields pop. inf. ... N no XXXXX Baseline

FDR Example: Plot of FDR Inequality p(i)  ( i/V ) ( q/c(V) )

FDR Example FWER Perm. Thresh. = 7.67 58 voxels FDR Threshold = 3.83 3,073 voxels Threshold Indep/PosDep u = 3.83 Arb Cov u = 13.15 Result 3,073 voxels above Indep/PosDep u <0.0001 minimum FDR-corrected p-value

FDR: Conclusions False Discovery Rate Benjamini & Hochberg FDR Method A new false positive metric Benjamini & Hochberg FDR Method Straightforward solution to fMRI MCP Valid under dependency Just one way of controlling FDR New methods under development Limitations Arbitrary dependence result less sensitive Start Ill http://www.sph.umich.edu/~nichols/FDR Prop

FDR Software for SPM http://www.sph.umich.edu/~nichols/FDR