Methods for Dummies Random Field Theory Annika Lübbert & Marian Schneider.

Slides:



Advertisements
Similar presentations
Mkael Symmonds, Bahador Bahrami
Advertisements

Random Field Theory Methods for Dummies 2009 Lea Firmin and Anna Jafarpour.
Concepts of SPM data analysis Marieke Schölvinck.
Multiple comparison correction
Statistical Inference and Random Field Theory Will Penny SPM short course, London, May 2003 Will Penny SPM short course, London, May 2003 M.Brett et al.
Multiple comparison correction
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research University of Zurich With many thanks for slides & images to: FIL Methods.
Topological Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course London, May 2014 Many thanks to Justin.
Classical inference and design efficiency Zurich SPM Course 2014
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
07/01/15 MfD 2014 Xin You Tai & Misun Kim
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
Multiple comparison correction Methods & models for fMRI data analysis 18 March 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
Comparison of Parametric and Nonparametric Thresholding Methods for Small Group Analyses Thomas Nichols & Satoru Hayasaka Department of Biostatistics U.
Multiple comparison correction Methods & models for fMRI data analysis 29 October 2008 Klaas Enno Stephan Branco Weiss Laboratory (BWL) Institute for Empirical.
False Discovery Rate Methods for Functional Neuroimaging Thomas Nichols Department of Biostatistics University of Michigan.
Lorelei Howard and Nick Wright MfD 2008
Giles Story Philipp Schwartenbeck
Voxel Based Morphometry
General Linear Model & Classical Inference
General Linear Model & Classical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM M/EEGCourse London, May.
Multiple Comparison Correction in SPMs Will Penny SPM short course, Zurich, Feb 2008 Will Penny SPM short course, Zurich, Feb 2008.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005 David Carmichael MfD 2006 David Carmichael.
Basics of fMRI Inference Douglas N. Greve. Overview Inference False Positives and False Negatives Problem of Multiple Comparisons Bonferroni Correction.
With a focus on task-based analysis and SPM12
Random field theory Rumana Chowdhury and Nagako Murase Methods for Dummies November 2010.
Computational Biology Jianfeng Feng Warwick University.
General Linear Model & Classical Inference London, SPM-M/EEG course May 2014 C. Phillips, Cyclotron Research Centre, ULg, Belgium
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
Multiple comparison correction Methods & models for fMRI data analysis October 2013 With many thanks for slides & images to: FIL Methods group & Tom Nichols.
Thresholding using FEAT
Multiple comparisons in M/EEG analysis Gareth Barnes Wellcome Trust Centre for Neuroimaging University College London SPM M/EEG Course London, May 2013.
1 Nonparametric Thresholding Methods (FWE inference w/ SnPM) Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics University of Michigan.
Classical Inference on SPMs Justin Chumbley SPM Course Oct 23, 2008.
Contrasts & Statistical Inference
**please note** Many slides in part 1 are corrupt and have lost images and/or text. Part 2 is fine. Unfortunately, the original is not available, so please.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005.
Random Field Theory Ciaran S Hill & Christian Lambert Methods for Dummies 2008.
The False Discovery Rate A New Approach to the Multiple Comparisons Problem Thomas Nichols Department of Biostatistics University of Michigan.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
1 Identifying Robust Activation in fMRI Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics University of Michigan
Statistical Analysis An Introduction to MRI Physics and Analysis Michael Jay Schillaci, PhD Monday, April 7 th, 2007.
FMRI Modelling & Statistical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Chicago, Oct.
Detecting connectivity: MS lesions, cortical thickness, and the “bubbles” task in the fMRI scanner Keith Worsley, McGill (and Chicago) Nicholas Chamandy,
Multiple comparisons problem and solutions James M. Kilner
The general linear model and Statistical Parametric Mapping I: Introduction to the GLM Alexa Morcom and Stefan Kiebel, Rik Henson, Andrew Holmes & J-B.
Topological Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course London, May 2015 With thanks to Justin.
Multiple comparison correction
False Discovery Rate for Functional Neuroimaging Thomas Nichols Department of Biostatistics University of Michigan Christopher Genovese & Nicole Lazar.
General Linear Model & Classical Inference London, SPM-M/EEG course May 2016 Sven Bestmann, Sobell Department, Institute of Neurology, UCL
General Linear Model & Classical Inference Short course on SPM for MEG/EEG Wellcome Trust Centre for Neuroimaging University College London May 2010 C.
What a Cluster F… ailure!
Topological Inference
General Linear Model & Classical Inference
The general linear model and Statistical Parametric Mapping
Statistical Inference
Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics
Detecting Sparse Connectivity: MS Lesions, Cortical Thickness, and the ‘Bubbles’ Task in an fMRI Experiment Keith Worsley, Nicholas Chamandy, McGill Jonathan.
Maxima of discretely sampled random fields
Methods for Dummies Random Field Theory
Multiple comparisons in M/EEG analysis
M/EEG Statistical Analysis & Source Localization
Topological Inference
Contrasts & Statistical Inference
Statistical Parametric Mapping
Contrasts & Statistical Inference
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
Statistical Inference
Contrasts & Statistical Inference
Presentation transcript:

Methods for Dummies Random Field Theory Annika Lübbert & Marian Schneider

Neural Correlates of Interspecies Perspective Taking in the Post-Mortem Atlantic Salmon task > rest contrast Bennett et al., (2010) in Journal of Serendipitous and Unexpected Results

Overview of Presentation 1. Multiple Comparisons Problem 2. Classical Approach to MCP 3. Random Field Theory 4. Implementation in SPM

Overview of Data Analysis so far

1. Multiple Comparisons Problem

Hypothesis Testing To test an hypothesis, we construct “test statistics” and ask how likely that our statistic could have come about by chance The Null Hypothesis H0 Typically what we want to disprove (no effect). The Alternative Hypothesis HA expresses outcome of interest. The Test Statistic T The test statistic summarises evidence about H0. Typically, test statistic is small in magnitude when the hypothesis H0 is true and large when false. We need to know the distribution of T under the null hypothesis Null Distribution of T

1. Multiple Comparisons Problem t uu  = p(t>u|H) Mass univariate analysis : perform t-tests for each voxel test t-statistic against the null hypothesis -> estimate how likely it is that our statistic could have come about by chance Decision rule (threshold) u, determines false positive rate  Choose u to give acceptable α  = P(type I error) i.e. chance we are wrong when rejecting the null hypothesis

1. Multiple Comparisons Problem t  uu t  uu t  uu t  uu t  uu Problem: fMRI – lots of voxels, lots of t-tests If use same threshold, inflated probability of obtaining false positives

Example -T-map for whole brain may contain say voxels -Each analysed separately would mean t-tests -At  = 0.05 this would be 3000 false positives (Type 1 Errors) -Adjust threshold so that any values above threshold are unlikely to under the null hypothesis (height thresholding & take into account the number of tests.) t > 0.5 t > 1.5 t > 2.5 t > 3.5 t > 4.5 t > 5.5t > 6.5 t > 0.5

Neural Correlates of Interspecies Perspective Taking in the Post-Mortem Atlantic Salmon task > rest contrast t(131) > 3.15, p(uncorrected) < 0.001,  resulting in 3 false positives

2. Classical Approach to MCP

Bonferroni Correction  = P FWE / n α = new single-voxel threshold n = number of tests (i.e. Voxels) FWE = family-wise error

Bonferroni Correction Example  = P FWE / n e.g. 100,000 t stats, all with 40 d.f. For P FWE of 0.05: 0.05/ = , corresponding t=5.77 => a voxel statistic of t>5.77 has only a 5% chance of arising anywhere in a volume of 100,000 t stats drawn from the null distribution single-voxel probability threshold  Number of voxels Family-wise error rate

Why Bonferroni is too conservative Functional imaging data has a degree of spatial correlation Number of independent values < number of voxels Why? The way that the scanner collects and reconstructs the image Physiology Spatial preprocessing (resampling, smoothing) Fundamental problem: image data represent situation in which we have a continuous statistic image, not a series of independent tests

Illustration Spatial Correlation Single slice image with 100 by 100 voxels Filling the voxel values with independent random numbers from the normal distribution  Bonferroni accurate

Illustration Spatial Correlation add spatial correlation: break up the image into squares of 10 by 10 pixels calculate the mean of the 100 values contained numbers in image but only 100 independent Bonferroni 100 times too conservative!

Smoothing contributes Spatial Correlation Smooth image by applying a Gaussian kernel with FWHM = 10 (at 5 pixels from centre, value is half peak value) Smoothing replaces each value in the image with weighted average of itself and neighbours Blurs the image -> reduces number of independent observations

Smoothing Kernel FWHM (Full Width at Half Maximum)

3. Random Field Theory

Using RFT to solve the Multiple Comparison Problem (get FWE) RFT approach treats SPMs as a discretisation of underlying continuous fields  Random fields have specified topological characteristics  Apply topological inference to detect activations in SPMs

Three stages of applying RFT: 1. Estimate smoothness 2. Find number of ‘resels’ (resolution elements) 3. Get an estimate of the Euler Characteristic at different thresholds

1. Estimate Smoothness Given: Image with applied and intrinsic smoothness Estimated using the residuals (error) from GLM  Estimate the increase in spatial correlation from i.i.d. noise to imaging data (‘from right to left’) estimate

2. Calculate the number of resels Look at your estimated smoothness (FWHM) Express your search volume in resels  Resolution element (R i = FWHM x x FWHM y x FWHM z )  # depends on smoothness of data (+volume of ROI)  “Restores” independence of the data

3. Get an estimate of the Euler Characteristic  Steps 1 and 2 yield a ‘fitted random field’ (appropriate smoothness)  Now: how likely is it to observe above threshold (‘significantly different’) local maxima (or clusters, sets) under H 0 ?  How to find out? EC!

Euler Characteristic – a topological property  Leonhard Euler 18 th century swiss mathematician  SPM  threshold  EC  EC = number of blobs (minus number of holes)* “Seven bridges of K ӧ nisberg” *Not relevant: we are only interested in EC at high thresholds (when it approximates P of FEW)

Euler Characteristic and FWE  The probability of a family wise error is approximately equivalent to the expected Euler Characteristic  Number of “above threshold blobs” Z t = 2.5 Z t = 2.75 EC = 3 EC = 1

How to get E [EC] at different thresholds # of resels Z (or T) -score threshold Expected Euler Characteristic * Or Monte Carlo Simulation, but the formula is good! It gives exact estimates, as long as your smoothness is appropriate

Given # of resels and E[EC], we can find the appropriate z/t-threshold E [EC] for an image of 100 resels, for Z score thresholds 0 – 5

How to get E [EC] at different thresholds # of resels Z (or T) -score threshold Expected Euler Characteristic From this equation, it looks like the threshold depends only on the number of resels in our image

Shape of search region matters, too! #resels  E [EC]  not strictly accurate!  close approximation if ROI is large compared to the size of a resel # + shape + size resel  E [EC]  Matters if small or oddly shaped ROI Example: 1.central 30x30 pixel box = max. 16 ‘resels’ (resel-width = 10pixel) 2.edge 2.5 pixel frame = same volume but max. 32 ‘resels’  Multiple comparison correction for frame must be more stringent

Different types of topological Inference Topological inference can be about  Peak height (voxel level)  Regional extent (cluster level)  Number of clusters (set level) space tt t clus  localizing power! …intermediate…  sensitivity! Different height and spatial extent thresholds

Assumptions made by RFT  Underlying fields are continuous with twice-differentiable autocorrelation function  Error fields are reasonable lattice approximation to underlying random field with multivariate Gaussian distribution lattice representation  Check: FWHM must be min. 3 voxels in each dimension  In general: if smooth enough + GLM correct (error = Gaussian)  RFT

Assumptions are not met if...  Data is not sufficiently smooth Small number of subjects High-resolution data  Non-parametric methods to get FWE (permutation)  Alternative ways of controlling for errors: FDR (false discovery rate)

4. Implementation in SPM

Another example:

Take-home message, overview: Large volume of imaging data Multiple comparison problem Bonferroni correction α=P FWE /n Corrected p value Null hypothesis rarely rejected Too many false negatives Mass univariate analysis Uncorrected p value Too many false positives Never use this. <smoothing with a Gaussian kernel, FWHM > Random field theory (RFT) α = P FWE ≒ E[EC] Corrected p value

Thank you for your attention! …and special thanks to Guillaume!

Sources Human Brain Function, Chapter 14, An Introduction to Random Field Theory (Brett, Penny, Kiebel) _MCPhttp:// _MCP Former MfD Presentations Expert Guillaume Flandin