1st level analysis: basis functions and correlated regressors

Slides:



Advertisements
Similar presentations
The SPM MfD course 12th Dec 2007 Elvina Chu
Advertisements

Basis Functions. What’s a basis ? Can be used to describe any point in space. e.g. the common Euclidian basis (x, y, z) forms a basis according to which.
1st level analysis: basis functions, parametric modulation and correlated regressors. 1 st of February 2012 Sylvia Kreutzer Max-Philipp Stenner Methods.
SPM 2002 C1C2C3 X =  C1 C2 Xb L C1 L C2  C1 C2 Xb L C1  L C2 Y Xb e Space of X C1 C2 Xb Space X C1 C2 C1  C3 P C1C2  Xb Xb Space of X C1 C2 C1 
Outline What is ‘1st level analysis’? The Design matrix
The General Linear Model Or, What the Hell’s Going on During Estimation?
Classical inference and design efficiency Zurich SPM Course 2014
Statistical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Zurich, February 2009.
The General Linear Model (GLM) Methods & models for fMRI data analysis in neuroeconomics November 2010 Klaas Enno Stephan Laboratory for Social & Neural.
The General Linear Model (GLM)
The General Linear Model (GLM) SPM Course 2010 University of Zurich, February 2010 Klaas Enno Stephan Laboratory for Social & Neural Systems Research.
Statistical Inference
Parametric modulation, temporal basis functions and correlated regressors Mkael Symmonds Antoinette Nicolle Methods for Dummies 21 st January 2008.
Methods for Dummies General Linear Model
SPM short course – May 2003 Linear Models and Contrasts The random field theory Hammering a Linear Model Use for Normalisation T and F tests : (orthogonal.
General Linear Model & Classical Inference
General Linear Model & Classical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM M/EEGCourse London, May.
With many thanks for slides & images to: FIL Methods group, Virginia Flanagin and Klaas Enno Stephan Dr. Frederike Petzschner Translational Neuromodeling.
Contrasts (a revision of t and F contrasts by a very dummyish Martha) & Basis Functions (by a much less dummyish Iroise!)
General Linear Model & Classical Inference London, SPM-M/EEG course May 2014 C. Phillips, Cyclotron Research Centre, ULg, Belgium
SPM short course – Oct Linear Models and Contrasts Jean-Baptiste Poline Neurospin, I2BM, CEA Saclay, France.
Contrasts & Statistical Inference
The General Linear Model (for dummies…) Carmen Tur and Ashwani Jha 2009.
Temporal Basis Functions Melanie Boly Methods for Dummies 27 Jan 2010.
Event-related fMRI SPM course May 2015 Helen Barron Wellcome Trust Centre for Neuroimaging 12 Queen Square.
Ch. 5 Bayesian Treatment of Neuroimaging Data Will Penny and Karl Friston Ch. 5 Bayesian Treatment of Neuroimaging Data Will Penny and Karl Friston 18.
Statistical Inference Christophe Phillips SPM Course London, May 2012.
FMRI Modelling & Statistical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Chicago, Oct.
The General Linear Model
The General Linear Model Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM fMRI Course London, May 2012.
SPM short – Mai 2008 Linear Models and Contrasts Stefan Kiebel Wellcome Trust Centre for Neuroimaging.
1 st level analysis: Design matrix, contrasts, and inference Stephane De Brito & Fiona McNabe.
The general linear model and Statistical Parametric Mapping I: Introduction to the GLM Alexa Morcom and Stefan Kiebel, Rik Henson, Andrew Holmes & J-B.
Contrasts & Statistical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course London, October 2008.
SPM and (e)fMRI Christopher Benjamin. SPM Today: basics from eFMRI perspective. 1.Pre-processing 2.Modeling: Specification & general linear model 3.Inference:
The general linear model and Statistical Parametric Mapping II: GLM for fMRI Alexa Morcom and Stefan Kiebel, Rik Henson, Andrew Holmes & J-B Poline.
The General Linear Model Christophe Phillips SPM Short Course London, May 2013.
The General Linear Model Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM fMRI Course London, October 2012.
SPM short course – Mai 2008 Linear Models and Contrasts Jean-Baptiste Poline Neurospin, I2BM, CEA Saclay, France.
General Linear Model & Classical Inference London, SPM-M/EEG course May 2016 Sven Bestmann, Sobell Department, Institute of Neurology, UCL
General Linear Model & Classical Inference Short course on SPM for MEG/EEG Wellcome Trust Centre for Neuroimaging University College London May 2010 C.
The General Linear Model (GLM)
The General Linear Model (GLM)
General Linear Model & Classical Inference
The general linear model and Statistical Parametric Mapping
The General Linear Model
Statistical Inference
Design Matrix, General Linear Modelling, Contrasts and Inference
Statistical Inference
SPM short course at Yale – April 2005 Linear Models and Contrasts
MfD 22/11/17 Kate Ledingham and Arabella Bird
and Stefan Kiebel, Rik Henson, Andrew Holmes & J-B Poline
Barbora Ondrejickova Methods for Dummies 23rd November 2016
The SPM MfD course 12th Dec 2007 Elvina Chu
The General Linear Model (GLM)
Sam Ereira Methods for Dummies 13th January 2016
Contrasts & Statistical Inference
The General Linear Model
The general linear model and Statistical Parametric Mapping
The General Linear Model
Karl Friston, Oliver Josephs
The General Linear Model (GLM)
Contrasts & Statistical Inference
MfD 04/12/18 Alice Accorroni – Elena Amoruso
The General Linear Model
The General Linear Model (GLM)
Statistical Inference
The General Linear Model
The General Linear Model
Contrasts & Statistical Inference
Presentation transcript:

1st level analysis: basis functions and correlated regressors Methods for Dummies 03/12/2014 Steffen Volz Faith Chiu

Overview Part 1: basis functions (Steffen Volz) Modeling of the BOLD signal What are basis functions Which choice of basis functions Part 2: correlated regressors (Faith Chiu) …

Statistical Inference Where are we? Normalisation Statistical Parametric Map Image time-series Parameter estimates General Linear Model Realignment Smoothing Design matrix Anatomical reference Spatial filter Statistical Inference RFT p <0.05

Modeling of the BOLD signal BOLD signal is not a direct measure of neuronal activity, but a function of the blood oxygenation, flow and volume (Buxton et al, 1998)  hemodynamic response function (HRF) The response is delayed compared to stimulus The response extends over about 20s  overlap with other stimuli Response looks different between regions (Schacter et al 1997) and subjects (Aguirre et al, 1998)  Model signal within General Linear Model (GLM)

Modeling of the BOLD signal Properties of BOLD response: Initial undershoot (Malonek & Grinvald, 1996) Peak after about 4-6s Final undershoot Back to baseline after 20-30s Brief Stimulus Undershoot Initial Peak

Temporal basis functions BOLD response can look different between ROIs and subjects To account for this temporal basis functions are used for modeling within General Linear Model (GLM) every time course can be constructed by a set of basis functions D

Temporal basis functions BOLD response can look different between ROIs and subjects To account for this temporal basis functions are used for modeling within General Linear Model (GLM) every time course can be constructed by a set of basis functions Different basis functions: Fourier basis Finite impulse response Gamma functions Informed basis set

Temporal basis functions Finite Impulse Response (FIR): poststimulus timebins (“mini- boxcars”) Captures any shape (bin width) Inference via F-test

Temporal basis functions Finite Impulse Response (FIR): poststimulus timebins (“mini- boxcars”) Captures any shape (bin width) Inference via F-test Fourier Base: Windowed sines & cosines Captures any shape (frequency limit)

Temporal basis functions Gamma Functions: Bounded, asymmetrical (like BOLD) Set of different lags Inference via F-test

Informed Basis Set (Friston et al. 1998) Canonical HRF: combination of 2 Gamma functions (best guess of BOLD response)

Informed Basis Set (Friston et al. 1998) Canonical HRF: combination of 2 Gamma functions (best guess of BOLD response) Variability captured by Taylor expansion: Temporal derivative (account for differences in the latency of response)

Informed Basis Set (Friston et al. 1998) Canonical HRF: combination of 2 Gamma functions (best guess of BOLD response) Variability captured by Taylor expansion: Temporal derivative (account for differences in the latency of response) Dispersion derivative (account for differences in the duration of response)

Basis functions in SPM

Basis functions in SPM

Which basis to choose? Canonical + Temporal + Dispersion + FIR Example: rapid motor response to faces (Henson et al, 2001) canonical HRF alone insufficient to capture full range of BOLD responses significant additional variability captured by including partial derivatives combination appears sufficient (little additional variability captured by FIR set) More complex with protracted processes (eg. stimulus-delay-response) could not be captured by canonical set, but benefit from FIR set Canonical + Temporal + Dispersion + FIR

Summary part 1 Basis functions are used in SPM to model the hemodynamic response either using a single basis function or a set of functions. The most common choice is the “Canonical HRF” (Default in SPM) time and dispersion derivatives additionally account for variability of signal change over voxels

Correlated Regressors Faith Chiu

>1 x-value Linear regression Multiple regression y = X.b + e Y = β1X1 + β2X2 + … + βLXL + ε Only 1 x-variable >1 x-variable

Multiple regression ^ y = b0 + b1.x1 + b2.x2

Why are you telling me about this? In the General Linear Model (GLM) of SPM, Coefficients (b/β) are parameters which weight the value of your… Regressors (x1, x2), the design matrix GLM deals with the time series in voxel in a linear combination Y = X . β + ε Observed data Design matrix Parameters Error/residual

>1 y-value Linear regression General linear model (GLM) y = X.b + e Multiple y variables: time series in voxel Single dependent variable y Y = vector y = scalar

Single voxel regression model error = + 1 2 + Time x1 x2 e BOLD signal

X y Mass-univariate analysis: voxel-wise GLM + = Model is specified by both Design matrix X Assumptions about e N: number of scans p: number of regressors The design matrix embodies all available knowledge about experimentally controlled factors and potential confounds.

Ordinary least squares estimation (OLS) (assuming i.i.d. error): Parameter estimation Objective: estimate parameters to minimize = + y X Ordinary least squares estimation (OLS) (assuming i.i.d. error):

A geometric perspective on the GLM y e Design space defined by X x1 x2 Smallest errors (shortest error vector) when e is orthogonal to X Ordinary Least Squares (OLS)

Orthogonalisation y x2 x2* x1 When x2 is orthogonalized w.r.t. x1, only the parameter estimate for x1 changes, not that for x2! Correlated regressors = explained variance is shared between regressors

Practicalities re: multicollinearity Interpreting results of multiple regression can be difficult: the overall p-value of a fitted model is very low i.e. the model fits the data well but individual p values for the regressors are high i.e. none of the X variables has a significant impact on predicting Y How is this possible? caused when two (or more) regressors are highly correlated: problem known as multicollinearity

Multicollinearity Are correlated regressors a problem? No When you want to predict Y from X1 & X2, because R2 and p will be correct Yes When you want to assess the impact of individual regressors Because individual p-values can be misleading: a p-value can be high, even though the variable is improtant

Final word When you have correlated regressors, it is very rare that orthogonalisation will be a solution. You usually don't have an a priori hypothesis about which regressor should be given the shared variance. The solution is rather at the stage of the experiment definition where you would make sure by experimental design to decorrelate as much as possible the regressors that you want to look at independently.

Thanks Guillaume SPM course video on GLM Slides from previous years Rik Henson’s MRC CBU page: http://imaging.mrc-cbu.cam.ac.uk/imaging/SpmMiniCourse?action=AttachFile&do=view&target=SPM-Henson-3-design.ppt http://imaging.mrc-cbu.cam.ac.uk/imaging/DesignEfficiency#Correlation_between_regressors