Andrea Banino & Punit Shah. Samples vs Populations Descriptive vs Inferential William Sealy Gosset (Student) Distributions, probabilities and P-values.

Slides:



Advertisements
Similar presentations
T-tests, ANOVA and regression Methods for Dummies February 1 st 2006 Jon Roiser and Predrag Petrovic.
Advertisements

Carles Falcon & Suz Prejawa
T tests, ANOVAs and regression
General Linear Model L ύ cia Garrido and Marieke Schölvinck ICN.
General Linear Model Beatriz Calvo Davina Bristow.
2nd level analysis – design matrix, contrasts and inference
Managerial Economics in a Global Economy
SPM 2002 C1C2C3 X =  C1 C2 Xb L C1 L C2  C1 C2 Xb L C1  L C2 Y Xb e Space of X C1 C2 Xb Space X C1 C2 C1  C3 P C1C2  Xb Xb Space of X C1 C2 C1 
Jennifer Siegel. Statistical background Z-Test T-Test Anovas.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Linear regression models
Chapter 13 Multiple Regression
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Chapter 12 Multiple Regression
The Simple Regression Model
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
REGRESSION AND CORRELATION
Today Concepts underlying inferential statistics
Intro to Parametric Statistics, Assumptions & Degrees of Freedom Some terms we will need Normal Distributions Degrees of freedom Z-values of individual.
Correlation and Regression Analysis
Simple Linear Regression and Correlation
Lorelei Howard and Nick Wright MfD 2008
Relationships Among Variables
Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.
Lecture 5 Correlation and Regression
CHAPTER 5 REGRESSION Discovering Statistics Using SPSS.
Correlation and Regression
Inferential Statistics
General Linear Model & Classical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM M/EEGCourse London, May.
Lecture 16 Correlation and Coefficient of Correlation
The Chi-Square Distribution 1. The student will be able to  Perform a Goodness of Fit hypothesis test  Perform a Test of Independence hypothesis test.
Inference for regression - Simple linear regression
3/2003 Rev 1 I – slide 1 of 33 Session I Part I Review of Fundamentals Module 2Basic Physics and Mathematics Used in Radiation Protection.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Introduction to Linear Regression
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Correlation and Regression Used when we are interested in the relationship between two variables. NOT the differences between means or medians of different.
SPM short course – Oct Linear Models and Contrasts Jean-Baptiste Poline Neurospin, I2BM, CEA Saclay, France.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
ITEC6310 Research Methods in Information Technology Instructor: Prof. Z. Yang Course Website: c6310.htm Office:
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Chapter 10 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 A perfect correlation implies the ability to predict one score from another perfectly.
Contrasts & Statistical Inference
Correlation & Regression Analysis
Research Methodology Lecture No :26 (Hypothesis Testing – Relationship)
Chapter 13 Understanding research results: statistical inference.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
SPM short course – Mai 2008 Linear Models and Contrasts Jean-Baptiste Poline Neurospin, I2BM, CEA Saclay, France.
Inference about the slope parameter and correlation
AP Statistics Chapter 14 Section 1.
Correlation and Simple Linear Regression
Regression 10/29.
Chapter 11 Simple Regression
Kin 304 Inferential Statistics
CHAPTER 29: Multiple Regression*
Comparing Several Means: ANOVA
Ass. Prof. Dr. Mogeeb Mosleh
Contrasts & Statistical Inference
I. Statistical Tests: Why do we use them? What do they involve?
Product moment correlation
Inferential Statistics
Contrasts & Statistical Inference
Contrasts & Statistical Inference
MGS 3100 Business Analysis Regression Feb 18, 2016
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Andrea Banino & Punit Shah

Samples vs Populations Descriptive vs Inferential William Sealy Gosset (Student) Distributions, probabilities and P-values Assumptions of t-tests

P values = the probability that the observed result was obtained by chance i.e. when the null hypothesis is true α level is set a priori (Usually.05) If p <.05 level then we reject the null hypothesis and accept the experimental hypothesis 95% certain that our experimental effect is genuine If however, p >.05 level then we reject the experimental hypothesis and accept the null hypothesis

Is there different activation of the FFG for faces vs objects Within-subjects design: Condition 1: Presented with face stimuli Condition 2: Presented with object stimuli Hypotheses H 0 = There is no difference in activation of the FFG during face vs object stimuli H A = There is a significant difference in activation of the FFG during face vs object stimuli

Mean BOLD signal change during object stimuli = % Mean BOLD signal change during facial stimuli = +4% Great- there is a difference, but how do we know this was not just a fluke?

Compare the mean between 2 conditions (Faces vs Objects) H 0 : μ A = μ B (null hypothesis)- no difference in brain activation between these 2 groups/conditions H A : μ A μ B (alternative hypothesis) = there is a difference in brain activation between these 2 groups/conditions if 2 samples are taken from the same population, then they should have fairly similar means if 2 means are statistically different, then the samples are likely to be drawn from 2 different populations, i.e they really are different Condition 1 (Objects)Condition 2 (Faces) BOLD response

t = differences between sample means / standard error of sample means The exact equation varies depending on which type of t-test used Condition 1 (Objects) Condition 2 (Faces) BOLD response * Independent Samples t-test

1 Sample t-test (sample vs. hypothesized mean) 2 Sample t-test (group/condition 1 vs group/condition 2)

The number of entities that are free to vary when estimating t n – 1 (for paired sample t) Larger sample or no. of observations = more df Putting it all together… t (df) = t= t-value, p = p-value

Subtraction / Multiple subtraction Techniques compare the means and standard deviations between various conditions each voxel considered an n – so Bonferroni correction is made for the number of voxels compared Time

Normalisation Statistical Parametric Map Image time-series Parameter estimates General Linear Model RealignmentSmoothing Design matrix Anatomical reference Spatial filter Statistical Inference RFT p <0.05

Y = X. β + ε Observed data: Y is the BOLD signal at various time points at a single voxel Design matrix: Several components which explain the observed data, i.e. the BOLD time series for the voxel Parameters: Define the contribution of each component of the design matrix to the value of Y Estimated so as to minimise the error, ε, i.e. least sums of squares Error: Difference between the observed data, Y, and that predicted by the model, Xβ.

GLM: Y= X β + ε 2 nd level analysis β 1 is an estimate of signal change over time attributable to the condition of interest (face vs object) Set up contrast (c T ) 1 0 for β 1 : 1xβ 1 +0xβ 2 +0xβ n /s.d Null hypothesis: c T β=0 No significant effect at each voxel for condition β 1 Contrast 1 -1 : Is the difference between 2 conditions significantly non-zero? t = c T β/sd[c T β] t-tests are simple combinations of the betas; they are either positive or negative (b1 – b2 is different from b2 – b1)

A contrast = a weighted sum of parameters: c´ ´ b c = divide by estimated standard deviation of b 1 T test - one dimensional contrasts – SPM {t } SPM{t} b 1 > 0 ? cb Compute 1 x b x b x b x b x b = cb c = [ ….] b 1 b 2 b 3 b 4 b T = contrast of estimated parameters T = cb variance estimate s 2 c(XX) - c

More that 2 groups and/or conditions- e.g. objects, faces and bodies Do this without inflating the Type I error rate Still compares the differences in means between groups/conditions but it uses the variance of data to calculate if means are significantly different (H A ) Tests the null hypothesis that the means are the same via the F- test Extra assumptions

By comparing the variance (SS T =SS M +SS R ) SS T (variability between scores) SS M (variability explained by model) SS R (variability due to individual difference) F- ratio Magnitude of the difference between the different conditions p-value associated with F is probability that differences between groups could occur by chance if null-hypothesis is correct need for post-hoc testing / planned contrasts (ANOVA can tell you if there is an effect but not where) F-ratio = MS M / MS R ÷df M ÷df R

One- way Repeated measures / between groups ANOVA- One Factor, 3+ levels 2 way (_ x _) ANOVA and even 3 way ANOVA - Two or more factors and many levels:

Convolution model Design and contrast SPM(t) or SPM(F) Fitted and adjusted data Application to fMRI

PART 2 Correlation - How much linear is the relationship of two variables? (descriptive) Regression - How good is a linear model to explain my data? (inferential)

Correlation: -How much depend the value of one variable on the value of the other one? Y X Y X Y X high positive correlation poor negative correlation no correlation

How to describe correlation (1): Covariance -The covariance is a statistic representing the degree to which 2 variables vary together (note that S x 2 = cov(x,x) )

cov(x,y) = mean of products of each point deviation from mean values Geometrical interpretation: mean of signed areas from rectangles defined by points and the mean value lines

sign of covariance = sign of correlation Y X Y X Y X Positive correlation: cov > 0 Negative correlation: cov < 0 No correlation. cov 0

How to describe correlation (2): Pearson correlation coefficient (r) -r is a kind of normalised (dimensionless) covariance -r takes values fom -1 (perfect negative correlation) to 1 (perfect positive correlation). r=0 means no correlation (S = st dev of sample)

Pearson correlation coefficient (r) Problems: -It is sensitive to outliers Limitations: -r is an estimate from the sample, but does it represent the population parameter?

They all have r=0.816 but… They all have the same regression line: y = x

But remember: Not causality Relationship not a prediction

Linear regression: - Regression: Prediction of one variable from knowledge of one or more other variables - How good is a linear model (y=ax+b) to explain the relationship of two variables? - If there is such a relationship, we can predict the value y for a given x. But, which error could we be doing? (25, 7.498)

Preliminars: Lineal dependence between 2 variables Two variables are linearly dependent when the increase of one variable is proportional to the increase of the other one x y

The equation y= β 1 x+ β 0 that connects both variables has two parameters: - β 1 is the unitary increase/decerease of y (how much increases or decreases y when x increases one unity) - Slope - β 0 the value of y when x is zero (usually zero) - Intrercept

Fiting data to a straight line (o viceversa): Here, ŷ = ax + b – ŷ : predicted value of y –β 1 : slope of regression line –β 0 : intercept Residual error (ε i ): Difference between obtained and predicted values of y (i.e. y i - ŷ i ) Best fit line (values of b and a) is the one that minimises the sum of squared errors (SS error ) (y i - ŷ i ) 2 ε iε i ε i = residual = y i, observed = ŷ i, predicted ŷ = β 1 x + β 0

Adjusting the straight line to data: Minimise (y i - ŷ i ) 2, which is (y i -ax i +b) 2 Minimum SS error is at the bottom of the curve where the gradient is zero – and this can found with calculus Take partial derivatives of (y i -ax i -b) 2 respect parametres a and b and solve for 0 as simultaneous equations, giving: This calculus can allways be done, whatever is the data!!

How good is the model? We can calculate the regression line for any data, but how well does it fit the data? Total variance = predicted variance + error variance: S y 2 = S ŷ 2 + S er 2 Also, it can be shown that r 2 is the proportion of the variance in y that is explained by our regression model r 2 = S ŷ 2 / S y 2 Insert r 2 S y 2 into S y 2 = S ŷ 2 + S er 2 and rearrange to get: S er 2 = S y 2 (1 – r 2 ) From this we can see that the greater the correlation the smaller the error variance, so the better our prediction

Is the model significant? i.e. do we get a significantly better prediction of y from our regression equation than by just predicting the mean? F-statistic: And it follows that: F (df ŷ,df er ) = sŷ2sŷ2 s er 2 r 2 (n - 2) 2 1 – r 2 =......= complicated rearranging t (n-2) = r (n - 2) 1 – r 2 So all we need to know are r and n !!!

Generalization to multiple variables Multiple regression is used to determine the effect of a number of independent variables, x 1, x 2, x 3 etc., on a single dependent variable, y The different x variables are combined in a linear way and each has its own regression coefficient: y = x x 2 +…..+ n x n + ε The a parameters reflect the independent contribution of each independent variable, x, to the value of the dependent variable, y i.e. the amount of variance in y that is accounted for by each x variable after all the other x variables have been accounted for

Geometric view, 2 variables: ŷ = x x 2 x1x1 x2x2 y ε Plane of regression: Plane nearest all the sample points distributed over a 3D space: y = x x 2 + ε -> Hyperplane

Last remarks: - Relationship between two variables doesnt mean causality (e.g suicide - icecream) - Cov(x,y)=0 doesnt mean x,y being independents (yes for linear relationship but it could be quadratic,…)

References Field, A. (2009). Discovering Statistics Using SPSS (2nd ed). London: Sage Publications Ltd. Various MfD Slides SPM Course slides Wikipedia Judd, C.M., McClelland, G.H., Ryan, C.S. Data Analysis: A Model Comparison Approach, Second Edition. Routledge; Slide from PSYCGR01 Statistic course - UCL (dr. Maarten Speekenbrink)