Research Support Center Chongming Yang

Slides:



Advertisements
Similar presentations
T-tests, ANOVA and regression Methods for Dummies February 1 st 2006 Jon Roiser and Predrag Petrovic.
Advertisements

Week 2 – PART III POST-HOC TESTS. POST HOC TESTS When we get a significant F test result in an ANOVA test for a main effect of a factor with more than.
Randomized Complete Block and Repeated Measures (Each Subject Receives Each Treatment) Designs KNNL – Chapters 21,
Mixed Designs: Between and Within Psy 420 Ainsworth.
Test of (µ 1 – µ 2 ),  1 =  2, Populations Normal Test Statistic and df = n 1 + n 2 – 2 2– )1– 2 ( 2 1 )1– 1 ( 2 where ] 2 – 1 [–
One-Way and Factorial ANOVA SPSS Lab #3. One-Way ANOVA Two ways to run a one-way ANOVA 1.Analyze  Compare Means  One-Way ANOVA Use if you have multiple.
Topic 12: Multiple Linear Regression
Inference for Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Classical Regression III
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
Chapter 10 Simple Regression.
Lesson #23 Analysis of Variance. In Analysis of Variance (ANOVA), we have: H 0 :  1 =  2 =  3 = … =  k H 1 : at least one  i does not equal the others.
ANCOVA Psy 420 Andrew Ainsworth. What is ANCOVA?
Lecture 19: Tues., Nov. 11th R-squared (8.6.1) Review
Chapter 14 Conducting & Reading Research Baumgartner et al Chapter 14 Inferential Data Analysis.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Analysis of Variance & Multivariate Analysis of Variance
Correlation and Regression Analysis
T-tests and ANOVA Statistical analysis of group differences.
Correlation & Regression
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Leedy and Ormrod Ch. 11 Gray Ch. 14
Statistics for the Social Sciences Psychology 340 Fall 2013 Thursday, November 21 Review for Exam #4.
ANCOVA Lecture 9 Andrew Ainsworth. What is ANCOVA?
Simple Linear Regression
بسم الله الرحمن الرحیم.. Multivariate Analysis of Variance.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
Statistical Power 1. First: Effect Size The size of the distance between two means in standardized units (not inferential). A measure of the impact of.
Chapter 10 Analysis of Variance.
GENERAL LINEAR MODELS Oneway ANOVA, GLM Univariate (n-way ANOVA, ANCOVA)
Inferential Statistics
I. Statistical Tests: A Repetive Review A.Why do we use them? Namely: we need to make inferences from incomplete information or uncertainty þBut we want.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Orthogonal Linear Contrasts This is a technique for partitioning ANOVA sum of squares into individual degrees of freedom.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Lecture 9-1 Analysis of Variance
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Correlation & Regression. The Data SPSS-Data.htmhttp://core.ecu.edu/psyc/wuenschk/SPSS/ SPSS-Data.htm Corr_Regr.
ANCOVA. What is Analysis of Covariance? When you think of Ancova, you should think of sequential regression, because really that’s all it is Covariate(s)
General Linear Model.
Remember You just invented a “magic math pill” that will increase test scores. On the day of the first test you give the pill to 4 subjects. When these.
CHAPTER 10 ANOVA - One way ANOVa.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Experimental Statistics - week 9
ANOVA Overview of Major Designs. Between or Within Subjects Between-subjects (completely randomized) designs –Subjects are nested within treatment conditions.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
Formula for Linear Regression y = bx + a Y variable plotted on vertical axis. X variable plotted on horizontal axis. Slope or the change in y for every.
ENGR 610 Applied Statistics Fall Week 8 Marshall University CITE Jack Smith.
Topics, Summer 2008 Day 1. Introduction Day 2. Samples and populations Day 3. Evaluating relationships Scatterplots and correlation Day 4. Regression and.
ENGR 610 Applied Statistics Fall Week 11 Marshall University CITE Jack Smith.
Differences Among Groups
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Analyze Of VAriance. Application fields ◦ Comparing means for more than two independent samples = examining relationship between categorical->metric variables.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Chapter 14 Introduction to Multiple Regression
Week 2 – PART III POST-HOC TESTS.
Why is this important? Requirement Understand research articles
Factorial Experiments
ANOVA Econ201 HSTS212.
Data Analysis and Interpretation
Least Squares ANOVA & ANCOV
STAT120C: Final Review.
Joanna Romaniuk Quanticate, Warsaw, Poland
I. Statistical Tests: Why do we use them? What do they involve?
LEARNING OUTCOMES After studying this chapter, you should be able to
Presentation transcript:

Research Support Center Chongming Yang SPSS Workshop Research Support Center Chongming Yang

Causal Inference If A, then B, under condition C If A, 95% Probability B, under condition C

Student T Test (William S. Gossett’s pen name = student) Assumptions Small Sample Normally Distributed t distributions: t = [ x - μ ] / [ s / sqrt( n ) ] df = degrees of freedom=number of independent observations

Type of T Tests One sample Two independent samples Paired test against a specific (population) mean Two independent samples compare means of two independent samples that represent two populations Paired compare means of repeated samples

One Sample T Test Conceputally convert sample mean to t score and examine if t falls within acceptable region of distribution

Two Independent Samples

Paired Observation Samples d = difference value between first and second observations

Multiple Group Issues Groups A B C comparisons AB AC BC .95 .95 .95 Joint Probability that one differs from another .95*.95*.95 = .91

Analysis of Variance (ANOVA) Completely randomized groups Compare group variances to infer group mean difference Sources of Total Variance Within Groups Between Groups F distribution SSB = between groups sum squares SSW = within groups sum squares

Fisher-Snedecor Distribution

F Test Null hypothesis: 𝑥 1 = 𝑥 2 = 𝑥 3 ...= 𝑥 𝑛 Given df1 and df2, and F value, Determine if corresponding probability is within acceptable distribution region

Issues of ANOVA Indicates some group difference Does not reveal which two groups differ Needs other tests to identify specific group difference Hypothetical comparisons Contrast No Hypothetical comparisons Post Hoc ANOVA has been replaced by multiple regressions, which can also be replaced by General Linear Modeling (GLM)

Multiple Linear Regression Causes 𝑥 cab be continuous or categorical Effect 𝑦 is continuous measure Mild causal terms  predictors Objective  identify important 𝑥

Assumptions of Linear Regression Y and X have linear relations Y is continuous or interval & unbounded expected or mean of  = 0  = normally distributed not correlated with predictors Predictors should not be highly correlated No measurement error in all variables

Least Squares Solution Choose 𝛽 0 , 𝛽 1 , 𝛽 2 , 𝛽 3 ,... 𝛽 𝑘 to minimize the sum of square of difference between observed 𝑦 𝑖 and model estimated/predicted 𝑦 𝑖 Through solving many equations

Explained Variance in 𝑦

Standard Error of 𝛽

T Test significant of 𝛽 t = 𝛽 / SE𝛽 If t > a critical value & p <.05 Then 𝛽 is significantly different from zero

Confidence Intervals of 𝛽

Standardized Coefficient (𝛽𝑒𝑡𝑎) Make 𝛽s comparable among variables on the same scale (standardized scores)

Interpretation of 𝛽 If x increases one unit, y increases 𝛽 unit, given other values of X

Model Comparisons Complete Model: Reduced Model: Test F = Msdrop / MSE MS = mean square MSE = mean square error

Variable Selection Select significant from a pool of predictors Stepwise undesirable, see http://en.wikipedia.org/wiki/Stepwise_regression Forward Backward (preferable)

Dummy-coding of Nominal 𝑥 R = Race(1=white, 2=Black, 3=Hispanic, 4=Others) R d1 d2 d3 1 1 0 0 2 0 1 0 3 0 0 1 4 0 0 0 Include all dummy variables in the model, even if not every one is significant.

Interaction Create a product term X2X3 Include X2 and X3 even effects are not significant Interpret interaction effect: X2 effect depends on the level of X3.

Plotting Interaction Write out model with main and interaction effects, Use standardized coefficient Plug in some plausible numbers of interacting variables and calculate y Use one X for X dimension and Y value for the Y dimension See examples http://frank.itlab.us/datamodel/node104.html

Diagnostic Linear relation of predicted and observed (plotting Collinearity Outliers Normality of residuals (save residual as new variable)

Repeated Measures (MANOVA, GLM) Measure(s) repeated over time Change in individual cases (within)? Group differences (between, categorical x)? Covariates effects (continuous x)? Interaction between within and between variables?

Assumptions Normality Sphericity: Variances are equal across groups so that Total sum of squares can be partitioned more precisely into Within subjects Between subjects Error

Model 𝜇 = grand mean 𝜋 𝑖 = constant of individual i 𝜏 𝑗 = constant of jth treatment 𝜀 𝑖𝑗 = error of i under treatment j 𝜋𝜏 = interaction

F Test of Effects F = MSbetween / Mswithin (simple repeated) F = Mstreatment / Mserror (with treatment) F = Mswithin / Msinteraction (with interaction)

Four Types Sum-Squares Type I  balanced design Type II  adjusting for other effects Type III  no empty cell unbalanced design Type VI  empty cells

Exercise http://www.ats.ucla.edu/stat/spss/seminars/Repeated_Measures/default.htm Copy data to spss syntax window, select and run Run Repeated measures GLM