ANOVA Greg C Elvers.

Slides:



Advertisements
Similar presentations
Analysis of Variance (ANOVA) ANOVA methods are widely used for comparing 2 or more population means from populations that are approximately normal in distribution.
Advertisements

Independent t -test Features: One Independent Variable Two Groups, or Levels of the Independent Variable Independent Samples (Between-Groups): the two.
Chapter 11 Analysis of Variance
Smith/Davis (c) 2005 Prentice Hall Chapter Thirteen Inferential Tests of Significance II: Analyzing and Interpreting Experiments with More than Two Groups.
C82MST Statistical Methods 2 - Lecture 4 1 Overview of Lecture Last Week Per comparison and familywise error Post hoc comparisons Testing the assumptions.
Analysis of Variance (ANOVA) Statistics for the Social Sciences Psychology 340 Spring 2010.
BHS Methods in Behavioral Sciences I April 25, 2003 Chapter 6 (Ray) The Logic of Hypothesis Testing.
Two Factor ANOVA.
Independent Sample T-test Formula
Lecture 10 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
Analysis of Variance: Inferences about 2 or More Means
Chapter 3 Analysis of Variance
PSY 307 – Statistics for the Behavioral Sciences
Intro to Statistics for the Behavioral Sciences PSYC 1900
Intro to Statistics for the Behavioral Sciences PSYC 1900
Lecture 9: One Way ANOVA Between Subjects
Two Groups Too Many? Try Analysis of Variance (ANOVA)
One-way Between Groups Analysis of Variance
Statistics for the Social Sciences
Today Concepts underlying inferential statistics
Statistical Methods in Computer Science Hypothesis Testing II: Single-Factor Experiments Ido Dagan.
Introduction to Analysis of Variance (ANOVA)
Chapter 14 Inferential Data Analysis
Richard M. Jacobs, OSA, Ph.D.
1 Two Factor ANOVA Greg C Elvers. 2 Factorial Designs Often researchers want to study the effects of two or more independent variables at the same time.
1 Nominal Data Greg C Elvers. 2 Parametric Statistics The inferential statistics that we have discussed, such as t and ANOVA, are parametric statistics.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Analysis of Variance (ANOVA) Quantitative Methods in HPELS 440:210.
Analysis of Variance. ANOVA Probably the most popular analysis in psychology Why? Ease of implementation Allows for analysis of several groups at once.
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
AM Recitation 2/10/11.
One-Way Analysis of Variance Comparing means of more than 2 independent samples 1.
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
January 31 and February 3,  Some formulae are presented in this lecture to provide the general mathematical background to the topic or to demonstrate.
t(ea) for Two: Test between the Means of Different Groups When you want to know if there is a ‘difference’ between the two groups in the mean Use “t-test”.
© Copyright McGraw-Hill CHAPTER 12 Analysis of Variance (ANOVA)
PSY 307 – Statistics for the Behavioral Sciences Chapter 16 – One-Factor Analysis of Variance (ANOVA)
Statistics 11 Confidence Interval Suppose you have a sample from a population You know the sample mean is an unbiased estimate of population mean Question:
Chapter 13 Analysis of Variance (ANOVA) PSY Spring 2003.
ANOVA (Analysis of Variance) by Aziza Munir
Psychology 301 Chapters & Differences Between Two Means Introduction to Analysis of Variance Multiple Comparisons.
Between-Groups ANOVA Chapter 12. >When to use an F distribution Working with more than two samples >ANOVA Used with two or more nominal independent variables.
Chapter 14 – 1 Chapter 14: Analysis of Variance Understanding Analysis of Variance The Structure of Hypothesis Testing with ANOVA Decomposition of SST.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
I. Statistical Tests: A Repetive Review A.Why do we use them? Namely: we need to make inferences from incomplete information or uncertainty þBut we want.
One-Way ANOVA ANOVA = Analysis of Variance This is a technique used to analyze the results of an experiment when you have more than two groups.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
ANOVA: Analysis of Variance.
Chapter 14 – 1 Chapter 14: Analysis of Variance Understanding Analysis of Variance The Structure of Hypothesis Testing with ANOVA Decomposition of SST.
Lecture 9-1 Analysis of Variance
Chapter 13 - ANOVA. ANOVA Be able to explain in general terms and using an example what a one-way ANOVA is (370). Know the purpose of the one-way ANOVA.
Analysis of Variance (One Factor). ANOVA Analysis of Variance Tests whether differences exist among population means categorized by only one factor or.
Chapter 17 Comparing Multiple Population Means: One-factor ANOVA.
1 ANALYSIS OF VARIANCE (ANOVA) Heibatollah Baghi, and Mastee Badii.
Chapter 12 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 Chapter 12: One-Way Independent ANOVA What type of therapy is best for alleviating.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: One-way ANOVA Marshall University Genomics Core.
Copyright © Cengage Learning. All rights reserved. 12 Analysis of Variance.
Hypothesis test flow chart frequency data Measurement scale number of variables 1 basic χ 2 test (19.5) Table I χ 2 test for independence (19.9) Table.
Statistics for the Social Sciences Psychology 340 Spring 2009 Analysis of Variance (ANOVA)
Copyright c 2001 The McGraw-Hill Companies, Inc.1 Chapter 11 Testing for Differences Differences betweens groups or categories of the independent variable.
Introduction to ANOVA Research Designs for ANOVAs Type I Error and Multiple Hypothesis Tests The Logic of ANOVA ANOVA vocabulary, notation, and formulas.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 11 Testing for Differences Differences betweens groups or categories of the independent.
BHS Methods in Behavioral Sciences I May 9, 2003 Chapter 6 and 7 (Ray) Control: The Keystone of the Experimental Method.
Formula for Linear Regression y = bx + a Y variable plotted on vertical axis. X variable plotted on horizontal axis. Slope or the change in y for every.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 4 Investigating the Difference in Scores.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Between-Groups ANOVA Chapter 12. Quick Test Reminder >One person = Z score >One sample with population standard deviation = Z test >One sample no population.
CHAPTER 10: ANALYSIS OF VARIANCE(ANOVA) Leon-Guerrero and Frankfort-Nachmias, Essentials of Statistics for a Diverse Society.
Reasoning in Psychology Using Statistics
Presentation transcript:

ANOVA Greg C Elvers

Multi-Level Experiments Often the research that psychologists perform has more conditions than just the control and experimental conditions You might want to compare the effectiveness of two treatments to each other and to no treatment Such research designs are called multi-level because there are more than two levels to the independent variable

Multiple t Tests How would you analyze data from a multi-level experiment? Given what we know so far, our only real option would be to perform multiple t-tests The problem with this approach to the data analysis is that you might have to perform many, many t-tests to test all possible combinations of the levels

Multiple t-Tests As the number of levels (or conditions) increases, the number of comparisons needed increases more rapidly # comparisons = (n2 - n) / 2 n = number of levels

Why Not Perform Multiple t Tests? With a computer, it is easy, quick, and error-free to perform multiple t-tests, so why shouldn’t we do it? The answer lies in a a is the probability of making a Type-I error, or incorrectly rejecting H0 when H0 is true a applies to a single comparison It is correctly called acomparison wise

Why Not Perform Multiple t Tests? What happens to the probability of making at least 1 Type-I error as the number of comparisons increases? It increases too The probability of making at least one Type-I error across all of your inferential statistical tests is called afamilywise or afw

afw If the probability of making a Type-I error on a given comparison is given by a, what is the probability of making at least 1 Type-I error when you make n comparisons? The simplest way to solve this problem is to realize that p(at least 1 Type-I error in n comparisons) + p(no Type-I errors in n comparisons) = 1 So, afw = 1 - p(no Type-I errors)

afw p(no Type-I errors in n comparisons) = p(no Type-I error on the 1st comparison) * p(no Type-I error on the 2nd comparison)* …* p(no Type-I error on the nth comparison) p(no Type-I errors in n comparisons) = p(no Type-I error on 1 comparison)n p(no Type-I error on 1 comparison) = 1 - a

afw afw = 1 - (1 - a)n As the number of comparisons increases, the probability of making at least 1 Type-I error increases rapidly

Why Not Multiple t-Tests? Performing multiple t-tests is bad because it increases the probability that you will make at least one Type-I error You can never determine which statistically significant results, if any, are probably due to chance

ANOVA Part of the solution to this problem rests in a statistical procedure known as the analysis of variance or ANOVA for short ANOVA replaces the multiple comparisons with a single omnibus null hypothesis Omnibus -- covers all aspects In ANOVA, H0 is always of the form: H0: m1 = m2 = m3 = …. = mn That is, H0 is that all the means are equal

ANOVA Alternative Hypothesis Given the H0 and H1 must be both mutually exclusive and exhaustive, what is H1 for ANOVA? Why isn’t this H1? H1: m1 ¹ m2 ¹ m3 ¹…. ¹ mn In ANOVA, the alternative hypothesis is always of the form: H1: not H0

Two Estimates of Variance ANOVA compares two estimates of the variability in the data in order to determine if H0 can be rejected: Between-groups variance Within-groups variance Do not confuse these terms with between-subject and within-subjects designs

Within-Groups Variance Within-groups variance is the weighted mean variability within each group or condition Which of the two distributions to the right has a larger within-groups variance? Why?

Within-Groups Variance What causes variability within a group? Within-groups variation is caused by factors that we cannot or did not control in the experiment, such as individual differences between the participants Do you want within-group variability to be large or small? Within-groups variation should be as small as possible as it is a source of error

Between-Groups Variance Between-groups variance is a measure of how different the groups are from each other Which distribution has a greater between-groups variance?

Sources of Between-Groups Variance What causes, or is the source of, between-groups variance? That is, why are not all the groups identical to each other? Between-groups variance is partially caused by the effect that the treatment has on the dependent variable The larger the effect is, the more different the groups become and the larger between-groups variance will be

Sources of Between-Groups Variance Between-groups variance also measures sampling error Even if the treatment had no effect on the dependent variable, you still would not expect the distributions to be equal because samples rarely are identical to each other That is, some of the between-groups variance is due to the fact that the groups are initially different due to random fluctuations (or errors) in our sampling

Variance Summary Within-groups variance measures error Between-groups variance measures the effect of the treatment on the dependent variable and error

F Ratio Fisher’s F ratio is defined as:

F Ratio What should F equal if H0 is true? When H0 is true, then there is no effect of the treatment on the DV Thus, when H0 is true, we have error / error which should equal 1 What should F be if H0 is not true? When H0 is not true, then there is an effect of the treatment on the DV Thus, when H0 is not true, F should be larger than 1

How Big Does F Have to Be? Several factors influence how big F has to be in order to reject the null hypothesis To reject H0, the calculated F must be larger than the critical F which can be found in a table Anything that makes the critical F value large will make it more difficult to reject H0

Factors That Influence the Size of the Critical F a level -- as a decreases, the size of the critical F increases Sample size -- as the sample size increase, the degrees of freedom for the within-groups variance increases, and the size of the critical F decreases As the sample becomes larger, it becomes more representative of the population

Factors That Influence the Size of the Critical F Number of conditions -- as the number of conditions increase, so does the degrees of freedom for the between-groups variance term If the degrees of freedom for the denominator are larger than 2, then the size of the critical F will decrease as the number of conditions increases The opposite is true if the degrees of freedom for the denominator equal 1 or 2

Factors That Influence the Size of the Observed F Experimental Control -- as experimental control increases, within-groups variance decreases, and the observed F increases

ANOVA Summary Table Source SS df MS F p Between-group 10 1 10 5 < .05 Within-group 20 10 2 Total 30 11 SS = sum of squares -- the numerator of the variance formula = (X-)2 df = degrees of freedom -- the denominator of the variance formula df between-groups = # levels - 1 df within-groups = S(# participants in a condition - 1) MS = SS / df F = MSbetween-groups / MSwithin-groups p = p value (significance level)

ANOVA Assumptions ANOVA makes certain assumptions: Sampling error is normal, or Gaussian in shape, and is centered around the mean of the distribution Homogeneity of variance -- the variability within each group is approximately equal Independence of observations

ANOVA Assumptions ANOVA is fairly robust (it will give good results even if the assumptions are violated) to the normality assumption and the homogeneity of variance assumption as long as: the number of participants in each group is equal the number of participants in each group is fairly large

Multiple Comparisons ANOVA only tells us if there is an effect That is, are the means of the groups not all equal? ANOVA does not tell us which means are different from which other means Multiple comparisons are used to determine which means are probably different from which other means

Multiple Comparisons Multiple comparisons are not fundamentally different from performing a t-test That is, both multiple comparisons and t-tests answer the same question: are two means different from each other The difference is that the multiple comparisons protect us from making a Type-I error even though we perform many such comparisons

Multiple Comparisons There are many types of multiple comparison E.g. Tukey, Newman-Keuls, Dunnett, Scheffé, Bonferroni There is little agreement about which particular one is best The different tests trade off statistical power for protection from Type-I errors

Tukey Tests We will consider only the Tukey test; other people may feel that other tests are more appropriate The Tukey test offers reasonable protection from Type-I errors while maintaining reasonable statistical power

Tukey Tests You should perform Tukey tests only when two criteria have been met: There are more than two levels to the IV The IV is statistically significant

Steps in Performing Tukey Tests Write the hypotheses H0: m1 = m2 H1: m1 ¹ m2 Specify the a level a = .05

Steps in Performing Tukey Tests Calculate the honestly significant difference (HSD) q,k,dfwithin-groups = tabled q value  =  level k = number of levels of the IV dfwithin-groups = degrees of freedom for MSwithin-groups MSwithin-groups = within-groups variance estimate n = number of participants

Steps in Performing Tukey Tests Take the difference of the means of the conditions you are comparing If the difference of the means is at least as large as the HSD, you can reject H0 Repeat for whatever other comparisons need to be made

A-Priori vs A-Posteriori Comparisons The Tukey comparisons we have been talking about are post-hoc or a-posteriori (after the fact) comparisons That is, we did not specify prior to the experiment that we would perform these particular comparisons A-posteriori comparisons are the most frequently used comparisons in the social sciences

A-Priori vs A-Posteriori Comparisons There are other comparison techniques that are appropriate to use when, prior to the study, you have decided which comparisons you would like to make These comparisons are a-priori (before the fact) comparisons

A-Priori vs A-Posteriori Comparisons A-posteriori comparisons should only be performed after finding a statistically significant result in an ANOVA This helps to reduce the probability of making a Type-I error A-priori comparisons can be, but need not be, preceded by an ANOVA A-priori tests tend to have greater statistical power than their a-posteriori counterparts

Between- vs Within-Subjects Designs Like the t-test, there are separate procedures to be used when you have between-subjects and when you have within-subjects designs Be sure to use the appropriate test SPSS (and many other statistical programs) call within-subjects ANOVA repeated measures ANOVA