One-way ANOVA: - Inference for one-way ANOVA IPS chapter 12.1 © 2006 W.H. Freeman and Company.

Slides:



Advertisements
Similar presentations
Chapter 9 Introduction to the t-statistic
Advertisements

BPS - 5th Ed. Chapter 241 One-Way Analysis of Variance: Comparing Several Means.
CHAPTER 25: One-Way Analysis of Variance Comparing Several Means
CHAPTER 25: One-Way Analysis of Variance: Comparing Several Means ESSENTIAL STATISTICS Second Edition David S. Moore, William I. Notz, and Michael A. Fligner.
The Analysis of Variance
Analysis of Variance (ANOVA) ANOVA can be used to test for the equality of three or more population means We want to use the sample results to test the.
ANOVA: Analysis of Variation
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. Analysis of Variance Chapter 16.
© 2010 Pearson Prentice Hall. All rights reserved Single Factor ANOVA.
Copyright ©2011 Brooks/Cole, Cengage Learning Analysis of Variance Chapter 16 1.
Statistics 303 Chapter 12 ANalysis Of VAriance. ANOVA: Comparing Several Means The statistical methodology for comparing several means is called analysis.
PSY 307 – Statistics for the Behavioral Sciences
Topic 2: Statistical Concepts and Market Returns
Inferences About Means of Two Independent Samples Chapter 11 Homework: 1, 2, 4, 6, 7.
ANOVA Single Factor Models Single Factor Models. ANOVA ANOVA (ANalysis Of VAriance) is a natural extension used to compare the means more than 2 populations.
Lecture 12 One-way Analysis of Variance (Chapter 15.2)
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved.
Inferences About Process Quality
Statistics for the Social Sciences
Two-sample problems for population means BPS chapter 19 © 2006 W.H. Freeman and Company.
IPS Chapter 12 © 2010 W.H. Freeman and Company  Inference for On-Way Analysis of Variance  Comparing the Means One-Way Analysis of Variance.
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
F-Test ( ANOVA ) & Two-Way ANOVA
Experimental Statistics - week 2
Hypothesis testing – mean differences between populations
1 1 Slide © 2005 Thomson/South-Western Chapter 13, Part A Analysis of Variance and Experimental Design n Introduction to Analysis of Variance n Analysis.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 13 Experimental Design and Analysis of Variance nIntroduction to Experimental Design.
Inference for distributions: - Comparing two means IPS chapter 7.2 © 2006 W.H. Freeman and Company.
 The idea of ANOVA  Comparing several means  The problem of multiple comparisons  The ANOVA F test 1.
Lecture Presentation Slides SEVENTH EDITION STATISTICS Moore / McCabe / Craig Introduction to the Practice of Chapter 12 One-Way Analysis of Variance.
ANOVA One Way Analysis of Variance. ANOVA Purpose: To assess whether there are differences between means of multiple groups. ANOVA provides evidence.
PSY 307 – Statistics for the Behavioral Sciences Chapter 16 – One-Factor Analysis of Variance (ANOVA)
One-Way ANOVA Ch 12.  Recall:  A categorical variable = factor.  Its values =levels  ANOVA in general studies the effect of categorical variables.
ANOVA (Analysis of Variance) by Aziza Munir
Testing Multiple Means and the Analysis of Variance (§8.1, 8.2, 8.6) Situations where comparing more than two means is important. The approach to testing.
Basic concept Measures of central tendency Measures of central tendency Measures of dispersion & variability.
1 Chapter 13 Analysis of Variance. 2 Chapter Outline  An introduction to experimental design and analysis of variance  Analysis of Variance and the.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Chapter 19 Analysis of Variance (ANOVA). ANOVA How to test a null hypothesis that the means of more than two populations are equal. H 0 :  1 =  2 =
Comparing Three or More Means ANOVA (One-Way Analysis of Variance)
Two-Way ANOVA Ch 13. In a 2-way design, 2 factors (independent variables) are studied in conjunction with the response (dependent) variable. There is.
Analysis of Variance (One Factor). ANOVA Analysis of Variance Tests whether differences exist among population means categorized by only one factor or.
One-way ANOVA: - Comparing the means IPS chapter 12.2 © 2006 W.H. Freeman and Company.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.1 One-Way ANOVA: Comparing.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.3 Two-Way ANOVA.
Copyright © Cengage Learning. All rights reserved. 12 Analysis of Variance.
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics S eventh Edition By Brase and Brase Prepared by: Lynn Smith.
AP Statistics: ANOVA Section 1. In section 13.1 A, we used a t-test to compare the means between two groups. An ANOVA (ANalysis Of VAriance) test is used.
CHAPTER 27: One-Way Analysis of Variance: Comparing Several Means
Objectives (IPS Chapter 12.1) Inference for one-way ANOVA  Comparing means  The two-sample t statistic  An overview of ANOVA  The ANOVA model  Testing.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
Analysis of Variance STAT E-150 Statistical Methods.
1/54 Statistics Analysis of Variance. 2/54 Statistics in practice Introduction to Analysis of Variance Analysis of Variance: Testing for the Equality.
Inference for a population mean BPS chapter 16 © 2006 W.H. Freeman and Company.
Objectives (PSLS Chapter 18) Comparing two means (σ unknown)  Two-sample situations  t-distribution for two independent samples  Two-sample t test 
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
One-Way ANOVA: One-Way Analysis of Variance PBS Chapter 14.1 © 2009 W.H. Freeman and Company.
Lecture notes 13: ANOVA (a.k.a. Analysis of Variance)
ANOVA: Analysis of Variation
24. One-way ANOVA: Comparing several means
ANOVA: Analysis of Variation
ANOVA: Analysis of Variation
ANOVA: Analysis of Variation
Basic Practice of Statistics - 5th Edition
Hypothesis testing using contrasts
11/ Inference for one-way ANOVA
Statistics for Business and Economics (13e)
Chapter 10 – Part II Analysis of Variance
ANOVA: Analysis of Variance
Presentation transcript:

One-way ANOVA: - Inference for one-way ANOVA IPS chapter 12.1 © 2006 W.H. Freeman and Company

Objectives (IPS chapter 12.1) Inference for one-way ANOVA  The concept of ANOVA  The ANOVA F-test  The ANOVA table  Using Table E  Computation details

The one-way layout  Suppose we have two or more experimental conditions (treatments) we would like to compare.  Usually that comparison takes the form of testing the hypothesis of equal means H o :  1 =  2 = … =  k.  In theory we can select a random sample from the k populations associated with each treatment.  More practically we can identify N “experimental units” and randomly assign the treatments to those units.

The concept of ANOVA Reminders: A categorical factor is a variable that can take on any of several levels used to differentiate one group from another. An experiment has a one-way, or completely randomized, design if several levels of one factor are being studied and the individuals are randomly assigned to those levels. (There is only one way to group the data. )  Example: Four levels of nematode quantity in seedling growth experiment.  Example: Student performance is evaluated with and without (2 levels) “computer aided” instruction Analysis of variance (ANOVA) is the technique used to test the equality of k > 2 means. One-way ANOVA is used for completely randomized, one-way designs.

How do we compare several means? We want to know if the observed differences in sample means are likely to have occurred by chance. Our decision depends partly on the amount of overlap between the groups which depends on the differences between the means and the amount of variability within the groups.

 We first examine the samples to test for overall significance as evidence of any difference among the means.  ANOVA F-test  If that overall test indicates statistical significance, then a follow-up comparison of combinations of means is in order.  If we planned our experiment with specific alternative hypotheses in mind (before gathering the data), we can test them using contrasts.  If we do not have specific alternatives, we can examine all pair-wise parameter comparisons to define which parameters differ from which, using multiple comparisons procedures.

Do nematodes affect plant growth? A botanist prepares 16 identical planting pots and adds different numbers of nematodes into the pots. Seedling growth (in mm) is recorded two weeks later. Nematodes and plant growth Hypotheses:  i are all equal (H 0 ) versus not All  i are the same (H a )

The ANOVA model Random sampling always produces chance variation. Any “factor effect” would thus show up in our data as the factor-driven differences plus chance variations (“error”): Data = fit (“factor/groups”) + residual (“error”) The one-way ANOVA model analyses situations where chance variations are normally distributed N(0,σ) so that:

The ANOVA F-test We have I independent SRSs, from I populations or treatments. The i th population has a normal distribution with unknown mean µ i. All I populations have the same standard deviation σ, unknown. The ANOVA F statistic tests: When H 0 is true, F has the F distribution with I − 1 ( numerator ) and N − I ( denominator ) degrees of freedom. H 0 :   =   = … =  I H a : not all the  i are equal.

The ANOVA F-test Alternatively and more practically we can randomly assign the treatments to a collection of N experimental units so that n 1 units get treatment 1, n 2 units get treatment 2, and so on. We then proceed as before. When H 0 is true, F has the F distribution with I − 1 ( numerator ) and N − I ( denominator ) degrees of freedom. H 0 :   =   = … =  I H a : not all the  i are equal.

Difference in means small relative to overall variability Difference in means large relative to overall variability Larger F-values lead to more significant results. How large it needs to be in order to be significant depends on the degrees of freedom (I − 1 and N − I). The ANOVA F-statistic compares variation due to treatments (levels of the factor) with variation among individuals who should be similar (individuals in the same sample).  F tends to be small  F tends to be large

Checking our assumptions The ANOVA F-test theory also requires that all populations have the same standard deviation . Practically: The results of the ANOVA F-test are approximately correct when the largest sample standard deviation is no more than twice as large as the smallest sample standard deviation. (Equal sample sizes also make ANOVA more robust to deviations from the equal  rule) “Theory” suggests each of the populations must be normally distributed. But the test is robust to deviations from normality for reasonably sized samples, thanks to the central limit theorem.

Do nematodes affect plant growth? Seedling growth x¯ i s i 0 nematode nematodes nematodes nematodes Conditions required: equal variances: checking that largest s i no more than twice smallest s i Largest s i = 2.053; smallest s i = Independent SRSs Four groups, assumed independent Distributions “roughly” normal It is hard to assess normality with only four points per condition. But the pots in each group are identical, and there are no outliers.

A study of the effect of smoking classifies subjects as nonsmokers, moderate smokers, and heavy smokers. The investigators interview a random sample of 200 people in each group and ask “How many hours do you sleep on a typical night?” 1. Study design? 2. Hypotheses? 3. ANOVA assumptions? 4. Degrees of freedom? 1. This is an observational study. Explanatory variable: smoking -- 3 levels: nonsmokers, moderate smokers, heavy smokers Response variable: # hours of sleep per night 2. H 0 : all 3  i equal (versus not all equal) 3. Three obviously independent SRS. Sample size of 200 should accommodate any departure from normality. Would still be good to check for s min /s max. 4. I = 3, n1 = n2 = n3 = 200, and N = 600, so there are I - 1 = 2 (numerator) and N - I = 597 (denominator) degrees of freedom. Smoking influence on sleep

The ANOVA table Source of variationSum of squares SS DFMean square MS FP valueF crit Among or between “groups” I -1SSG/DFGMSG/MSETail area above F Value of F for  Within groups or “error” N - ISSE/DFE TotalSST=SSG+SSEN – 1 R 2 = SSG/SST√MSE = s p Coefficient of determinationPooled standard deviation The sum of squares represents variation in the data: SST = SSG + SSE. The degrees of freedom likewise reflect the ANOVA model: DFT = DFG + DFE. Data (“Total”) = fit (“Groups”) + residual (“Error”)

Using Table E The F distribution is asymmetrical and has two distinct degrees of freedom. This was discovered by Fisher, hence the label “F.” Once again, what we do is calculate the value of F for our sample data and then look up the corresponding area under the curve in Table E.

df num = I − 1 df den = N − I F p For df: 5,4 Table E

F critical for  5% is 3.49 F = > Thus p < 0.001

Yogurt preparation and taste Yogurt can be made using three distinct commercial preparation methods: traditional, ultra filtration, and reverse osmosis. To study the effect of these methods on taste, an experiment was designed where three batches of yogurt were prepared for each of the three methods. A trained expert tasted each of the nine samples, presented in random order, and judged them on a scale of 1 to 10. Variables, hypotheses, assumptions, calculations? ANOVA table Source of variationSSdfMSFP-valueF crit Between groups17.3I-1= Within groups4.6N-I= Total17.769

df num = I − 1 df den = N − I F

MSG, the mean square for groups, measures how different the individual means are from the overall mean (~ weighted average of square distances of sample averages to the overall mean). SSG is the sum of squares for groups. MSE, the mean square for error is the pooled sample variance s p 2 and estimates the common variance σ 2 of the I populations (~ weighted average of the variances from each of the I samples). SSG is the sum of squares for error. Computation details

Note: Two sample t-test and ANOVA A two sample t-test assuming equal variance and an ANOVA comparing only two groups will give you the exact same p-value (for a two-sided hypothesis). H 0 :    =   H a :    ≠   t-test assuming equal variance t-statistic H 0 :    =   H a :    ≠   One-way ANOVA F-statistic F = t 2 and both p-values are the same. But the t-test is more flexible: You may choose a one-sided alternative instead, or you may want to run a t-test assuming unequal variance if you are not sure that your two populations have the same standard deviation .