Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.1 One-Way ANOVA: Comparing.

Slides:



Advertisements
Similar presentations
BPS - 5th Ed. Chapter 241 One-Way Analysis of Variance: Comparing Several Means.
Advertisements

Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. Analysis of Variance Chapter 16.
Chapter 13 Multiple Regression
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
The Two Factor ANOVA © 2010 Pearson Prentice Hall. All rights reserved.
Copyright ©2011 Brooks/Cole, Cengage Learning Analysis of Variance Chapter 16 1.
Part I – MULTIVARIATE ANALYSIS
Chapter 12 Multiple Regression
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Copyright © 2006 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide Are the Means of Several Groups Equal? Ho:Ha: Consider the following.
Inferences About Process Quality
Chi-Square and F Distributions Chapter 11 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
Chapter 12: Analysis of Variance
F-Test ( ANOVA ) & Two-Way ANOVA
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 12 Analyzing the Association Between Quantitative Variables: Regression Analysis Section.
Chapter 13: Inference in Regression
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 14 Analysis.
Hypothesis Testing in Linear Regression Analysis
QNT 531 Advanced Problems in Statistics and Research Methods
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.2 Estimating Differences.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
STA291 Statistical Methods Lecture 31. Analyzing a Design in One Factor – The One-Way Analysis of Variance Consider an experiment with a single factor.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Comparing Three or More Means 13.
Agresti/Franklin Statistics, 1 of 82 Chapter 13 Comparing Groups: Analysis of Variance Methods Learn …. How to use Statistical inference To Compare Several.
More About Significance Tests
MADAM SITI AISYAH ZAKARIA EQT 271 SEM /2015 CHAPTER 3 ANOVA (EXTRA NOTE & EXERCISE)
Chapter 14 Introduction to Multiple Regression
CHAPTER 14 MULTIPLE REGRESSION
Agresti/Franklin Statistics, 1 of 106  Section 9.4 How Can We Analyze Dependent Samples?
© Copyright McGraw-Hill CHAPTER 12 Analysis of Variance (ANOVA)
Copyright © 2004 Pearson Education, Inc.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Copyright © 2011 Pearson Education, Inc. Analysis of Variance Chapter 26.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Section Inference about Two Means: Independent Samples 11.3.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
Chap 14-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics.
+ Chapter 12: More About Regression Section 12.1 Inference for Linear Regression.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 10 Comparing Two Groups Section 10.4 Analyzing Dependent Samples.
Section 9-1: Inference for Slope and Correlation Section 9-3: Confidence and Prediction Intervals Visit the Maths Study Centre.
Analysis of Variance 1 Dr. Mohammed Alahmed Ph.D. in BioStatistics (011)
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.3 Using Multiple Regression to Make Inferences.
Analysis of Variance (One Factor). ANOVA Analysis of Variance Tests whether differences exist among population means categorized by only one factor or.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.1 One-Way ANOVA: Comparing.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Agresti/Franklin Statistics, 1 of 88 Chapter 11 Analyzing Association Between Quantitative Variables: Regression Analysis Learn…. To use regression analysis.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.3 Two-Way ANOVA.
Copyright © Cengage Learning. All rights reserved. 12 Analysis of Variance.
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics S eventh Edition By Brase and Brase Prepared by: Lynn Smith.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
CHAPTER 27: One-Way Analysis of Variance: Comparing Several Means
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 14-1 Chapter 14 Introduction to Multiple Regression Statistics for Managers using Microsoft.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 10 Comparing Two Groups Section 10.1 Categorical Response: Comparing Two Proportions.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 11 Analyzing the Association Between Categorical Variables Section 11.2 Testing Categorical.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Lecture Slides Elementary Statistics Eleventh Edition and the Triola.
McGraw-Hill, Bluman, 7th ed., Chapter 12
Analysis of Variance STAT E-150 Statistical Methods.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Overview and One-Way ANOVA.
Copyright © 2014, 2011 Pearson Education, Inc. 1 Chapter 26 Analysis of Variance.
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Lecture Slides Elementary Statistics Tenth Edition and the.
Chapter 13 Comparing Groups: Analysis of Variance Methods
Two-Sample Hypothesis Testing
Lecture Slides Elementary Statistics Twelfth Edition
Comparing Three or More Means
Analyzing the Association Between Categorical Variables
STATISTICS INFORMED DECISIONS USING DATA
Presentation transcript:

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.1 One-Way ANOVA: Comparing Several Means

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 3 Analysis of Variance The analysis of variance method compares means of several groups.  Let g denote the number of groups.  Each group has a corresponding population of subjects.  The means of the response variable for the g populations are denoted by.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 4 Hypotheses and Assumptions for the ANOVA Test Comparing Means The analysis of variance is a significance test of the null hypothesis of equal population means:  The alternative hypothesis is:  : at least two of the population means are unequal.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 5 The assumptions for the ANOVA test comparing population means are as follows:  The population distributions of the response variable for the g groups are normal with the same standard deviation for each group.  Randomization (depends on data collection method):  In a survey sample, independent random samples are selected from each of the g populations.  For an experiment, subjects are randomly assigned separately to the g groups. Hypotheses and Assumptions for the ANOVA Test Comparing Means

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 6 Example: Tolerance of Being on Hold? An airline has a toll-free telephone number for reservations. Often the call volume is heavy, and callers are placed on hold until a reservation agent is free to answer. The airline hopes a caller remains on hold until the call is answered, so as not to lose a potential customer.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 7 The airline recently conducted a randomized experiment to analyze whether callers would remain on hold longer, on the average, if they heard:  An advertisement about the airline and its current promotion  Muzak (“elevator music”)  Classical music Example: Tolerance of Being on Hold?

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 8 The company randomly selected one out of every 1000 calls in a week. For each call, they randomly selected one of the three recorded messages. They measured the number of minutes that the caller stayed on hold before hanging up (these calls were purposely not answered). Example: Tolerance of Being on Hold?

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 9 Table 14.1 Telephone Holding Times by Type of Recorded Message. Each observation is the number of minutes a caller remained on hold before hanging up, rounded to the nearest minute. Example: Tolerance of Being on Hold?

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 10 Denote the holding time means for the populations that these three random samples represent by:  = mean for the advertisement  = mean for the Muzak  = mean for the classical music Example: Tolerance of Being on Hold?

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 11 The hypotheses for the ANOVA test are:   : at least two of the population means are different Example: Tolerance of Being on Hold?

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 12 Here is a display of the sample means: Figure 14.1 Sample Means of Telephone Holding Times for Callers Who Hear One of Three Recorded Messages. Question: Since the sample means are quite different, can we conclude that the population means differ? Example: Tolerance of Being on Hold?

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 13 As you can see from the output on the previous page, the sample means are quite different. But even if the population means are equal, we expect the sample means to differ some because of sampling variability. This alone is not sufficient evidence to enable us to reject. Example: Tolerance of Being on Hold?

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 14 Variability Between Groups and Within Groups Is the Key to Significance The ANOVA method is used to compare population means. It is called analysis of variance because it uses evidence about two types of variability.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 15 Two examples of data sets with equal means but unequal variability: Figure 14.2 Data from Table 14.1 in Figure 14.2a and Hypothetical Data in Figure 14.2b That Have the Same Means but Less Variability Within Groups Variability Between Groups and Within Groups Is the Key to Significance

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 16 Which case do you think gives stronger evidence against ? What is the difference between the data in these two cases? Variability Between Groups and Within Groups Is the Key to Significance

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 17 In both cases the variability between pairs of means is the same. In ‘Case b’ the variability within each sample is much smaller than in ‘Case a.’ The fact that ‘Case b’ has less variability within each sample gives stronger evidence against. Variability Between Groups and Within Groups Is the Key to Significance

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 18 ANOVA F-Test Statistic The analysis of variance (ANOVA) F-test statistic is: The larger the variability between groups relative to the variability within groups, the larger the F test statistic tends to be.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 19 The test statistic for comparing means has the F-distribution. The larger the F-test statistic value, the stronger the evidence against. ANOVA F-Test Statistic

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 20 SUMMARY: ANOVA F-test for Comparing Population Means of Several Groups 1. Assumptions:  Independent random samples  Normal population distributions with equal standard deviations 2. Hypotheses:   : at least two of the population means are different

Copyright © 2013, 2009, and 2007, Pearson Education, Inc Test statistic:  F- sampling distribution has, (total sample size – no. of groups) SUMMARY: ANOVA F-test for Comparing Population Means of Several Groups

Copyright © 2013, 2009, and 2007, Pearson Education, Inc P-value: Right-tail probability above the observed F- value 5. Conclusion: If decision is needed, reject if P-value significance level (such as 0.05) SUMMARY: ANOVA F-test for Comparing Population Means of Several Groups

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 23 The Variance Estimates and the ANOVA Table Let denote the standard deviation for each of the g population distributions  One assumption for the ANOVA F-test is that each population has the same standard deviation,.  The F-test statistic is the ratio of two estimates of, the population variance for each group.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 24  The estimate of in the denominator of the F-test statistic uses the variability within each group.  The estimate of in the numerator of the F-test statistic uses the variability between each sample mean and the overall mean for all the data. The Variance Estimates and the ANOVA Table

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 25  Computer software displays the two estimates of in the ANOVA table similar to tables displayed in regression.  The MS column contains the two estimates, which are called mean squares.  The ratio of the two mean squares is the F- test statistic.  This F- statistic has a P-value. The Variance Estimates and the ANOVA Table

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 26 Example: Telephone Holding Times This example is a continuation of a previous example in which an airline conducted a randomized experiment to analyze whether callers would remain on hold longer, on the average, if they heard:  An advertisement about the airline and its current promotion  Muzak (“elevator music”)  Classical music

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 27 Denote the holding time means for the populations that these three random samples represent by:  = mean for the advertisement  = mean for the Muzak  = mean for the classical music Example: Telephone Holding Times

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 28 The hypotheses for the ANOVA test are:   : at least two of the population means are different Example: Telephone Holding Times

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 29 Table 14.2 ANOVA Table for F Test Using Data From Table 14.1 Example: Telephone Holding Times

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 30 Since P-value < 0.05, there is sufficient evidence to reject. We conclude that a difference exists among the three types of messages in the population mean amount of time that customers are willing to remain on hold. Example: Telephone Holding Times

Copyright © 2013, 2009, and 2007, Pearson Education, Inc Population distributions are normal  Moderate violations of the normality assumption are not serious. 2.These distributions have the same standard deviation  Moderate violations are also not serious. 3.The data resulted from randomization. Assumptions and the Effects of Violating Them

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 32 You can construct box plots or dot plots for the sample data distributions to check for extreme violations of normality. Misleading results may occur with the F-test if the distributions are highly skewed and the sample size N is small. Assumptions and the Effects of Violating Them

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 33  Misleading results may also occur with the F-test if there are relatively large differences among the standard deviations (the largest sample standard deviation being more than double the smallest one).  The ANOVA methods presented here are for independent samples. For dependent methods, other techniques must be used. Assumptions and the Effects of Violating Them

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 34 Using One F Test or Several t Tests to Compare the Means Why Not Use Multiple t-tests?  When there are several groups, using the F test instead of multiple t tests allows us to control the probability of a type I error.  If separate t tests are used, the significance level applies to each individual comparison, not the overall type I error rate for all the comparisons.  However, the F test does not tell us which groups differ or how different they are.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.2 Estimating Differences in Groups for a Single Factor

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 36 Confidence Intervals Comparing Pairs of Means Follow Up to an ANOVA F-Test:  When an analysis of variance F-test has a small P-value, the test does not specify which means are different or how different they are.  We can estimate differences between population means with confidence intervals.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 37 For two groups i and j, with sample means and having sample sizes n i and n j, the 95% confidence interval for is: The t-score has total sample size - # groups SUMMARY: Confidence Interval Comparing Means

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 38 Confidence Intervals Comparing Pairs of Means In the context of follow-up analyses after the ANOVA F test by forming this confidence interval to compare a pair of means, some software (such as MINITAB) refers to this method of comparing means as the Fisher method. When the confidence interval does not contain 0, we can infer that the population means are different. The interval shows just how different the means may be.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 39 A recent GSS study asked: “About how many good friends do you have?” The study also asked each respondent to indicate whether they were ‘very happy,’ ‘pretty happy,’ or ‘not too happy’. Example: Number of Good Friends and Happiness

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 40 Let the response variable y = number of good friends Let the categorical explanatory variable x = happiness level Example: Number of Good Friends and Happiness

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 41 Table 14.3 Summary of ANOVA for Comparing Mean Number of Good Friends for Three Happiness Categories. The analysis is based on GSS data. Example: Number of Good Friends and Happiness

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 42 Construct a 95% CI to compare the population mean number of good friends for the three pairs of happiness categories—very happy with pretty happy, very happy with not too happy, and pretty happy with not too happy. 95% CI formula: Example: Number of Good Friends and Happiness

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 43 First, use the output to find s:  df=828  Use software or a table to find the t-value of Example: Number of Good Friends and Happiness

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 44 For comparing the very happy and pretty happy categories, the confidence interval for is Since the CI contains only positive numbers, this suggests that, on average, people who are very happy have more good friends than people who are pretty happy. Example: Number of Good Friends and Happiness

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 45 The Effects of Violating Assumptions The t confidence intervals have the same assumptions as the ANOVA F test: 1.normal population distributions with 2.identical standard deviations 3.data obtained from randomization When the sample sizes are large and the ratio of the largest standard deviation to the smallest is less than 2, these procedures are robust to violations of these assumptions. If the ratio of the largest standard deviation to the smallest exceeds 2, use the confidence interval formulas that use separate standard deviations for the groups.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 46 Controlling Overall Confidence with Many Confidence Intervals The confidence interval method just discussed is mainly used when g is small or when only a few comparisons are of main interest. The confidence level of 0.95 applies to any particular confidence interval that we construct.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 47 How can we construct the intervals so that the 95% confidence extends to the entire set of intervals rather than to each single interval? Methods that control the probability that all confidence intervals will contain the true differences in means are called multiple comparison methods. For these methods, all intervals are designed to contain the true parameters simultaneously with an overall fixed probability. Controlling Overall Confidence with Many Confidence Intervals

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 48  The method that we will use is called the Tukey method.  It is designed to give overall confidence level very close to the desired value (such as 0.95).  This method is available in most software packages. Controlling Overall Confidence with Many Confidence Intervals

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 49 Example: Number of Good Friends Table 14.4 Multiple Comparisons of Mean Good Friends for Three Happiness Categories. An asterisk * indicates a significant difference, with the confidence interval not containing 0.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 50 ANOVA and Regression ANOVA can be presented as a special case of multiple regression by using indicator variables to represent the factors. For example, with 3 groups we need 2 indicator variables to indicate group membership: The first indicator variable is  x1 = 1 for observations from the first group, = 0 otherwise

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 51 The second indicator variable is  for observations from the second group  otherwise The indicator variables identify the group to which an observation belongs as follows: ANOVA and Regression

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 52 The multiple regression equation for the mean of y is Table 14.5 Interpretation of Coefficients of Indicator Variables in Regression Model The indicator variables represent a categorical predictor with three categories specifying three groups. ANOVA and Regression

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 53 Using Regression for the ANOVA Comparison of Means For three groups, the null hypothesis for the ANOVA F test is   If is true, then and In the Multiple Regression model:  with and Thus, ANOVA hypothesis is equivalent to in the regression model.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.3 Two-Way ANOVA

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 55 Type of ANOVA One-way ANOVA is a bivariate method:  It has a quantitative response variable  It has one categorical explanatory variable Two-way ANOVA is a multivariate method:  It has a quantitative response variable  It has two categorical explanatory variables

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 56 Example: Amounts of Fertilizer and Manure A recent study at Iowa State University:  A field was portioned into 20 equal-size plots.  Each plot was planted with the same amount of corn seed.  The goal was to study how the yield of corn later harvested depended on the levels of use of nitrogen-based fertilizer and manure.  Each factor (fertilizer and manure) was measured in a binary manner.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 57 There are four treatments you can compare with this experiment found by cross-classifying the two binary factors: fertilizer level and manure level. Table 14.7 Four Groups for Comparing Mean Corn Yield These result from the two-way cross classification of fertilizer level with manure level. Example: Amounts of Fertilizer and Manure

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 58 Inference about Effects in Two-Way ANOVA In two-way ANOVA, a null hypothesis states that the population means are the same in each category of one factor, at each fixed level of the other factor. We could test: : Mean corn yield is equal for plots at the low and high levels of fertilizer, for each fixed level of manure. Example: Amounts of Fertilizer and Manure

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 59 We could also test: : Mean corn yield is equal for plots at the low and high levels of manure, for each fixed level of fertilizer. The effect of individual factors tested with the two null hypotheses (the previous two pages) are called the main effects. Example: Amounts of Fertilizer and Manure

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 60 Assumptions for the Two-way ANOVA F-test  The population distribution for each group is normal.  The population standard deviations are identical.  The data result from a random sample or randomized experiment.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 61 SUMMARY: F-test Statistics in Two-Way ANOVA For testing the main effect for a factor, the test statistic is the ratio of mean squares:  The MS for the factor is a variance estimate based on between-groups variation for that factor.  The MS error is a within-groups variance estimate that is always unbiased.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 62  When the null hypothesis of equal population means for the factor is true, the F-test statistic values tend to fluctuate around 1.  When it is false, they tend to be larger.  The P-value is the right-tail probability above the observed F-value. SUMMARY: F-test Statistics in Two-Way ANOVA

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 63 Example: Corn Yield Data and sample statistics for each group: Table 14.9 Corn Yield by Fertilizer Level and Manure Level

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 64 Output from Two-way ANOVA: Table Two-Way ANOVA for Corn Yield Data in Table 14.9 Example: Corn Yield

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 65 First consider the hypothesis: : Mean corn yield is equal for plots at the low and high levels of fertilizer, for each fixed level of manure. From the output, you can obtain the F-test statistic of 6.33 with its corresponding P-value of The small P-value indicates strong evidence that the mean corn yield depends on fertilizer level. Example: Corn Yield

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 66 Next consider the hypothesis: : Mean corn yield is equal for plots at the low and high levels of manure, for each fixed level of fertilizer. From the output, you can obtain the F-test statistic of 6.88 with its corresponding P-value of The small P-value indicates strong evidence that the mean corn yield depends on manure level. Example: Corn Yield

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 67 Exploring Interaction between Factors in Two-Way ANOVA No interaction between two factors means that the effect of either factor on the response variable is the same at each category of the other factor.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 68 Figure 14.5 Mean Corn Yield, by Fertilizer and Manure Levels, Showing No Interaction. Exploring Interaction between Factors in Two-Way ANOVA

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 69 A graph showing interaction: Figure 14.6 Mean Corn Yield, by Fertilizer and Manure Levels, Displaying Interaction. Exploring Interaction between Factors in Two-Way ANOVA

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 70 Testing for Interaction In conducting a two-way ANOVA, before testing the main effects, it is customary to test a third null hypothesis stating that their is no interaction between the factors in their effects on the response.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 71 The test statistic providing the sample evidence of interaction is: When is false, the F-statistic tends to be large. Testing for Interaction

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 72 Example: Corn Yield Data ANOVA table for a model that allows interaction: Table Two-Way ANOVA of Mean Corn Yield by Fertilizer Level and Manure Level, Allowing Interaction

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 73 The test statistic for : no interaction is: F = (MS for interaction)/(MS error) = 3.04 / 2.78 = 1.10 ANOVA table reports corresponding P-value of  There is not much evidence of interaction.  We would not reject at the usual significance levels, such as Example: Corn Yield Data

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 74 Check Interaction Before Main Effects In practice, in two-way ANOVA, you should first test the hypothesis of no interaction. It is not meaningful to test the main effects hypotheses when there is interaction.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 75 If the evidence of interaction is not strong (that is, if the P-value is not small), then test the main effects hypotheses and/or construct confidence intervals for those effects. Check Interaction Before Main Effects

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 76 If important evidence of interaction exists, plot and compare the cell means for a factor separately at each category of the other factor. Check Interaction Before Main Effects

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 77 Why Not Instead Perform Two Separate One-Way ANOVAs? When you have two factors, you could perform two separate One-Way ANOVAs rather than a Two-Way ANOVA but  you learn more with a Two-Way ANOVA -it indicates whether there is interaction.  more cost effective to study the variables together rather than running two separate experiments.  the residual variability tends to decrease so we get better predictions, larger test statistics and hence greater power for rejecting false null hypotheses.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 78 Factorial ANOVA The methods of two-way ANOVA can be extended to the analysis of several factors. A multifactor ANOVA with observations from all combinations of the factors is called factorial ANOVA, e.g., with three factors - three-way ANOVA considers main effects for all three factors as well as possible interactions.

Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 79 Use Regression With Categorical and Quantitative Predictors In practice, when you have several predictors, both categorical and quantitative, it is sensible to build a multiple regression model containing both types of predictors.