Download presentation
Presentation is loading. Please wait.
1
© 2011 Pearson Education, Inc
2
Statistics for Business and Economics
Chapter 8 Design of Experiments and Analysis of Variance © 2011 Pearson Education, Inc
3
© 2011 Pearson Education, Inc
Contents 8.1 Elements of a Designed Experiment 8.2 The Completely Randomized Design: Single Factor 8.3 Multiple Comparisons of Means 8.4 The Randomized Block Design 8.5 Factorial Experiments: Two Factors As a result of this class, you will be able to ... © 2011 Pearson Education, Inc
4
© 2011 Pearson Education, Inc
Learning Objectives Discuss critical elements in the design of a sampling experiment Learn how to set up three experimental designs for comparing more than two population means: completely randomized, randomized block, and factorial design Show how to analyze data collected from a designed experiment using a technique called an analysis of variance (ANOVA) As a result of this class, you will be able to ... © 2011 Pearson Education, Inc
5
Elements of a Designed Experiment
8.1 Elements of a Designed Experiment :1, 1, 3 © 2011 Pearson Education, Inc
6
© 2011 Pearson Education, Inc
Response Variable The response variable is the variable of interest to be measured in the experiment. We also refer to the response as the dependent variable. © 2011 Pearson Education, Inc
7
© 2011 Pearson Education, Inc
Factors Factors are those variables whose effect on the response is of interest to the experimenter. Quantitative factors are measured on a numerical scale, whereas qualitative factors are those that are not (naturally) measured on a numerical scale. Factors are also referred to as independent variables. © 2011 Pearson Education, Inc
8
© 2011 Pearson Education, Inc
Factor Levels Factor levels are the values of the factor used in the experiment. Treatments The treatments of an experiment are the factor-level combinations used. © 2011 Pearson Education, Inc
9
Designed and Observational Experiment
A designed experiment is one for which the analyst controls the specification of the treatments and the method of assigning the experimental units to each treatment. An observational experiment is one for which the analyst simply observes the treatments and the response on a sample of experimental units. © 2011 Pearson Education, Inc
10
© 2011 Pearson Education, Inc
Experimental Unit An experimental unit is the object on which the response and factors are observed or measured. © 2011 Pearson Education, Inc
11
© 2011 Pearson Education, Inc
Experiment Investigator controls one or more independent variables Called treatment variables or factors Contain two or more levels (subcategories) Observes effect on dependent variable Response to levels of independent variable Experimental design: plan used to test hypotheses © 2011 Pearson Education, Inc
12
Examples of Experiments
Thirty stores are randomly assigned 1 of 4 (levels) store displays (independent variable) to see the effect on sales (dependent variable). Two hundred consumers are randomly assigned 1 of 3 (levels) brands of juice (independent variable) to study reaction (dependent variable). © 2011 Pearson Education, Inc
13
The Completely Randomized Design: Single Factor
8.2 The Completely Randomized Design: Single Factor :1, 1, 3 © 2011 Pearson Education, Inc
14
Completely Randomized Design
A completely randomized design is a design for which independent random samples of experimental units are selected for each treatment. © 2011 Pearson Education, Inc
15
Completely Randomized Design
Experimental units (subjects) are assigned randomly to treatments Subjects are assumed homogeneous One factor or independent variable Two or more treatment levels or classifications Analyzed by one-way ANOVA © 2011 Pearson Education, Inc
16
Randomized Design Example
Factor (Training Method) Factor levels Level 1 Level 2 Level 3 (Treatments) Are the mean training times the same for 3 different methods? 9 subjects 3 methods (factor levels) Experimental units 21 hrs. 17 hrs. 31 hrs. Dependent variable 27 hrs. 25 hrs. 28 hrs. (Response) 29 hrs. 20 hrs. 22 hrs. © 2011 Pearson Education, Inc 78
17
© 2011 Pearson Education, Inc
ANOVA F-Test Tests the equality of two or more (k) population means Variables One nominal scaled independent variable Two or more (k) treatment levels or classifications One interval or ratio scaled dependent variable Used to analyze completely randomized experimental designs Note: There is one dependent variable in the ANOVA model. MANOVA has more than one dependent variable. Ask, what are nominal & interval scales? © 2011 Pearson Education, Inc
18
© 2011 Pearson Education, Inc
ANOVA F-Test to Compare k Treatment Means: Completely Randomized Design H0: µ1 = µ2 = … = µk Ha: At least two treatment means differ Rejection region: F > F, where F is based on (k – 1) numerator degrees of freedom (associated with MST) and (n – k) denominator degrees of freedom (associated with MSE). Note: There is one dependent variable in the ANOVA model. MANOVA has more than one dependent variable. Ask, what are nominal & interval scales? © 2011 Pearson Education, Inc
19
© 2011 Pearson Education, Inc
Conditions Required for a Valid ANOVA F-test: Completely Randomized Design 1. The samples are randomly selected in an independent manner from the k treatment populations. (This can be accomplished by randomly assigning the experimental units to the treatments.) 2. All k sampled populations have distributions that are approximately normal. 3. The k population variances are equal (i.e., © 2011 Pearson Education, Inc
20
ANOVA F-Test Hypotheses
H0: 1 = 2 = 3 = ... = k All population means are equal No treatment effect Ha: Not All i Are Equal At least 2 pop. means are different Treatment effect 1 2 ... k is Wrong x f(x) 1 = 2 3 1 2 3 x f(x) = © 2011 Pearson Education, Inc
21
Variances AMONG differ Variances WITHIN differ
Why Variances? Same treatment variation Different random variation Pop 1 Pop 2 Pop 3 Pop 4 Pop 6 Pop 5 Variances AMONG differ B Different treatment variation Same random variation Pop 1 Pop 2 Pop 3 Pop 4 Pop 6 Pop 5 Variances WITHIN differ A In all cases, the population means are DIFFERENT! Should reject Ho. Panel A: Same treatment effect - means are equal. Different random variation - standard deviations are different. As variances WITHIN get bigger, we are more likely to conclude equal means. In Populations 4, 5, & 6, it is possible to draw a sample and falsely conclude population means are equal. Panel B: Different treatment effect - means are different. Same random variation - standard deviations are equal. As variances AMONG get bigger, we are more likely to conclude population means are different. In Populations 4,5, & 6, it is possible to draw a sample and falsely conclude population means are equal. Possible to conclude means are equal! © 2011 Pearson Education, Inc 82
22
© 2011 Pearson Education, Inc
ANOVA Basic Idea Compares two types of variation to test equality of means Comparison basis is ratio of variances If treatment variation is significantly greater than random variation then means are not equal Variation measures are obtained by ‘partitioning’ total variation © 2011 Pearson Education, Inc
23
© 2011 Pearson Education, Inc
What Do You Do When the Assumptions Are Not Satisfied for an ANOVA for a Completely Randomized Design? Answer: Use a nonparametric statistical method such as the Kruskal-Wallis H-test. © 2011 Pearson Education, Inc
24
Steps for Conducting an ANOVA for a Completely Randomized Design
1. Be sure the design is truly completely randomized, with independent random samples for each treatment. 2. Check the assumptions of normality and equal variances. © 2011 Pearson Education, Inc
25
Steps for Conducting an ANOVA for a Completely Randomized Design
3. Create an ANOVA summary table that specifies the variability attributable to treatments and error, making sure that it leads to the calculation of the F-statistic for testing the null hypothesis that the treatment means are equal in the population. Use a statistical software program to obtain the numerical results. If no such package is available, use the calculation formulas in Appendix C. © 2011 Pearson Education, Inc
26
Steps for Conducting an ANOVA for a Completely Randomized Design
4. If the F-test leads to the conclusion that the means differ, a. Conduct a multiple comparisons procedure for as many of the pairs of means as you wish to compare. Use the results to summarize the statistically significant differences among the treatment means. b. If desired, form confidence intervals for one or more individual treatment means. © 2011 Pearson Education, Inc
27
Steps for Conducting an ANOVA for a Completely Randomized Design
5. If the F-test leads to the nonrejection of the null hypothesis that the treatment means are equal, consider the following possibilities: a. The treatment means are equal–that is, the null hypothesis is true. © 2011 Pearson Education, Inc
28
Steps for Conducting an ANOVA for a Completely Randomized Design
b. The treatment means really differ, but other important factors affecting the response are not accounted for by the completely randomized design. These factors inflate the sampling variability, as measured by MSE, resulting in smaller values of the F-statistic. Either increase the sample size for each treatment or use a different experimental design that accounts for the other factors affecting the response. © 2011 Pearson Education, Inc
29
Steps for Conducting an ANOVA for a Completely Randomized Design
Note: Be careful not to automatically conclude that the treatment means are equal because the possibility of a Type II error must be considered if you accept H0. © 2011 Pearson Education, Inc
30
ANOVA Partitions Total Variation
Variation due to treatment Total variation Variation due to random sampling Variation due to Random Sampling are due to Individual Differences Within Groups. Sum of Squares Among Sum of Squares Between Sum of Squares Treatment Among Groups Variation Sum of Squares Within Sum of Squares Error Within Groups Variation © 2011 Pearson Education, Inc 85
31
© 2011 Pearson Education, Inc
Total Variation Response, x x Group 1 Group 2 Group 3 © 2011 Pearson Education, Inc
32
© 2011 Pearson Education, Inc
Treatment Variation Response, x x3 x x2 x1 Group 1 Group 2 Group 3 © 2011 Pearson Education, Inc
33
Random (Error) Variation
Response, x x3 x2 x1 Group 1 Group 2 Group 3 © 2011 Pearson Education, Inc
34
ANOVA F-Test Test Statistic
F = MST / MSE MST is Mean Square for Treatment MSE is Mean Square for Error Degrees of Freedom 1 = k – 1 numerator degrees of freedom 2 = n – k denominator degrees of freedom k = Number of groups n = Total sample size © 2011 Pearson Education, Inc
35
Mean Square (Variance)
ANOVA Summary Table Degrees of Freedom Mean Square (Variance) Source of Variation Sum of Squares F Treatment k – 1 SST MST = SST/(k – 1) MST MSE n = sum of sample sizes of all populations. k = number of factor levels All values are positive. Why? (squared terms) Degrees of Freedom & Sum of Squares are additive; Mean Square is NOT. Error n – k SSE MSE = SSE/(n – k) Total n – 1 SS(Total) = SST+SSE © 2011 Pearson Education, Inc
36
ANOVA F-Test Critical Value
If means are equal, F = MST / MSE 1. Only reject large F! Reject H Do Not Reject H F F(α; k – 1, n – k) Always One-Tail! © T/Maker Co. © 2011 Pearson Education, Inc
37
© 2011 Pearson Education, Inc
ANOVA F-Test Example As production manager, you want to see if three filling machines have different mean filling times. You assign 15 similarly trained and experienced workers, 5 per machine, to the machines. At the .05 level of significance, is there a difference in mean filling times? Mach1 Mach2 Mach © 2011 Pearson Education, Inc
38
© 2011 Pearson Education, Inc
ANOVA F-Test Solution H0: Ha: = 1 = 2 = Critical Value(s): 1 = 2 = 3 Not all equal .05 F 3.89 = .05 © 2011 Pearson Education, Inc
39
Summary Table Solution
From Computer Degrees of Freedom Mean Square (Variance) Source of Variation Sum of Squares F Treatment (Machines) 3 – 1 = 2 25.60 Error 15 – 3 = 12 .9211 Total 15 – 1 = 14 © 2011 Pearson Education, Inc
40
© 2011 Pearson Education, Inc
ANOVA F-Test Solution H0: Ha: = 1 = 2 = Critical Value(s): 1 = 2 = 3 Not all equal Test Statistic: Decision: Conclusion: .05 F MST MSE 23 5820 9211 25.6 . F 3.89 = .05 Reject at = .05 There is evidence population means are different © 2011 Pearson Education, Inc
41
ANOVA F-Test Thinking Challenge
You’re a trainer for Microsoft Corp. Is there a difference in mean learning times of 12 people using 4 different training methods ( =.05)? M1 M2 M3 M Use the following table. You assign randomly 3 people to each method, making sure that they are similar in intelligence etc. © T/Maker Co. © 2011 Pearson Education, Inc
42
Summary Table (Partially Completed)
Degrees of Freedom Mean Square (Variance) Source of Variation Sum of Squares F Treatment 348 (Methods) Error 80 Total © 2011 Pearson Education, Inc
43
© 2011 Pearson Education, Inc
ANOVA F-Test Solution H0: Ha: = 1 = 2 = Critical Value(s): 1 = 2 = 3 = 4 Not all equal Test Statistic: Decision: Conclusion: .05 F 4.07 = .05 © 2011 Pearson Education, Inc
44
Summary Table Solution
Degrees of Freedom Mean Square (Variance) Source of Variation Sum of Squares F Treatment (Methods) 4 – 1 = 3 348 116 11.6 Error 12 – 4 = 8 80 10 Total 12 – 1 = 11 428 © 2011 Pearson Education, Inc
45
© 2011 Pearson Education, Inc
ANOVA F-Test Solution H0: Ha: = 1 = 2 = Critical Value(s): 1 = 2 = 3 = 4 Not all equal Test Statistic: Decision: Conclusion: F MST MSE 116 10 11 6 . .05 F 4.07 = .05 Reject at = .05 There is evidence population means are different © 2011 Pearson Education, Inc
46
Multiple Comparisons of Means
8.3 Multiple Comparisons of Means :1, 1, 3 © 2011 Pearson Education, Inc
47
Determining the Number of Pairwise Comparisons of Treatment Means
In general, if there are k treatment means, there are pairs of means that can be compared. © 2011 Pearson Education, Inc
48
© 2011 Pearson Education, Inc
Tukey Methods Tukey methods. For each of these procedures, the risk of making a Type I error applies to the comparisons of the treatment means in the experiment; thus, the value of selected is called an experimentwise error rate (in contrast to a comparisonwise error rate). © 2011 Pearson Education, Inc
49
© 2011 Pearson Education, Inc
Bonferroni Method The Bonferroni method (see Miller, 1981), like the Tukey procedure, can be applied when pairwise comparisons are of interest; however, Bonferroni’s method does not require equal sample sizes. © 2011 Pearson Education, Inc
50
© 2011 Pearson Education, Inc
Scheffé Method Scheffé (1953) developed a more general procedure for comparing all possible linear combinations of treatment means (called contrasts). Consequently, when making pairwise comparisons, the confidence intervals produced by Scheffé’s method will generally be wider than the Tukey or Bonferroni confidence intervals. © 2011 Pearson Education, Inc
51
© 2011 Pearson Education, Inc
Tukey Procedure Tells which population means are significantly different Example: μ1 = μ 2 ≠ μ 3 Post hoc procedure Done after rejection of equal means in ANOVA Output from many statistical computer programs x f(x) m 1 = 2 3 2 Groupings © 2011 Pearson Education, Inc 105
52
© 2011 Pearson Education, Inc
Guidelines © 2011 Pearson Education, Inc 105
53
The Randomized Block Design
8.4 The Randomized Block Design :1, 1, 3 © 2011 Pearson Education, Inc
54
Randomized Block Design
The randomized block design consists of a two- step procedure: 1. Matched sets of experimental units, called blocks, are formed, each block consisting of k experimental units (where k is the number of treatments). The b blocks should consist of experimental units that are as similar as possible. 2. One experimental unit from each block is randomly assigned to each treatment, resulting in a total of n = bk responses. © 2011 Pearson Education, Inc
55
Randomized Block Design
Reduces sampling variability (MSE) Matched sets of experimental units (blocks) One experimental unit from each block is randomly assigned to each treatment © 2011 Pearson Education, Inc
56
Randomized Block Design Total Variation Partitioning
© 2011 Pearson Education, Inc
57
ANOVA F-Test to Compare k Treatment Means: Randomized Block Design
H0: µ1 = µ2 = … = µk Ha: At least two treatment means differ Rejection region: F > F, where F is based on (k – 1) numerator degrees of freedom and (n – b – k +1) denominator degrees of freedom. Note: There is one dependent variable in the ANOVA model. MANOVA has more than one dependent variable. Ask, what are nominal & interval scales? © 2011 Pearson Education, Inc
58
Conditions Required for a Valid ANOVA F-test: Randomized Block Design
The b blocks are randomly selected, and all k treatments are applied (in random order) to each block. The distributions of observations corresponding to all bk block-treatment combinations are approximately normal. All bk block-treatment distributions have equal variances. © 2011 Pearson Education, Inc
59
Randomized Block Design F-Test Test Statistic
F = MST / MSE MST is Mean Square for Treatment MSE is Mean Square for Error Degrees of Freedom 1 = k – 1 (numerator) 2 = n – k – b + 1 (denominator) k = Number of groups n = Total sample size b = Number of blocks © 2011 Pearson Education, Inc
60
Randomized Block Design Example
A production manager wants to see if three assembly methods have different mean assembly times (in minutes). Five employees were selected at random and assigned to use each assembly method. At the .05 level of significance, is there a difference in mean assembly times? Employee Method 1 Method 2 Method 3 © 2011 Pearson Education, Inc
61
Random Block Design F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): 1 = 2 = 3 Not all equal .05 F 4.46 = .05 © 2011 Pearson Education, Inc
62
Summary Table Solution
Degrees of Freedom Mean Square (Variance) Source of Variation Sum of Squares F Treatment (Methods) 3 – 1 = 2 5.43 2.71 12.9 Block (Employee) 5 – 1 = 4 10.69 2.67 12.7 Error 15 – 3 – = 8 1.68 .21 Total 15 – 1 = 14 17.8 © 2011 Pearson Education, Inc
63
Random Block Design F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): 1 = 2 = 3 Not all equal Test Statistic: Decision: Conclusion: F MST MSE 2.71 .21 12.9 .05 F 4.46 = .05 Reject at = .05 There is evidence population means are different © 2011 Pearson Education, Inc
64
Steps for Conducting an ANOVA for a Randomized Block Design
1. Be sure the design consists of blocks (preferably, blocks of homogeneous experimental units) and that each treatment is randomly assigned to one experimental unit in each block. 2. If possible, check the assumptions of normality and equal variances for all block-treatment combinations. [Note: This may be difficult to do because the design will likely have only one observation for each block-treatment combination.] © 2011 Pearson Education, Inc
65
Steps for Conducting an ANOVA for a Randomized Block Design
3. Create an ANOVA summary table that specifies the variability attributable to Treatments, Blocks, and Error, which leads to the calculation of the F-statistic to test the null hypothesis that the treatment means are equal in the population. Use a statistical software package or the calculation formulas in Appendix C to obtain the necessary numerical ingredients. © 2011 Pearson Education, Inc
66
Steps for Conducting an ANOVA for a Randomized Block Design
4. If the F-test leads to the conclusion that the means differ, use the Bonferroni, Tukey, or similar procedure to conduct multiple comparisons of as many of the pairs of means as you wish. Use the results to summarize the statistically significant differences among the treatment means. Remember that, in general, the randomized block design cannot be used to form confidence intervals for individual treatment means. © 2011 Pearson Education, Inc
67
Steps for Conducting an ANOVA for a Randomized Block Design
5. If the F-test leads to the nonrejection of the null hypothesis that the treatment means are equal, consider the following possibilities: a. The treatment means are equal–that is, the null hypothesis is true. © 2011 Pearson Education, Inc
68
Steps for Conducting an ANOVA for a Randomized Block Design
b. The treatment means really differ, but other important factors affecting the response are not accounted for by the randomized block design. These factors inflate the sampling variability, as measured by MSE, resulting in smaller values of the F-statistic. Either increase the sample size for each treatment or conduct an experiment that accounts for the other factors affecting the response. Do not automatically reach the former conclusion because the possibility of a Type II error must be considered if you accept H0. © 2011 Pearson Education, Inc
69
Steps for Conducting an ANOVA for a Randomized Block Design
6. If desired, conduct the F-test of the null hypothesis that the block means are equal. Rejection of this hypothesis lends statistical support to using the randomized block design. © 2011 Pearson Education, Inc
70
Steps for Conducting an ANOVA for a Randomized Block Design
Note: It is often difficult to check whether the assumptions for a randomized block design are satisfied. There is usually only one observation for each block-treatment combination. When you feel these assumptions are likely to be violated, a nonparametric procedure is advisable. © 2011 Pearson Education, Inc
71
© 2011 Pearson Education, Inc
What Do You Do When the Assumptions Are Not Satisfied for an ANOVA for a Completely Randomized Design? Answer: Use a nonparametric statistical method such as the Friedman Fr test. © 2011 Pearson Education, Inc
72
Factorial Experiments: Two Factors
8.5 Factorial Experiments: Two Factors :1, 1, 3 © 2011 Pearson Education, Inc
73
© 2011 Pearson Education, Inc
Factorial Design A complete factorial experiment is one in which every factor-level combination is employed–that is, the number of treatments in the experiment equals the total number of factor-level combinations. Also referred to as a two-way classification. © 2011 Pearson Education, Inc
74
© 2011 Pearson Education, Inc
Factorial Design To determine the nature of the treatment effect, if any, on the response in a factorial experiment, we need to break the treatment variability into three components: Interaction between Factors A and B, Main Effect of Factor A, and Main Effect of Factor B. The Factor Interaction component is used to test whether the factors combine to affect the response, while the Factor Main Effect components are used to determine whether the factors separately affect the response. © 2011 Pearson Education, Inc
75
© 2011 Pearson Education, Inc
Factorial Design Experimental units (subjects) are assigned randomly to treatments Subjects are assumed homogeneous Two or more factors or independent variables Each has two or more treatments (levels) Analyzed by two-way ANOVA © 2011 Pearson Education, Inc
76
Procedure for Analysis of Two-Factor Factorial Experiment
1. Partition the Total Sum of Squares into the Treatments and Error components. Use either a statistical software package or the calculation formulas in Appendix C to accomplish the partitioning. © 2011 Pearson Education, Inc
77
Procedure for Analysis of Two-Factor Factorial Experiment
Use the F-ratio of Mean Square for Treatments to Mean Square for Error to test the null hypothesis that the treatment means are equal. a. If the test results in nonrejection of the null hypothesis, consider refining the experiment by increasing the number of replications or introducing other factors. Also consider the possibility that the response is unrelated to the two factors. b. If the test results in rejection of the null hypothesis, then proceed to step 3. © 2011 Pearson Education, Inc
78
Procedure for Analysis of Two-Factor Factorial Experiment
3. Partition the Treatments Sum of Squares into the Main Effect and Interaction Sum of Squares. Use either a statistical software package or the calculation formulas in Appendix C to accomplish the partitioning. © 2011 Pearson Education, Inc
79
Procedure for Analysis of Two-Factor Factorial Experiment
4. Test the null hypothesis that factors A and B do not interact to affect the response by computing the F-ratio of the Mean Square for Interaction to the Mean Square for Error. a. If the test results in nonrejection of the null hypothesis, proceed to step 5. b. If the test results in rejection of the null hypothesis, conclude that the two factors interact to affect the mean response. Then proceed to step 6a. © 2011 Pearson Education, Inc
80
Procedure for Analysis of Two-Factor Factorial Experiment
5. Conduct tests of two null hypotheses that the mean response is the same at each level of factor A and factor B. Compute two F-ratios by comparing the Mean Square for each Factor Main Effect to the Mean Square for Error. a. If one or both tests result in rejection of the null hypothesis, conclude that the factor affects the mean response. Proceed to step 6b. © 2011 Pearson Education, Inc
81
Procedure for Analysis of Two-Factor Factorial Experiment
b. If both tests result in nonrejection, an apparent contradiction has occurred. Although the treatment means apparently differ (step 2 test), the interaction (step 4) and main effect (step 5) tests have not supported that result. Further experimentation is advised. © 2011 Pearson Education, Inc
82
Procedure for Analysis of Two-Factor Factorial Experiment
6. Compare the means: a. If the test for interaction (step 4) is significant, use a multiple comparisons procedure to compare any or all pairs of the treatment means. b. If the test for one or both main effects (step 5) is significant, use a multiple comparisons procedure to compare the pairs of means corresponding to the levels of the significant factor(s). © 2011 Pearson Education, Inc
83
© 2011 Pearson Education, Inc
ANOVA Tests Conducted for Factorial Experiments: Completely Randomized Design, r Replicates per Treatment Test for Treatment Means H0: No difference among the ab treatment means Ha: At least two treatment means differ Rejection region: F > F, based on (ab – 1) numerator and (n – ab) denominator degrees of freedom [Note: n = abr.] Note: There is one dependent variable in the ANOVA model. MANOVA has more than one dependent variable. Ask, what are nominal & interval scales? © 2011 Pearson Education, Inc
84
© 2011 Pearson Education, Inc
ANOVA Tests Conducted for Factorial Experiments: Completely Randomized Design, r Replicates per Treatment Test for Factor Interaction H0: Factors A and B do not interact to affect the response mean Ha: Factors A and B do interact to affect the response mean Rejection region: F > F, based on (a – 1)(b – 1) numerator and (n – ab) denominator degrees of freedom Note: There is one dependent variable in the ANOVA model. MANOVA has more than one dependent variable. Ask, what are nominal & interval scales? © 2011 Pearson Education, Inc
85
© 2011 Pearson Education, Inc
ANOVA Tests Conducted for Factorial Experiments: Completely Randomized Design, r Replicates per Treatment Test for Main Effect of Factor A H0: No difference among the a mean levels of factor A Ha: At least two factor A mean levels differ Rejection region: F > F, based on (a – 1) numerator and (n – ab) denominator degrees of freedom Note: There is one dependent variable in the ANOVA model. MANOVA has more than one dependent variable. Ask, what are nominal & interval scales? © 2011 Pearson Education, Inc
86
© 2011 Pearson Education, Inc
ANOVA Tests Conducted for Factorial Experiments: Completely Randomized Design, r Replicates per Treatment Test for Main Effect of Factor B H0: No difference among the b mean levels of factor B Ha: At least two factor B mean levels differ Rejection region: F > F, based on (b – 1) numerator and (n – ab) denominator degrees of freedom Note: There is one dependent variable in the ANOVA model. MANOVA has more than one dependent variable. Ask, what are nominal & interval scales? © 2011 Pearson Education, Inc
87
Conditions Required for Valid F-tests in Factorial Experiments
1. The response distribution for each factor-level combination (treatment) is normal. 2. The response variance is constant for all treatments. 3. Random and independent samples of experimental units are associated with each treatment. © 2011 Pearson Education, Inc
88
© 2011 Pearson Education, Inc
ANOVA Data Table Factor Factor B A 1 2 ... b Observation k 1 x111 x121 ... x1b1 xijk x112 x122 ... x1b2 2 x211 x221 ... x2b1 x212 x222 ... x2b2 Level i Factor A Level j Factor B : : : : : Treatment a xa11 xa21 ... xab1 xa12 xa22 ... xab2 © 2011 Pearson Education, Inc
89
Factorial Design Example
Factor 2 (Training Method) Factor Levels Level 1 Level 2 Level 3 Level 1 15 hr. 10 hr. 22 hr. (High) Factor 1 (Motivation) 11 hr. 12 hr. 17 hr. Treatment Level 2 27 hr. 15 hr. 31 hr. (Low) 29 hr. 17 hr. 49 hr. © 2011 Pearson Education, Inc
90
Advantages of Factorial Designs
Saves time and effort e.g., Could use separate completely randomized designs for each variable Controls confounding effects by putting other variables into model Can explore interaction between variables © 2011 Pearson Education, Inc
91
© 2011 Pearson Education, Inc
Two-Way ANOVA Tests the equality of two or more population means when several independent variables are used Same results as separate one-way ANOVA on each variable No interaction can be tested Used to analyze factorial designs © 2011 Pearson Education, Inc
92
© 2011 Pearson Education, Inc
Interaction Occurs when effects of one factor vary according to levels of other factor When significant, interpretation of main effects (A and B) is complicated Can be detected In data table, pattern of cell means in one row differs from another row In graph of cell means, lines cross © 2011 Pearson Education, Inc
93
© 2011 Pearson Education, Inc
Graphs of Interaction Effects of motivation (high or low) and training method (A, B, C) on mean learning time Interaction No Interaction Average Average Response High Response High Low Low A B C A B C © 2011 Pearson Education, Inc
94
Two-Way ANOVA Total Variation Partitioning
© 2011 Pearson Education, Inc
95
Two-Way ANOVA Summary Table
Source of Variation Degrees of Freedom Sum of Squares Mean Square F A (Row) a – 1 SS(A) MS(A) MS(A) MSE B (Column) b – 1 SS(B) MS(B) MS(B) MSE AB (Interaction) (a – 1)(b – 1) SS(AB) MS(AB) MS(AB) MSE Error n – ab SSE MSE Same as other designs Total n – 1 SS(Total) © 2011 Pearson Education, Inc
96
Factorial Design Example
Human Resources wants to determine if training time is different based on motivation level and training method. Conduct the appropriate ANOVA tests. Use α = .05 for each test. Training Method Factor Levels Self– paced Classroom Computer 15 hr. 10 hr. 22 hr. Motivation High 11 hr. 12 hr. 17 hr. 27 hr. 31 hr. Low 29 hr. 49 hr. © 2011 Pearson Education, Inc
97
Treatment Means F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): The 6 treatment means are equal At least 2 differ .05 F 4.39 = .05 © 2011 Pearson Education, Inc
98
Two-Way ANOVA Summary Table
Source of Variation Degrees of Freedom Sum of Squares Mean Square F Model 5 1201.8 240.35 7.65 Error 6 188.5 31.42 Corrected Total 11 1390.3 © 2011 Pearson Education, Inc
99
Treatment Means F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): The 6 treatment means are equal At least 2 differ Test Statistic: Decision: Conclusion: .05 Reject at = .05 F 4.39 = .05 There is evidence population means are different © 2011 Pearson Education, Inc
100
Factor Interaction F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): The factors do not interact The factors interact .05 F 5.14 = .05 © 2011 Pearson Education, Inc
101
Two-Way ANOVA Summary Table
Source of Variation Degrees of Freedom Sum of Squares Mean Square F A (Row) 1 546.75 546.75 17.40 B (Column) 2 531.5 265.75 8.46 AB (Interaction) 2 123.5 61.76 1.97 Error 6 188.5 31.42 Same as other designs Total 11 SS(Total) © 2011 Pearson Education, Inc
102
Factor Interaction F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): The factors do not interact The factors interact Test Statistic: Decision: Conclusion: .05 Do not reject at = .05 F 5.14 = .05 There is no evidence the factors interact © 2011 Pearson Education, Inc
103
Main Factor A F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): No difference between motivation levels Motivation levels differ .05 F 5.99 = .05 © 2011 Pearson Education, Inc
104
Two-Way ANOVA Summary Table
Source of Variation Degrees of Freedom Sum of Squares Mean Square F A (Row) 1 546.75 546.75 17.40 B (Column) 2 531.5 265.75 8.46 AB (Interaction) 2 123.5 61.76 1.97 Error 6 188.5 31.42 Same as other designs Total 11 SS(Total) © 2011 Pearson Education, Inc
105
Main Factor A F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): No difference between motivation levels Motivation levels differ Test Statistic: Decision: Conclusion: .05 Reject at = .05 F 5.99 = .05 There is evidence that motivations levels differ © 2011 Pearson Education, Inc
106
Main Factor B F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): No difference between training methods Training methods differ .05 F 5.14 = .05 © 2011 Pearson Education, Inc
107
Two-Way ANOVA Summary Table
Source of Variation Degrees of Freedom Sum of Squares Mean Square F A (Row) 1 546.75 546.75 17.40 B (Column) 2 531.5 265.75 8.46 AB (Interaction) 2 123.5 61.76 1.97 Error 6 188.5 31.42 Same as other designs Total 11 SS(Total) © 2011 Pearson Education, Inc
108
Main Factor B F-Test Solution
H0: Ha: = 1 = 2 = Critical Value(s): No difference between training methods Training methods differ Test Statistic: Decision: Conclusion: .05 Reject at = .05 F 5.14 = .05 There is evidence training methods differ © 2011 Pearson Education, Inc
109
© 2011 Pearson Education, Inc
Key Ideas Key Elements of a Designed Experiment 1. Response (dependent) variable – quantitative 2. Factors (independent variables) – quantitative or qualitative 3. Factor levels (values of factors) – selected by the experimenter 4. Treatments – combinations of factor levels 5. Experimental units – assign treatments to experimental units and measure response for each As a result of this class, you will be able to... © 2011 Pearson Education, Inc
110
© 2011 Pearson Education, Inc
Key Ideas Key Elements of a Designed Experiment Balanced Design Sample sizes for each treatment are equal. Test for main effects in a factorial design Only appropriate if the test for factor interaction is nonsignificant. Robust method Slight to moderate departures from normality do not have impact on validity of the ANOVA results. As a result of this class, you will be able to... © 2011 Pearson Education, Inc
111
© 2011 Pearson Education, Inc
Key Ideas Conditions Required for Valid F-Test in a Completely Randomized Design 1. All k treatment populations are approximately normal. 2. As a result of this class, you will be able to... © 2011 Pearson Education, Inc
112
© 2011 Pearson Education, Inc
Key Ideas Conditions Required for Valid F-Test in a Randomized Block Design 1. All treatment-block populations are approximately normal. 2. All treatment-block populations have the same variance. As a result of this class, you will be able to... © 2011 Pearson Education, Inc
113
© 2011 Pearson Education, Inc
Key Ideas Conditions Required for Valid F-Test in a Complete Factorial Design 1. All treatment populations are approximately normal. 2. All treatment populations have the same variance. Experimentwise error rate Risk of making at least one Type I error when making multiple comparisons of means in ANOVA As a result of this class, you will be able to... © 2011 Pearson Education, Inc
114
© 2011 Pearson Education, Inc
Key Ideas Multiple Comparison of Means Methods Number of pairwise comparisons with k treatment means c = k(k – 1)/2 As a result of this class, you will be able to... © 2011 Pearson Education, Inc
115
© 2011 Pearson Education, Inc
Key Ideas Multiple Comparison of Means Methods Tukey method: 1. Balanced design 2. Pairwise comparison of means As a result of this class, you will be able to... © 2011 Pearson Education, Inc
116
© 2011 Pearson Education, Inc
Key Ideas Multiple Comparison of Means Methods Bonferroni method: 1. Either balanced or unbalanced design 2. Pairwise comparison of means As a result of this class, you will be able to... © 2011 Pearson Education, Inc
117
© 2011 Pearson Education, Inc
Key Ideas Multiple Comparison of Means Methods Scheffé method: 1. Either balanced or unbalanced design 2. General contrast of means As a result of this class, you will be able to... © 2011 Pearson Education, Inc
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.