The Completely Randomized Design (§8.3)

Slides:



Advertisements
Similar presentations
A. The Basic Principle We consider the multivariate extension of multiple linear regression – modeling the relationship between m responses Y 1,…,Y m and.
Advertisements

Latin Square Designs (§15.4)
CHAPTER 25: One-Way Analysis of Variance Comparing Several Means
Chapter 11 Analysis of Variance
Multiple regression analysis
Statistics for Managers Using Microsoft® Excel 5th Edition
Be humble in our attribute, be loving and varying in our attitude, that is the way to live in heaven.
Chapter 11 Analysis of Variance
ANalysis Of VAriance (ANOVA) Comparing > 2 means Frequently applied to experimental data Why not do multiple t-tests? If you want to test H 0 : m 1 = m.
Chapter 3 Analysis of Variance
Every achievement originates from the seed of determination. 1Random Effect.
13-1 Designing Engineering Experiments Every experiment involves a sequence of activities: Conjecture – the original hypothesis that motivates the.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 10-1 Chapter 10 Analysis of Variance Statistics for Managers Using Microsoft.
Chap 10-1 Analysis of Variance. Chap 10-2 Overview Analysis of Variance (ANOVA) F-test Tukey- Kramer test One-Way ANOVA Two-Way ANOVA Interaction Effects.
22-1 Factorial Treatments (§ 14.3, 14.4, 15.4) Completely randomized designs - One treatment factor at t levels. Randomized block design - One treatment.
One-Factor Experiments Andy Wang CIS 5930 Computer Systems Performance Analysis.
1 Experimental Statistics - week 7 Chapter 15: Factorial Models (15.5) Chapter 17: Random Effects Models.
QNT 531 Advanced Problems in Statistics and Research Methods
12-1 Chapter Twelve McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc., All Rights Reserved.
© 2002 Prentice-Hall, Inc.Chap 14-1 Introduction to Multiple Regression Model.
1 Experimental Statistics - week 6 Chapter 15: Randomized Complete Block Design (15.3) Factorial Models (15.5)
1 Experimental Statistics - week 4 Chapter 8: 1-factor ANOVA models Using SAS.
12-1 Chapter Twelve McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved.
1 Experimental Statistics - week 10 Chapter 11: Linear Regression and Correlation.
23-1 Analysis of Covariance (Chapter 16) A procedure for comparing treatment means that incorporates information on a quantitative explanatory variable,
1 Experimental Statistics - week 10 Chapter 11: Linear Regression and Correlation Note: Homework Due Thursday.
6-3 Multiple Regression Estimation of Parameters in Multiple Regression.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Chapter 10 Analysis of Variance.
Testing Multiple Means and the Analysis of Variance (§8.1, 8.2, 8.6) Situations where comparing more than two means is important. The approach to testing.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap th Lesson Analysis of Variance.
Randomized Block Design (Kirk, chapter 7) BUSI 6480 Lecture 6.
5-5 Inference on the Ratio of Variances of Two Normal Populations The F Distribution We wish to test the hypotheses: The development of a test procedure.
Testing Hypotheses about Differences among Several Means.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 10-1 Chapter 10 Analysis of Variance Statistics for Managers Using Microsoft.
6-1 Introduction To Empirical Models Based on the scatter diagram, it is probably reasonable to assume that the mean of the random variable Y is.
6-3 Multiple Regression Estimation of Parameters in Multiple Regression.
AOV Assumption Checking and Transformations (§ )
Lecture 9-1 Analysis of Variance
1 Analysis of Variance & One Factor Designs Y= DEPENDENT VARIABLE (“yield”) (“response variable”) (“quality indicator”) X = INDEPENDENT VARIABLE (A possibly.
Analysis of Variance (One Factor). ANOVA Analysis of Variance Tests whether differences exist among population means categorized by only one factor or.
1 Experimental Statistics - week 14 Multiple Regression – miscellaneous topics.
PSYC 3030 Review Session April 19, Housekeeping Exam: –April 26, 2004 (Monday) –RN 203 –Use pencil, bring calculator & eraser –Make use of your.
By: Corey T. Williams 03 May Situation Objective.
Analysis Overheads1 Analyzing Heterogeneous Distributions: Multiple Regression Analysis Analog to the ANOVA is restricted to a single categorical between.
1 Experimental Statistics - week 9 Chapter 17: Models with Random Effects Chapter 18: Repeated Measures.
12-1 Chapter Twelve McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved.
1 Experimental Statistics Spring week 6 Chapter 15: Factorial Models (15.5)
Experimental Statistics - week 3
Topic 20: Single Factor Analysis of Variance. Outline Analysis of Variance –One set of treatments (i.e., single factor) Cell means model Factor effects.
Chapter 4 Analysis of Variance
IE241: Introduction to Design of Experiments. Last term we talked about testing the difference between two independent means. For means from a normal.
One-Way Analysis of Variance Recapitulation Recapitulation 1. Comparing differences among three or more subsamples requires a different statistical test.
ANOVA By Heather Nydam-Fragapane ANOVA Anova is an acronym for  Analysis of Variance.
The Mixed Effects Model - Introduction In many situations, one of the factors of interest will have its levels chosen because they are of specific interest.
Experimental Statistics - week 9
Copyright © 2004 by The McGraw-Hill Companies, Inc. All rights reserved.
Formula for Linear Regression y = bx + a Y variable plotted on vertical axis. X variable plotted on horizontal axis. Slope or the change in y for every.
ANOVA and Multiple Comparison Tests
1 Experimental Statistics - week 8 Chapter 17: Mixed Models Chapter 18: Repeated Measures.
1 Experimental Statistics - week 11 Chapter 11: Linear Regression and Correlation.
Chapter 11 Analysis of Variance
CHAPTER 13 Design and Analysis of Single-Factor Experiments:
i) Two way ANOVA without replication
Chapter 11 Analysis of Variance
6-1 Introduction To Empirical Models
Experimental Statistics - week 8
Presentation transcript:

The Completely Randomized Design (§8.3) Introduction to the simplest experimental design - the Completely Randomized Design. Introduce a statistical model for the observations in a completely randomized design.

Completely Randomized Design Two different Names for the Same Design: Experimental Study - Completely randomized design (CRD) Sampling Study - One-way classification design Randomization: The t treatments are randomly allocated to the experimental units in such a way that n1 units receive treatment 1, n2 receive treatment 2, etc. Assumptions: Independent random samples (response from one experimental unit does not affect responses from other experimental units). Responses follow a normal distribution. Common true variance, s2, across all groups/treatments. True mean for population i is mi. Interest is in comparing means.

AOV Model of Responses/Effects random error ~ N(0,s2) overall mean effect due to population i Requirement for  to be the overall mean: Expected response Estimate All ai = 0 implies all groups have the same mean (m)

Example A manufacturer of concrete bridge supports is interested in determining the effect of varying the sand content on the strength of the supports. Five supports are made for each of five different amounts of sand in the concrete mix and each is tested for compression resistance. Percent Sand 15 20 25 30 35 7 17 14 12 18 24 10 11 22 19 9 23

Basic Statistics and AOV Effects Percent Sand 15 20 25 30 35 7 17 14 12 18 24 10 11 22 19 9 23 9.6 15.4 17.6 21.6 10.8 -5.4 0.4 2.6 6.6 -4.2 Overall Mean MEAN EFFECT Sum of Effects

Decomposing the Data m = overall mean ai = i – m = group i effect ij = yij – m – ai = residual (Note that sum of residuals for each treatment is zero) Sum of squares

Decomposing Sums of Squares SSB SSW 6275.0 -5625.0 =650.0 -486.4 =163.6 -163.6 =0.0 TSS SSB SSW

Best Treatment? Is 30% significantly better than 25%?

Estimation

Reference Group/Cell Model random error ~ N(0,s2) reference group mean effect due to population i This is the model SASâ uses. Mean for the last group (i=t) is mt. Mean for the first group (i=1) is mt + b1 Thus, b1 is the difference between the mean of the reference group (cell) and the target group mean. Any group can be the reference group. All bi = 0 implies all groups have the same mean.

Basic Statistics and Reference Cell Effects Percent Sand 15 20 25 30 35 7 17 14 12 18 24 10 11 22 19 9 23 9.6 15.4 17.6 21.6 10.8 -1.2 4.6 6.8 21 Reference Cell Mean MEAN EFFECT Sum of Effects

Reference Cell Decomposition Note: Sums of squares don’t quite add up. Due to fact that sum of bi is not zero. 6275.0 -2916.0 =3369.0 -927.4 =2441.6 -163.6 =2278.0

Decomposing Sums of Squares 6275 = 2916.0 + 927.4 + 163.4 + 2278

Reference Cell Model

SAS Program options ls=78 ps=49 nodate; data stress; input sand resistance @@; datalines; 15 7 15 7 15 10 15 15 15 9 20 17 20 12 20 11 20 18 20 19 25 14 25 18 25 18 25 19 25 19 30 20 30 24 30 22 30 19 30 23 35 7 35 10 35 11 35 15 35 11 ; proc glm data=stress; class sand; model resistance = sand / solution; title2 'Compression resistance in concrete beams as'; title2 ' a function of percent sand in the mix'; run;

SAS Output(1) Compression resistance in concrete beams as a function of percent sand in the mix The GLM Procedure Dependent Variable: resistance Sum of Source DF Squares Mean Square F Value Pr > F Model 4 486.4000000 121.6000000 14.87 <.0001 Error 20 163.6000000 8.1800000 Corrected Total 24 650.0000000 R-Square Coeff Var Root MSE resistance Mean 0.748308 19.06713 2.860070 15.00000

SAS Output(2) Source DF Type I SS Mean Square F Value Pr > F sand 4 486.4000000 121.6000000 14.87 <.0001 Source DF Type III SS Mean Square F Value Pr > F Standard Parameter Estimate Error t Value Pr > |t| Intercept 10.80000000 B 1.27906216 8.44 <.0001 sand 15 -1.20000000 B 1.80886705 -0.66 0.5146 sand 20 4.60000000 B 1.80886705 2.54 0.0194 sand 25 6.80000000 B 1.80886705 3.76 0.0012 sand 30 10.80000000 B 1.80886705 5.97 <.0001 sand 35 0.00000000 B . . . NOTE: The X'X matrix has been found to be singular, and a generalized inverse was used to solve the normal equations. Terms whose estimates are followed by the letter 'B' are not uniquely estimable.

Minitab One-way ANOVA: Resist versus Sand Analysis of Variance for Resist Source DF SS MS F P Sand 4 486.40 121.60 14.87 0.000 Error 20 163.60 8.18 Total 24 650.00 Individual 95% CIs For Mean Based on Pooled StDev Level N Mean StDev -------+---------+---------+--------- 15 5 9.600 3.286 (----*-----) 20 5 15.400 3.647 (-----*----) 25 5 17.600 2.074 (----*-----) 30 5 21.600 2.074 (----*-----) 35 5 10.800 2.864 (-----*----) -------+---------+---------+--------- Pooled StDev = 2.860 10.0 15.0 20.0

Minitab Stat  ANOVA  One-Way Multiple comparisons (later)

Minitab Dot Plot

SPSS AOV Table ANOVA RESIST Sum of Squares df Mean Square F Sig. Between Groups 486.400 4 121.600 14.866 .000 Within Groups 163.600 20 8.180 Total 650.000 24

SPSS Descriptives Descriptives RESIST 95% Confidence Interval for Std. Mean N Mean Deviation Std. Error Lower Bound Upper Bound 15.00 5 9.6000 3.28634 1.46969 5.5195 13.6805 20.00 5 15.4000 3.64692 1.63095 10.8718 19.9282 25.00 5 17.6000 2.07364 .92736 15.0252 20.1748 30.00 5 21.6000 2.07364 .92736 19.0252 24.1748 35.00 5 10.8000 2.86356 1.28062 7.2444 14.3556 Total 25 15.0000 5.20416 1.04083 12.8518 17.1482 Model Fixed Effects 2.86007 .57201 13.8068 16.1932 Random Effects 2.20545 8.8767 21.1233 7.00 15.00 11.00 19.00 14.00 24.00 22.68400 Minimum Maximum Between- Component Variance

CRD Analysis in R > resist <- c(7,7,10,15,9,17,12,11,18,19,14, …,19,23,7,10,11,15,11) > sand <- factor(rep(seq(15,35,5),rep(5,5))) > myfit <- aov(resist~sand) > summary(myfit) Df Sum Sq Mean Sq F value Pr(>F) sand 4 486.40 121.60 14.866 8.655e-06 *** Residuals 20 163.60 8.18 --- Signif. codes: 0 `***' 0.001 `**' 0.01 `*' 0.05 `.' 0.1 ` ' 1 > coef(myfit) (Intercept) sand20 sand25 sand30 sand35 9.6 5.8 8.0 12.0 1.2 R functions aov() & lm() by default reference first cell mean!

Fixed Effects Normally, the “effect” of a particular treatment is assumed to be a constant value (ai) added to the response of all units in the group receiving the treatment. If the treatments are well defined, easily replicable and are expected to produce the same effect on average in each replicate, we have a fixed set of treatments and the AOV model is said to describe a fixed effects model. Examples: A scientist develops 3 new fungicides. Her interest is in these fungicides only. The impact of 4 specific soil types on plant growth are of interest. Three particular milling machines are being compared. Four particular lakes are of interest in their weed biomass densities. Three tests for assessing developmental learning are being compared.

Random Effects If the treatments cannot be assumed to be from a prespecified or known set of treatments, they are assumed to be a random sample from some larger population of potential treatments. In this case, the AOV model is called a random effects model and the ai are called random effects. Examples: A scientist is interested in how fungicides work. Ten (10) fungicides are selected (at random) to represent the population of all fungicides in the research (plots as replicates). Four soil sub groups are selected for examining plant growth (pots as replicates). Three milling machines selected at random from the production line are compared (runs as replicates). 16 lakes selected at random are measured for their weed biomass densities (water samples as replicates). A standard test for development is given to 20 middle school classes selected at random from the over 200 available among all middle schools in the county (student as replicate). In each case, we assume the values for the effects would change if our sample had changed. Inference is directed not to answering “which treatment is different from which other treatment?” but to the issue of “is the variability among treatments significantly greater than the residual variability?”.

Closing Comments on CRD Even though we have introduced several variations on the same basic model for defining “effects”, the final F-test for the hypothesis of overall equal group means is the same one developed as part of the analysis of variance. It turns out that there may be computational advantages to using the one formulation of the model over another, but this has absolutely no effect on the hypothesis test. We will see this in the next Section. For simple one-factor designs, whether the treatment effect is considered random or fixed, the F-test is the same, the interpretation is different.