When we think only of sincerely helping all others, not ourselves,

Slides:



Advertisements
Similar presentations
Intro to ANOVA.
Advertisements

Analysis of Variance (ANOVA) Statistics for the Social Sciences Psychology 340 Spring 2010.
One-Way ANOVA Multiple Comparisons.
MARE 250 Dr. Jason Turner Analysis of Variance (ANOVA)
Lecture 10 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
Be humble in our attribute, be loving and varying in our attitude, that is the way to live in heaven.
ANOVA Determining Which Means Differ in Single Factor Models Determining Which Means Differ in Single Factor Models.
Comparing Means.
PSY 307 – Statistics for the Behavioral Sciences
Analysis of Variance (ANOVA) MARE 250 Dr. Jason Turner.
Analysis of Variance Chapter 15 - continued Two-Factor Analysis of Variance - Example 15.3 –Suppose in Example 15.1, two factors are to be examined:
Lecture 9: One Way ANOVA Between Subjects
Two Groups Too Many? Try Analysis of Variance (ANOVA)
8. ANALYSIS OF VARIANCE 8.1 Elements of a Designed Experiment
Statistics for the Social Sciences Psychology 340 Spring 2005 Analysis of Variance (ANOVA)
One-way Between Groups Analysis of Variance
Lecture 12 One-way Analysis of Variance (Chapter 15.2)
The Analysis of Variance
Comparing Means.
Statistical Methods in Computer Science Hypothesis Testing II: Single-Factor Experiments Ido Dagan.
Linear Contrasts and Multiple Comparisons (Chapter 9)
Chapter 12: Analysis of Variance
Analysis of Variance (ANOVA) Quantitative Methods in HPELS 440:210.
Intermediate Applied Statistics STAT 460
1 Multiple Comparison Procedures Once we reject H 0 :   =   =...  c in favor of H 1 : NOT all  ’s are equal, we don’t yet know the way in which.
ANOVA Greg C Elvers.
Comparing Means. Anova F-test can be used to determine whether the expected responses at the t levels of an experimental factor differ from each other.
1 1 Slide © 2005 Thomson/South-Western Chapter 13, Part A Analysis of Variance and Experimental Design n Introduction to Analysis of Variance n Analysis.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Comparing Three or More Means 13.
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Design and Analysis of Experiments Dr. Tai-Yue Wang Department of Industrial and Information Management National Cheng Kung University Tainan, TAIWAN,
t(ea) for Two: Test between the Means of Different Groups When you want to know if there is a ‘difference’ between the two groups in the mean Use “t-test”.
Statistics 11 Confidence Interval Suppose you have a sample from a population You know the sample mean is an unbiased estimate of population mean Question:
ANOVA (Analysis of Variance) by Aziza Munir
Everyday is a new beginning in life. Every moment is a time for self vigilance.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, October 15, 2013 Analysis of Variance (ANOVA)
Chapter 19 Analysis of Variance (ANOVA). ANOVA How to test a null hypothesis that the means of more than two populations are equal. H 0 :  1 =  2 =
Be humble in our attribute, be loving and varying in our attitude, that is the way to live in heaven.
STA MCP1 Multiple Comparisons: Example Study Objective: Test the effect of six varieties of wheat to a particular race of stem rust. Treatment:
1 Analysis of Variance Chapter 14 2 Introduction Analysis of variance helps compare two or more populations of quantitative data. Specifically, we are.
1 Always be mindful of the kindness and not the faults of others.
Analysis of Variance (One Factor). ANOVA Analysis of Variance Tests whether differences exist among population means categorized by only one factor or.
One-way ANOVA: - Comparing the means IPS chapter 12.2 © 2006 W.H. Freeman and Company.
1 Orthogonality One way to delve further into the impact a factor has on the yield is to break the Sum of Squares (SSQ) into “orthogonal” components. If.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: One-way ANOVA Marshall University Genomics Core.
Copyright © Cengage Learning. All rights reserved. 12 Analysis of Variance.
ANOVA P OST ANOVA TEST 541 PHL By… Asma Al-Oneazi Supervised by… Dr. Amal Fatani King Saud University Pharmacy College Pharmacology Department.
Two-Way (Independent) ANOVA. PSYC 6130A, PROF. J. ELDER 2 Two-Way ANOVA “Two-Way” means groups are defined by 2 independent variables. These IVs are typically.
McGraw-Hill, Bluman, 7th ed., Chapter 12
Statistics for the Social Sciences Psychology 340 Spring 2009 Analysis of Variance (ANOVA)
Introduction to ANOVA Research Designs for ANOVAs Type I Error and Multiple Hypothesis Tests The Logic of ANOVA ANOVA vocabulary, notation, and formulas.
Topic 22: Inference. Outline Review One-way ANOVA Inference for means Differences in cell means Contrasts.
Chapters Way Analysis of Variance - Completely Randomized Design.
Formula for Linear Regression y = bx + a Y variable plotted on vertical axis. X variable plotted on horizontal axis. Slope or the change in y for every.
Posthoc Comparisons finding the differences. Statistical Significance What does a statistically significant F statistic, in a Oneway ANOVA, tell us? What.
The 2 nd to last topic this year!!.  ANOVA Testing is similar to a “two sample t- test except” that it compares more than two samples to one another.
Everyday is a new beginning in life.
CHAPTER 13 Design and Analysis of Single-Factor Experiments:
Comparing Three or More Means
Statistics Analysis of Variance.
STAT 6304 Final Project Fall, 2016.
Linear Contrasts and Multiple Comparisons (§ 8.6)
Multiple Comparisons: Example
Always be mindful of the kindness and not the faults of others.
Comparing Means.
Joyful mood is a meritorious deed that cheers up people around you
Be humble in our attribute, be loving and varying in our attitude, that is the way to live in heaven.
1-Way Analysis of Variance - Completely Randomized Design
Presentation transcript:

When we think only of sincerely helping all others, not ourselves, We will find that we receive all that we wish for. multiple comparisons

Chapter 9: Multiple Comparisons Error rate of control Pairwise comparisons Comparisons to a control Linear contrasts multiple comparisons

Multiple Comparison Procedures Once we reject H0: ==...t in favor of H1: NOT all ’s are equal, we don’t yet know the way in which they’re not all equal, but simply that they’re not all the same. If there are 4 columns (levels), are all 4 ’s different? Are 3 the same and one different? If so, which one? etc. multiple comparisons

These “more detailed” inquiries into the process are called MULTIPLE COMPARISON PROCEDURES. Errors (Type I): We set up “” as the significance level for a hypothesis test. Suppose we test 3 independent hypotheses, each at = .05; each test has type I error (rej H0 when it’s true) of .05. However, P(at least one type I error in the 3 tests) = 1-P( accept all ) = 1 - (.95)3  .14 3, given true multiple comparisons

In other words, Probability is In other words, Probability is .14 that at least one type one error is made. For 5 tests, prob = .23. Question - Should we choose = .05, and suffer (for 5 tests) a .23 Experimentwise Error rate (“a” or aE)? OR Should we choose/control the overall error rate, “a”, to be .05, and find the individual test  by 1 - (1-)5 = .05, (which gives us  = .011)? multiple comparisons

would be valid only if the tests are independent; often they’re not. The formula 1 - (1-)5 = .05 would be valid only if the tests are independent; often they’re not. [ e.g., 1=22= 3, 1= 3 IF accepted & rejected, isn’t it more likely that rejected? ] 1 2 3 1 2 3 multiple comparisons

Error Rates When the tests are not independent, it’s usually very difficult to arrive at the correct for an individual test so that a specified value results for the experimentwise error rate (or called family error rate). multiple comparisons

There are many multiple comparison procedures. We’ll cover only a few. Pairwise Comparisons Method 1: (Fisher Test) Do a series of pairwise t-tests, each with specified  value (for individual test). This is called “Fisher’s LEAST SIGNIFICANT DIFFERENCE” (LSD). multiple comparisons

Example: Broker Study A financial firm would like to determine if brokers they use to execute trades differ with respect to their ability to provide a stock purchase for the firm at a low buying price per share. To measure cost, an index, Y, is used. Y=1000(A-P)/A where P=per share price paid for the stock; A=average of high price and low price per share, for the day. “The higher Y is the better the trade is.” multiple comparisons

} n=6 Five brokers were in the study and six trades CoL: broker 1 12 3 5 -1 6 2 7 17 13 11 12 3 8 1 7 4 5 4 21 10 15 12 20 6 14 5 24 13 14 18 19 17 } n=6 Five brokers were in the study and six trades were randomly assigned to each broker. multiple comparisons

 = .05, FTV = 2.76 (reject equal column MEANS) “MSW”  = .05, FTV = 2.76 (reject equal column MEANS) multiple comparisons

For any comparison of 2 columns, Yi -Yj /2 /2 CL Cu AR: 0+ ta/2 x MSW x 1 + 1 ni nj dfw (ni = nj = 6, here) Pooled Variance, the estimate for the common variance MSW : multiple comparisons

In our example, with=.05 0  2.060 (21.2 x 1 + 1 ) 0 5.48 This value, 5.48 is called the Least Significant Difference (LSD). When same number of data points, n, in each column, LSD = ta/2 x 2xMSW. n multiple comparisons

Underline Diagram Col: 3 1 2 4 5 Summarize the comparison results. (p. 443) Now, rank order and compare: Col: 3 1 2 4 5 5 6 12 14 17 multiple comparisons

Step 2: identify difference > 5.48, and mark accordingly: 5 6 12 14 17 3 1 2 4 5 3: compare the pair of means within each subset: Comparison difference vs. LSD 3 vs. 1 2 vs. 4 2 vs. 5 4 vs. 5 < * 5 * Contiguous; no need to detail multiple comparisons

> Conclusion : 3, 1 2, 4, 5 Can get “inconsistency”: Suppose col 5 were 18: 3 1 2 4 5 5 6 12 14 18 Now: Comparison |difference| vs. LSD < > 3 vs. 1 2 vs. 4 2 vs. 5 4 vs. 5 * 6 Conclusion : 3, 1 2 4 5 ??? multiple comparisons

Conclusion : 3, 1 2 4 5 Broker 1 and 3 are not significantly different but they are significantly different to the other 3 brokers. Broker 2 and 4 are not significantly different, and broker 4 and 5 are not significantly different, but broker 2 is different to (smaller than) broker 5 significantly. multiple comparisons

multiple comparisons

Minitab: Stat>>ANOVA>>One-Way Anova then click “comparisons”. Fisher's pairwise comparisons (Minitab) Family error rate = 0.268 Individual error rate = 0.0500 Critical value = 2.060  t_a/2 Intervals for (column level mean) - (row level mean) 1 2 3 4 2 -11.476 -0.524 3 -4.476 1.524 6.476 12.476 4 -13.476 -7.476 -14.476 -2.524 3.476 -3.524 5 -16.476 -10.476 -17.476 -8.476 -5.524 0.476 -6.524 2.476 Col 1 < Col 2 Cannot reject Col 2 = Col 4 multiple comparisons

Pairwise comparisons Method 2: (Tukey Test) A procedure which controls the experimentwise error rate is “TUKEY’S HONESTLY SIGNIFICANT DIFFERENCE TEST ”. multiple comparisons

or, for equal number of data points/col Tukey’s method works in a similar way to Fisher’s LSD, except that the “LSD” counterpart (“HSD”) is not ta/2 x MSW x  1 + 1 ni nj ( or, for equal number of data points/col ) = , ta/2 x 2xMSW n but tuk X 2xMSW , a/2 n where tuk has been computed to take into account all the inter-dependencies of the different comparisons. multiple comparisons

HSD = tuka/2x2MSW n _______________________________________ A more general approach is to write HSD = qaxMSW n where qa = tuka/2 x2 --- q = (Ylargest - Ysmallest) / MSW n ---- probability distribution of q is called the “Studentized Range Distribution”. --- q = q(t, df), where t =number of columns, and df = df of MSW multiple comparisons

With t = 5 and df = v= 25, from Table 10: q = 4. 15 for a= 5% tuk = 4 Then, HSD = 4.15 21.2/6 = 7.80 also, 2.93 2x21.2/6 = 7.80 multiple comparisons

In our earlier example: 3 1 2 4 5 5 6 12 14 17 Rank order: (No differences [contiguous] > 7.80) multiple comparisons

Comparison |difference| >or< 7.80 3 vs. 1 3 vs. 2 3 vs. 4 (contiguous) * 7 9 12 * 8 11 5 3, 1, 2 4, 5 2 is “same as 1 and 3, but also same as 4 and 5.” multiple comparisons

Minitab: Stat>>ANOVA>>One-Way Anova then click “comparisons”. Tukey's pairwise comparisons (Minitab) Family error rate = 0.0500 Individual error rate = 0.00706 Critical value = 4.15  q_a Intervals for (column level mean) - (row level mean) 1 2 3 4 2 -13.801 1.801 3 -6.801 -0.801 8.801 14.801 4 -15.801 -9.801 -16.801 -0.199 5.801 -1.199 5 -18.801 -12.801 -19.801 -10.801 -3.199 2.801 -4.199 4.801 multiple comparisons

Method 3: Dunnett’s test Special Multiple Comp. Method 3: Dunnett’s test Designed specifically for (and incorporating the interdependencies of) comparing several “treatments” to a “control.” Col Example: 1 2 3 4 5 } n=6 6 12 5 14 17 CONTROL Analog of LSD (=t/2 x 2 MSW ) D = Dut/2 x 2 MSW n n From table or Minitab multiple comparisons

Comparison |difference| >or< 6.94 D= Dut/2 x 2 MSW/n = 2.61 (2(21.2) ) = 6.94 CONTROL 6 1 2 3 4 5 In our example: 6 12 5 14 17 Comparison |difference| >or< 6.94 1 vs. 2 1 vs. 3 1 vs. 4 1 vs. 5 < > 6 1 8 11 - Cols 4 and 5 differ from the control [ 1 ]. - Cols 2 and 3 are not significantly different from control. multiple comparisons

Minitab: Stat>>ANOVA>>General Linear Model then click “comparisons”. Dunnett's comparisons with a control (Minitab) Family error rate = 0.0500  controlled!! Individual error rate = 0.0152 Critical value = 2.61  Dut_a/2 Control = level (1) of broker Intervals for treatment mean minus control mean Level Lower Center Upper --+---------+---------+---------+----- 2 -0.930 6.000 12.930 (---------*--------) 3 -7.930 -1.000 5.930 (---------*--------) 4 1.070 8.000 14.930 (--------*---------) 5 4.070 11.000 17.930 (---------*---------) --+---------+---------+---------+----- -7.0 0.0 7.0 14.0 multiple comparisons

What Method Should We Use? Fisher procedure can be used only after the F-test in the Anova is significant at 5%. Otherwise, use Tukey procedure. Note that to avoid being too conservative, the significance level of Tukey test can be set bigger (10%), especially when the number of levels is big. Or use S-N-K procedure. multiple comparisons

Contrast 1 2 3 4 Consider the following data, which, let’s say, are the column means of a one factor ANOVA, with the one factor being “DRUG”: 1 2 3 4 Consider 4 column means: Y.1 Y.2 Y.3 Y.4 6 4 1 -3 Grand Mean = Y.. = 2 # of rows (replicates) = R = 8

Contrast Example 1 Suppose the questions of interest are 2 3 4 Sulfa Type S1 Sulfa Type S2 Anti-biotic Type A Placebo Suppose the questions of interest are (1) Placebo vs. Non-placebo (2) S1 vs. S2 (3) (Average) S vs. A multiple comparisons

For (1), we would like to test if the mean of Placebo is equal to the mean of other levels, i.e. the mean value of {Y.1-(Y.2 +Y.3 +Y.4)/3} is equal to 0. For (2), we would like to test if the mean of S1 is equal to the mean of S2, i.e. the mean value of (Y.2-Y.3) is equal to 0. For (3), we would like to test if the mean of Types S1 and S2 is equal to the mean of Type A, i.e. the mean value of {(Y.2 +Y.3 )/2-Y.4} is equal to 0.

with restriction that Saj = 0. In general, a question of interest can be expressed by a linear combination of column means such as with restriction that Saj = 0. Such linear combinations are called (linear) contrasts. multiple comparisons

Test if a contrast has mean 0 The sum of squares for contrast Z is where n is the number of rows (replications). The test statistic Fcalc = SSZ/MSW is distributed as F with 1 and (df of error) degrees of freedom. Reject E[Z]= 0 if the observed Fcalc is too large (say, > F0.05(1,df of error) at 5% significant level). multiple comparisons

Example 1 (cont.): aj’s for the 3 contrasts 1 2 3 4 P vs. P: Z1 S1 vs. S2:Z2 S vs. A: Z3 -3 1 1 1 0 -1 1 0 0 -1 -1 2 multiple comparisons

  Calculating top row middle row bottom row       multiple comparisons

Y.1 Y.2 Y.3 Y.4 Placebo vs. drugs S1 vs. S2 Average S vs. A 5 6 7 10 P S1 S2 A Placebo vs. drugs S1 vs. S2 Average S vs. A -3 1 1 1 5.33 0.50 -1 1 8.17 -1 -1 2 14.00 multiple comparisons

5.33 42.64 .50 4.00 (Y.j - Y..)2 = 14. SSBc = 14.R; R = # rows= 8. 8.17 65.36 SSBc ! 14.00 112.00

Orthogonal Contrasts ai1j . ai2j = 0 for all i1, i2, j i1 = i2. A set of k contrasts { Zi = , i=1,2,…,k } are called orthogonal if ai1j . ai2j = 0 for all i1, i2, j i1 = i2. If k = c -1 (the df of “column” term and c: # of columns), then

Orthogonal Contrasts If a set of contrasts are orthogonal, their corresponding questions are called independent because the probabilities of Type I and Type II errors in the ensuing hypothesis tests are independent, and “stand alone”. That is, the variability in Y due to one contrast (question) will not be affected by that due to the other contrasts (questions).

Orthogonal Breakdown Since SSBcol has (C-1) df (which corresponds with having C levels, or C columns ), the SSBcol can be broken up into (C-1) individual SSQ values, each with a single degree of freedom, each addressing a different inquiry into the data’s message (one question). A set of C-1 orthogonal contrasts (questions) provides an orthogonal breakdown.

Recall Data in Example 1: { Placebo . 5 S1 . 6 S2 . 7 A . 10 R=8 Y..= 7

ANOVA F1-.05(3,28)=2.95

An Orthogonal Breakdown Source SSQ df MSQ F Z1 Z2 Z3 { { 42.64 4.00 65.36 { 42.64 4.00 65.36 8.53 .80 13.07 1 3 Drugs Error 112 140 28 5 F1-.05(1,28)=4.20

Example 1 (Conti.): Conclusions The mean response for Placebo is significantly different to that for Non-placebo. There is no significant difference between using Types S1 and S2. Using Type A is significantly different to using Type S on average. multiple comparisons

What if contrasts of interest are not orthogonal? Let k be the number of contrasts of interest; c be the number of levels If k <= c-1  Bonferroni method If k > c-1  Bonferroni or Scheffe method *Bonferroni Method: The same F test but use a = a/k, where a is the desired family error rate (usual at 5%). *Scheffe Method: To test all linear combinations at once. Very conservative. (Section 9.8)

Special Pairwise Comp. Method 4: MCB Procedure (Compare to the best) This procedure provides a subset of treatments that cannot distinguished from the best. The probability of that the “best” treatment is included in this subset is controlled at 1-a. *Assume that the larger the better. If not, change response to –y.

Identify the subset of the best brokers Minitab: Stat>>ANOVA>>One-Way Anova then click “comparisons”, HSU’s MCB Hsu's MCB (Multiple Comparisons with the Best) Family error rate = 0.0500 Critical value = 2.27 Intervals for level mean minus largest of other level means Level Lower Center Upper ---+---------+---------+---------+---- 1 -17.046 -11.000 0.000 (------*-------------) 2 -11.046 -5.000 1.046 (-------*------) 3 -18.046 -12.000 0.000 (-------*--------------) 4 -9.046 -3.000 3.046 (------*-------) 5 -3.046 3.000 9.046 (-------*------) ---+---------+---------+---------+---- -16.0 -8.0 0.0 8.0 Brokers 2, 4, 5 Not included; only if the interval (excluding ends) covers 0, this level is selected.