Always be mindful of the kindness and not the faults of others.

Slides:



Advertisements
Similar presentations
Intro to ANOVA.
Advertisements

Chapter 11 Analysis of Variance
Analysis and Interpretation Inferential Statistics ANOVA
© 2010 Pearson Prentice Hall. All rights reserved The Complete Randomized Block Design.
Statistics for Managers Using Microsoft® Excel 5th Edition
Chapter 11 Analysis of Variance
ANOVA Determining Which Means Differ in Single Factor Models Determining Which Means Differ in Single Factor Models.
Analysis of Variance: Inferences about 2 or More Means
Lesson #23 Analysis of Variance. In Analysis of Variance (ANOVA), we have: H 0 :  1 =  2 =  3 = … =  k H 1 : at least one  i does not equal the others.
Chapter 3 Analysis of Variance
Statistics for Managers Using Microsoft® Excel 5th Edition
Basic concept of statistics Measures of central Measures of central tendency Measures of dispersion & variability.
Basic concept of statistics Measures of central Measures of central tendency Measures of dispersion & variability.
Chapter 17 Analysis of Variance
1 Pertemuan 10 Analisis Ragam (Varians) - 1 Matakuliah: I0262 – Statistik Probabilitas Tahun: 2007 Versi: Revisi.
Lecture 12 One-way Analysis of Variance (Chapter 15.2)
Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 7 th Edition Chapter 15 Analysis of Variance.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 10-1 Chapter 10 Analysis of Variance Statistics for Managers Using Microsoft.
Chap 10-1 Analysis of Variance. Chap 10-2 Overview Analysis of Variance (ANOVA) F-test Tukey- Kramer test One-Way ANOVA Two-Way ANOVA Interaction Effects.
6.1 - One Sample One Sample  Mean μ, Variance σ 2, Proportion π Two Samples Two Samples  Means, Variances, Proportions μ 1 vs. μ 2.
Chapter 12: Analysis of Variance
ANOVA Chapter 12.
Analysis of Variance Nutan S. Mishra Department of Mathematics and Statistics University of South Alabama.
QNT 531 Advanced Problems in Statistics and Research Methods
1 Multiple Comparison Procedures Once we reject H 0 :   =   =...  c in favor of H 1 : NOT all  ’s are equal, we don’t yet know the way in which.
When we think only of sincerely helping all others, not ourselves,
1 1 Slide © 2005 Thomson/South-Western Chapter 13, Part A Analysis of Variance and Experimental Design n Introduction to Analysis of Variance n Analysis.
INFERENTIAL STATISTICS: Analysis Of Variance ANOVA
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Comparing Three or More Means 13.
Chapter 10 Analysis of Variance.
ANOVA (Analysis of Variance) by Aziza Munir
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap th Lesson Analysis of Variance.
Everyday is a new beginning in life. Every moment is a time for self vigilance.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 10-1 Chapter 10 Analysis of Variance Statistics for Managers Using Microsoft.
Chapter 19 Analysis of Variance (ANOVA). ANOVA How to test a null hypothesis that the means of more than two populations are equal. H 0 :  1 =  2 =
One-Way Analysis of Variance
Be humble in our attribute, be loving and varying in our attitude, that is the way to live in heaven.
Lecture 9-1 Analysis of Variance
1 Always be mindful of the kindness and not the faults of others.
1 Analysis of Variance & One Factor Designs Y= DEPENDENT VARIABLE (“yield”) (“response variable”) (“quality indicator”) X = INDEPENDENT VARIABLE (A possibly.
Previous Lecture: Phylogenetics. Analysis of Variance This Lecture Judy Zhong Ph.D.
Basic concept of statistics Measures of central Measures of central tendency Measures of dispersion & variability.
Chapter 4 Analysis of Variance
Chap 11-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 11 Analysis of Variance.
Inferential Statistics Inferential statistics: The part of statistics that allows researchers to generalize their findings beyond data collected. Statistical.
Formula for Linear Regression y = bx + a Y variable plotted on vertical axis. X variable plotted on horizontal axis. Slope or the change in y for every.
 List the characteristics of the F distribution.  Conduct a test of hypothesis to determine whether the variances of two populations are equal.  Discuss.
DSCI 346 Yamasaki Lecture 4 ANalysis Of Variance.
Chapter 11 Analysis of Variance
Chapter 10 Two-Sample Tests and One-Way ANOVA.
Everyday is a new beginning in life.
Statistics for Managers Using Microsoft Excel 3rd Edition
Factorial Experiments
ANOVA Econ201 HSTS212.
CHAPTER 13 Design and Analysis of Single-Factor Experiments:
The greatest blessing in life is in giving and not taking.
i) Two way ANOVA without replication
Comparing Three or More Means
10 Chapter Chi-Square Tests and the F-Distribution Chapter 10
Chapter 10: Analysis of Variance: Comparing More Than Two Means
Chapter 10 Two-Sample Tests and One-Way ANOVA.
Comparing k Populations
Chapter 13: Comparing Several Means (One-Way ANOVA)
Chapter 11 Analysis of Variance
Be humble in our attribute, be loving and varying in our attitude, that is the way to live in heaven.
Chapter 15 Analysis of Variance
Quantitative Methods ANOVA.
STATISTICS INFORMED DECISIONS USING DATA
Presentation transcript:

Always be mindful of the kindness and not the faults of others.

One-way Anova: Inferences about More than Two Population Means What is Anova? One-Way Anova; F tests Pairwise comparisons: Bonferroni procedure

Analysis of Variance & One Factor Designs Y= DEPENDENT VARIABLE (“yield”) (“response variable”) (“quality indicator”) X = INDEPENDENT VARIABLE (A possibly influential FACTOR)

= Many other factors (possibly, some we’re unaware of) OBJECTIVE: To determine the impact of X on Y Mathematical Model: Y = f (x, ) , where  = (impact of) all factors other than X Ex: Y = Forced expiratory volume in one second (liters) X = Medical center (John Hopkins, Rancho Los Amigos, St. Louis) = Many other factors (possibly, some we’re unaware of)

Statistical Model • Yij “LEVEL” OF Center Yij = + j + ij (Brand is, of course, represented as “categorical”) “LEVEL” OF Center 1 2 • • •  •  •  • • • C 1 2 • n Y11 Y12 • • • • • • •Y1c Yij = + j + ij i = 1, . . . . . , nj j = 1, . . . . . , C Y21 • YnI • Yij Ync •   •  •   •    •   •    •    • 

Let mj = AVERAGE associated with jth level of X Where = OVERALL AVERAGE j = index for FACTOR (center) LEVEL i= index for “replication” j = Differential effect (response) associated with jth level of X and ij = “noise” or “error” associated with the (particular) (i,j)th data value. Let mj = AVERAGE associated with jth level of X  tj = mj – m and m = AVERAGE of mj .

•••• Y1, Y2, etc., are Column Means 1 2 3 C Y11 Y12 • • • • • •Y1c Y21 YRI •••• YRc •  •  •  •  •  •  •  •  •  Y 1 Y 2 •  •  •   (Y j) •  •     Y c Y1, Y2, etc., are Column Means               

Y • = Y j /C = “GRAND MEAN” (assuming same # data points in each column) (otherwise, Y • = mean of all the data) c j=1

These estimates are based on Gauss’ (1796) PRINCIPLE OF LEAST SQUARES MODEL: Yij =  + j + ij Y• estimates  Yj - Y • estimatesj (= mj – m) (for all j) These estimates are based on Gauss’ (1796) PRINCIPLE OF LEAST SQUARES and on COMMON SENSE

MODEL: Yij =  + j + ij If you insert the estimates into the MODEL, (1) Yij = Y • + (Yj - Y • ) + ij. < it follows that our estimate of ij is (2) ij = Yij - Yj <

{ { { Then, Yij = Y• + (Yj - Y• ) + ( Yij - Yj) or, (Yij - Y• ) = (Yj - Y•) + (Yij - Yj ) { { { (3) TOTAL VARIABILITY in Y Variability in Y associated with X Variability in Y associated with all other factors + =

SUM OF SQUARES BETWEEN COLUMNS SUM OF SQUARES WITHIN COLUMNS If you square both sides of (3), and double sum both sides (over i and j), you get, [after some unpleasant algebra, but lots of terms which “cancel”] {{ C nj C C nj (Yij - Y• )2 =  nj(Yj - Y•)2 + (Yij - Yj)2 { j=1 i=1 j=1 j=1 i=1 TSS TOTAL SUM OF SQUARES ( SSB SUM OF SQUARES BETWEEN COLUMNS = + SSW (SSE) SUM OF SQUARES WITHIN COLUMNS ( ( ( ( (

ANOVA TABLE SSB SSB C - 1 = MSB C - 1 SSW = MSW SSW N - C N-C SOURCE OF VARIABILITY SSQ DF Mean square (M.S.) Between Columns (due to center) SSB SSB C - 1 = MSB C - 1 Within Columns (due to other factors) SSW = MSW SSW N - C N-C TOTAL TSS N - 1

ANOVA TABLE Source of Variability df SSQ M.S. CENTER 1.583 2 0.791 = 3 - 1 0.791 ERROR 14.480 57 = 59 - 2 0.254 TOTAL 115.84 59 = 60 -1

> 1 , < 1 , We can show: E ( MSB ) = 2 + VCOL E ( MSW ) = 2 This suggests that There’s some evidence of non-zero VCOL, or “level of X affects Y” if MSB > 1 , MSW if MSB No evidence that VCOL > 0, or that “level of X affects Y” < 1 , MSW

With HO: Level of X has no impact on Y HI: Level of X does have impact on Y, We need MSB > > 1 MSW to reject HO.

More Formally, HO: 1 = 2 = • • • c = 0 HI: not all j = 0 OR (All column means are equal) HO: 1 = 2 = • • • • c HI: not all j are EQUAL

 The distribution of MSB = “Fcalc” , is MSW The F - distribution with (dfB, dfw) degrees of freedom  Assuming HO true. Ca = Table Value

In our problem: ANOVA TABLE Source of Variability df SSQ M.S. Fcalc CENTER 1.583 2 = 3 - 1 0.791 3.12=0.791/0.254 ERROR 14.480 57 = 59 - 2 0.254 TOTAL 115.84 59 = 60 -1

F table: Table A-5  = .05 C0.5 = 3.15 Fcal =3.12 (2, 57 DF)

Hence, at  =. 05, Do Not Reject Ho , i. e Hence, at  = .05, Do Not Reject Ho , i.e., Conclude that centers don’t differ significantly on FEV1 at 5% level. P-value is .052, so it is significant at 6% level

Multiple Comparison Procedures Once we reject H0: ==...c in favor of H1: NOT all ’s are equal, we don’t yet know the way in which they’re not all equal, but simply that they’re not all the same. If there are 4 columns, are all 4 ’s different? Are 3 the same and one different? If so, which one? etc.

Overall Type I Error Rate We set up “” as the significance level for a hypothesis test. Suppose we test 3 independent hypotheses, each at = .05; each test has type I error (rej H0 when it’s true) of .05. However, P(at least one type I error in the 3 tests) = 1-P( accept all ) = 1 - (.95)3  .14 3, given true

Pairwise Comparisons Bonferroni Correction: Do a series of pairwise t-tests, each with specified  value divided by # of comparisons. Pairwise Comparisons

MINITAB INPUT center fev1 1 3.23 1 3.47 1 1.86 1 2.47 . . 3 2.85 1 3.23 1 3.47 1 1.86 1 2.47 . . 3 2.85 3 2.43 3 3.20 3 3.53

ONE FACTOR ANOVA (MINITAB) MINITAB: STAT>>ANOVA>>ONE-WAY Click for comparisons

Minitab Outputs Fisher 98.3% Individual Confidence Intervals All Pairwise Comparisons among Levels of center Simultaneous confidence level = 95.58% center = 1 subtracted from: center Lower Center Upper ------+---------+---------+---------+--- 2 -0.0049 0.4063 0.8176 (-----------*----------) 3 -0.1215 0.2525 0.6266 (---------*----------) ------+---------+---------+---------+--- -0.35 0.00 0.35 0.70 center = 2 subtracted from: center Lower Center Upper ------+---------+---------+---------+--- 3 -0.5572 -0.1538 0.2496 (-----------*----------)