ANOVA 3/19/12 Mini Review of simulation versus formulas and theoretical distributions Analysis of Variance (ANOVA) to compare means: testing for a difference.

Slides:



Advertisements
Similar presentations
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan ANOVA SECTION 8.1 Testing for a difference in means across multiple categories.
Advertisements

Chi-Square Tests 3/14/12 Testing the distribution of a single categorical variable :  2 goodness of fit Testing for an association between two categorical.
Statistics: Unlocking the Power of Data Lock 5 Testing Goodness-of- Fit for a Single Categorical Variable Kari Lock Morgan Section 7.1.
BPS - 5th Ed. Chapter 241 One-Way Analysis of Variance: Comparing Several Means.
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan 10/23/12 Sections , Single Proportion, p Distribution (6.1)
Hypothesis Testing: Intervals and Tests
Hypothesis Testing I 2/8/12 More on bootstrapping Random chance
ANOVA: Analysis of Variation
1 1 Slide © 2009, Econ-2030 Applied Statistics-Dr Tadesse Chapter 10: Comparisons Involving Means n Introduction to Analysis of Variance n Analysis of.
Copyright ©2011 Brooks/Cole, Cengage Learning Analysis of Variance Chapter 16 1.
Analysis of Variance: Inferences about 2 or More Means
Statistics Are Fun! Analysis of Variance
Today’s Agenda Review of ANOVA Module 9 Review for Exam 2 Please log in with your UMID and your participation will be graded by the number of questions.
Lecture 12 One-way Analysis of Variance (Chapter 15.2)
STAT 101 Dr. Kari Lock Morgan Exam 2 Review.
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved.
Inference for Categorical Variables 2/29/12 Single Proportion, p Distribution Intervals and tests Difference in proportions, p 1 – p 2 One proportion or.
Introducing Inference with Simulation Methods; Implementation at Duke University Kari Lock Morgan Department of Statistical Science, Duke University
Chapter 12: Analysis of Variance
Synthesis and Review 3/26/12 Multiple Comparisons Review of Concepts Review of Methods - Prezi Essential Synthesis 3 Professor Kari Lock Morgan Duke University.
1 Advances in Statistics Or, what you might find if you picked up a current issue of a Biological Journal.
AM Recitation 2/10/11.
F-Test ( ANOVA ) & Two-Way ANOVA
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 14 Analysis.
STAT 3130 Statistical Methods I Session 2 One Way Analysis of Variance (ANOVA)
1 1 Slide © 2006 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 Tests with two+ groups We have examined tests of means for a single group, and for a difference if we have a matched sample (as in husbands and wives)
1 1 Slide © 2005 Thomson/South-Western Chapter 13, Part A Analysis of Variance and Experimental Design n Introduction to Analysis of Variance n Analysis.
1 1 Slide Analysis of Variance Chapter 13 BA 303.
Analysis of Variance ST 511 Introduction n Analysis of variance compares two or more populations of quantitative data. n Specifically, we are interested.
Introducing Inference with Simulation Methods; Implementation at Duke University Kari Lock Morgan Department of Statistical Science, Duke University
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan 11/1/12 ANOVA SECTION 8.1 Testing for a difference in means across multiple.
ANOVA One Way Analysis of Variance. ANOVA Purpose: To assess whether there are differences between means of multiple groups. ANOVA provides evidence.
PSY 307 – Statistics for the Behavioral Sciences Chapter 16 – One-Factor Analysis of Variance (ANOVA)
ANOVA (Analysis of Variance) by Aziza Munir
Basic concept Measures of central tendency Measures of central tendency Measures of dispersion & variability.
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan Multiple Regression SECTIONS 9.2, 10.1, 10.2 Multiple explanatory variables.
1 Chapter 13 Analysis of Variance. 2 Chapter Outline  An introduction to experimental design and analysis of variance  Analysis of Variance and the.
Multiple Regression I 4/9/12 Transformations The model Individual coefficients R 2 ANOVA for regression Residual standard error Section 9.4, 9.5 Professor.
Chapter 19 Analysis of Variance (ANOVA). ANOVA How to test a null hypothesis that the means of more than two populations are equal. H 0 :  1 =  2 =
Inference after ANOVA, Multiple Comparisons 3/21/12 Inference after ANOVA The problem of multiple comparisons Bonferroni’s Correction Section 8.2 Professor.
Introducing Inference with Bootstrapping and Randomization Kari Lock Morgan Department of Statistical Science, Duke University with.
ANOVA: Analysis of Variance. The basic ANOVA situation Two variables: 1 Nominal, 1 Quantitative Main Question: Do the (means of) the quantitative variables.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Statistical Testing of Differences CHAPTER fifteen.
ANALYSIS OF VARIANCE STAT 200. OUTLINE Introduction of concepts without numbers, notation or details Motivation The four steps: Hypothesis in words The.
Statistics: Unlocking the Power of Data Lock 5 Exam 2 Review STAT 101 Dr. Kari Lock Morgan 11/13/12 Review of Chapters 5-9.
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan 10/30/12 Chi-Square Tests SECTIONS 7.1, 7.2 Testing the distribution of a.
Copyright © Cengage Learning. All rights reserved. 12 Analysis of Variance.
AP Statistics: ANOVA Section 1. In section 13.1 A, we used a t-test to compare the means between two groups. An ANOVA (ANalysis Of VAriance) test is used.
Statistics: Unlocking the Power of Data Lock 5 STAT 250 Dr. Kari Lock Morgan SECTION 7.1 Testing the distribution of a single categorical variable : χ.
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan 11/20/12 Multiple Regression SECTIONS 9.2, 10.1, 10.2 Multiple explanatory.
Econ 3790: Business and Economic Statistics Instructor: Yogesh Uppal
Econ 3790: Business and Economic Statistics Instructor: Yogesh Uppal
Section 6.4 Inferences for Variances. Chi-square probability densities.
Synthesis and Review 2/20/12 Hypothesis Tests: the big picture Randomization distributions Connecting intervals and tests Review of major topics Open Q+A.
Statistics: Unlocking the Power of Data Lock 5 STAT 250 Dr. Kari Lock Morgan SECTION 7.1 Testing the distribution of a single categorical variable : 
1/54 Statistics Analysis of Variance. 2/54 Statistics in practice Introduction to Analysis of Variance Analysis of Variance: Testing for the Equality.
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
 List the characteristics of the F distribution.  Conduct a test of hypothesis to determine whether the variances of two populations are equal.  Discuss.
The 2 nd to last topic this year!!.  ANOVA Testing is similar to a “two sample t- test except” that it compares more than two samples to one another.
Chapter 13 Analysis of Variance (ANOVA). ANOVA can be used to test for differences between three or more means. The hypotheses for an ANOVA are always:
ANOVA: Analysis of Variation
ANOVA: Analysis of Variation
ANOVA: Analysis of Variation
ANOVA: Analysis of Variation
AP Statistics FINAL EXAM ANALYSIS OF VARIANCE.
Statistics Analysis of Variance.
Econ 3790: Business and Economic Statistics
Chapter 10 – Part II Analysis of Variance
ANOVA: Analysis of Variance
Presentation transcript:

ANOVA 3/19/12 Mini Review of simulation versus formulas and theoretical distributions Analysis of Variance (ANOVA) to compare means: testing for a difference in means between multiple groups Section 8.1 Professor Kari Lock Morgan Duke University

Anonymous Midterm Evaluation (due TODAY, 5pm) Project 1 (due Thursday, 3/22, 5pm) Project 1 Homework 7 (due Monday, 3/26) Homework 7 NO LATE HOMEWORK ACCEPTED! Turn in by Friday, 3/23, 5pm to get it graded before Exam 2. To Do

Two Options for p-values We have learned two ways of calculating p-values: The only difference is how to create a distribution of the statistic, assuming the null is true: 1)Simulation (Randomization Test): Directly simulate what would happen, just by random chance, if the null were true 2)Formulas and Theoretical Distributions: Use a formula to create a test statistic for which we know the theoretical distribution when the null is true, if sample sizes are large enough

Two Options for Intervals We have learned two ways of calculating intervals: 1)Simulation (Bootstrap): Assess the variability in the statistic by creating many bootstrap statistics 2)Formulas and Theoretical Distributions: Use a formula to calculate the standard error of the statistic, and use the normal or t- distribution to find z* or t*, if sample sizes are large enough

Pros and Cons 1)Simulation Methods PROS: Methods tied directly to concepts, emphasizing conceptual understanding Same procedure for every statistic No formulas or theoretical distributions to learn and distinguish between Works for any sample size Minimal math needed CONS: Need entire dataset (if quantitative variables) Need a computer Newer approach, so different from the way most people do statistics

Pros and Cons 2)Formulas and Theoretical Distributions PROS: Only need summary statistics Only need a calculator The approach most people take CONS: Plugging numbers into formulas does little for conceptual understanding Many different formulas and distributions to learn and distinguish between Harder to see the big picture when the details are different for each statistic Doesn’t work for small sample sizes Requires more math and background knowledge

Two Options If the sample size is small, you have to use simulation methods If the sample size is large, you can use whichever method you prefer It is redundant to use both methods, unless you want to check your answers

Accuracy The accuracy of simulation methods depends on the number of simulations (more simulations = more accurate) The accuracy of formulas and theoretical distributions depends on the sample size (larger sample size = more accurate) If the sample size is large and you have generated many simulations, the two methods should give essentially the same answer

Multiple Categories So far, we’ve learned how to do inference for a difference in means IF the categorical variable has only two categories Today, we’ll learn how to do hypothesis tests for a difference in means across multiple categories

Hypothesis Testing 1.State Hypotheses 2.Calculate a statistic, based on your sample data 3.Create a distribution of this statistic, as it would be observed if the null hypothesis were true 4.Measure how extreme your test statistic from (2) is, as compared to the distribution generated in (3) test statistic

Hypotheses To test for a difference in means across k groups:

Test Statistic Why can’t use the familiar formula to get the test statistic? More than one sample statistic More than one null value We need something a bit more complicated…

Difference in Means Whether or not two means are significantly different depends on How far apart the means are How much variability there is within each group

Difference in Means

Analysis of Variance Analysis of Variance (ANOVA) compares the variability between groups to the variability within groups Total Variability Variability Between Groups Variability Within Groups

Analysis of Variance If the groups are actually different, then a)the variability between groups should be higher than the variability within groups b)the variability within groups should be higher than the variability between groups

Discoveries for Today How to measure variability between groups? How to measure variability within groups? How to compare the two measures? How to determine significance?

Notation k = number of groups n j = number of units in group j n = overall number of units = n 1 + n 2 + … + n k

Discoveries for Today How to measure variability between groups? How to measure variability within groups? How to compare the two measures? How to determine significance?

Sums of Squares We will measure variability as sums of squared deviations (aka sums of squares) familiar?

Sums of Squares Total Variability Variability Between Groups Variability Within Groups overall mean data value i overall mean mean in group j i th data value in group j Sum over all data valuesSum over all groups Sum over all data values

Deviations Group 1 Group 2 Overall Mean Group 1 Mean

Sums of Squares Total Variability Variability Between Groups Variability Within Groups SST (Total sum of squares) SSG (sum of squares due to groups) SSE (“Error” sum of squares)

Source Groups Error Total df k-1 n-kn-k n-1 Sum of Squares SSG SSE SST Mean Square MSG = SSG/(k-1) MSE = SSE/(n-k) ANOVA Table The “mean square” is the sum of squares divided by the degrees of freedom variability average variability

Discoveries for Today How to measure variability between groups? How to measure variability within groups? How to compare the two measures? How to determine significance?

F-Statistic The F-statistic is a ratio of the average variability between groups to the average variability within groups The F-statistic is the test statistic for testing for a difference in means across more than 2 groups

Source Groups Error Total df k-1 n-kn-k n-1 Sum of Squares SSG SSE SST Mean Square MSG = SSG/(k-1) MSE = SSE/(n-k) F Statistic MSG MSE ANOVA Table

Jumping and Bone Density Does jumping improve bone density? 30 rats were randomized to three treatment groups: No jumping (10 rats - group 1) 30 cm jump (10 rats - group 2) 60 cm jump (10 rats - group 3) Rats performed 10 jumps per day, 5 days per week. Bone density was measured after 8 weeks.

Jumping and Bone Density mean sd n Control Lowjump Highjump

Jumping and Bone Density

Source Groups Error Total df k-1 = 2 n-k =27 n-1=29 Sum of Squares Mean Square /2 = /27 = F Statistic =7.98 ANOVA Table

Discoveries for Today How to measure variability between groups? How to measure variability within groups? How to compare the two measures? How to determine significance?

We have a test statistic. What else do we need to perform the hypothesis test? A distribution of the test statistic assuming H 0 is true How do we get this? Two options: 1)Simulation 2)Distributional Theory

F-statistic If there really is a difference between the groups, we would expect the F-statistic to be a)Higher than we would observe by random chance b)Lower than we would observe by random chance If the null hypothesis is true, what kind of F- statistics would we observe just by random chance?

Simulation Rerandomize (reallocate) the rats to treatment groups, keeping response values fixed Calculate the F-statistic Repeat this simulation many times to form a randomization distribution Calculate the p-value as the proportion as extreme or more extreme than the observed F-statistic

F-distribution p-value = Because a difference in groups would make the F-statistic higher, calculate probability in the upper tail

F-distribution

F-Distribution If the following conditions hold, 1.Sample sizes in each group are large (each n j ≥ 30) OR the data are relatively normally distributed 2.Variability is similar in all groups 3.The null hypothesis is true then the F-statistic follows an F-distribution The F-distribution has two degrees of freedom, one for the numerator of the ratio (k – 1) and one for the denominator (n – k)

Equal Variance The F-distribution assumes equal within group variability for each group As a rough rule of thumb, this assumption is violated if the standard deviation of one group is more than double the standard deviation of another group

F-distribution Can we use the F-distribution to calculate the p-value for the jumping and bone density F-statistic? a)Yes b)No c)I need more information mean sd n Control Lowjump Highjump

Jumping and Bone Density

F-distribution p-values 1)Online applet: 2)RStudio: > pf(7.98,2,27,lower.tail=FALSE) [1] ) TI-83: 2 nd  DISTR  9:Fcdf(  7.98, 9999, 2, 27

Source Groups Error Total df k-1 n-kn-k n-1 Sum of Squares SSG SSE SST Mean Square MSG = SSG/(k-1) MSE = SSE/(n-k) F Statistic MSG MSE p-value Use F k-1,n-k ANOVA Table

Source Groups Error Total df Sum of Squares Mean Square F- Statistic 7.98 ANOVA Table p-value We have strong evidence that jumping does increase bone density, at least in rats.

Study Hours by Class Year Is there a difference in the average hours spent studying per week by class year at Duke? (a)Yes (b)No (c)Cannot tell from this data (d)I didn’t finish

Source Groups Error Total df Sum of Squares Mean Square F- Statistic 1.24 ANOVA Table p-value 0.29

Summary Analysis of variance is used to test for a difference in means between groups by comparing the variability between groups to the variability within groups Sums of squares are used to measure variability The F-statistic is the ratio of average variability between groups to average variability within groups The F-statistic follows an F-distribution, if sample sizes are large (or data is normal), variability is equal across groups, and the null hypothesis is true