Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.

Slides:



Advertisements
Similar presentations
STA305 Spring 2014 This started with excerpts from STA2101f13
Advertisements

Factorial ANOVA More than one categorical explanatory variable.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Copyright © 2009 Pearson Education, Inc. Chapter 29 Multiple Regression.
Sample Size Power and other methods. Non-central Chisquare.
Inferential Statistics
Hypothesis Testing A hypothesis is a claim or statement about a property of a population (in our case, about the mean or a proportion of the population)
Chapter 12 Tests of Hypotheses Means 12.1 Tests of Hypotheses 12.2 Significance of Tests 12.3 Tests concerning Means 12.4 Tests concerning Means(unknown.
Analysis of variance (ANOVA)-the General Linear Model (GLM)
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
4.3 Confidence Intervals -Using our CLM assumptions, we can construct CONFIDENCE INTERVALS or CONFIDENCE INTERVAL ESTIMATES of the form: -Given a significance.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Analysis of Variance (ANOVA) Statistics for the Social Sciences Psychology 340 Spring 2010.
The Two Factor ANOVA © 2010 Pearson Prentice Hall. All rights reserved.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Lecture 14 – Thurs, Oct 23 Multiple Comparisons (Sections 6.3, 6.4). Next time: Simple linear regression (Sections )
Comparing Means.
Lecture 9: One Way ANOVA Between Subjects
Hypothesis Tests for Means The context “Statistical significance” Hypothesis tests and confidence intervals The steps Hypothesis Test statistic Distribution.
Chapter 11 Multiple Regression.
Chapter 3 Hypothesis Testing. Curriculum Object Specified the problem based the form of hypothesis Student can arrange for hypothesis step Analyze a problem.
One-way Between Groups Analysis of Variance
Lecture 12 One-way Analysis of Variance (Chapter 15.2)
Inferences About Process Quality
Ch. 14: The Multiple Regression Model building
Chapter 9 Hypothesis Testing.
Comparing Means.
BCOR 1020 Business Statistics
Today Concepts underlying inferential statistics
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
INFERENTIAL STATISTICS – Samples are only estimates of the population – Sample statistics will be slightly off from the true values of its population’s.
Chapter 12: Analysis of Variance
AM Recitation 2/10/11.
1 STATISTICAL HYPOTHESES AND THEIR VERIFICATION Kazimieras Pukėnas.
4-1 Statistical Inference The field of statistical inference consists of those methods used to make decisions or draw conclusions about a population.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.2 Estimating Differences.
Intermediate Applied Statistics STAT 460
Comparing Means. Anova F-test can be used to determine whether the expected responses at the t levels of an experimental factor differ from each other.
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Statistics 11 Confidence Interval Suppose you have a sample from a population You know the sample mean is an unbiased estimate of population mean Question:
ANOVA (Analysis of Variance) by Aziza Munir
Copyright © 2011 Pearson Education, Inc. Analysis of Variance Chapter 26.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Multiple Regression BPS chapter 28 © 2006 W.H. Freeman and Company.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
3-1 MGMG 522 : Session #3 Hypothesis Testing (Ch. 5)
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Categorical Independent Variables STA302 Fall 2013.
Ex St 801 Statistical Methods Inference about a Single Population Mean.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: One-way ANOVA Marshall University Genomics Core.
Copyright © Cengage Learning. All rights reserved. 12 Analysis of Variance.
Chapter 12 Introduction to Analysis of Variance PowerPoint Lecture Slides Essentials of Statistics for the Behavioral Sciences Eighth Edition by Frederick.
Statistics for the Social Sciences Psychology 340 Spring 2009 Analysis of Variance (ANOVA)
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
One-way ANOVA Example Analysis of Variance Hypotheses Model & Assumptions Analysis of Variance Multiple Comparisons Checking Assumptions.
Categorical Independent Variables STA302 Fall 2015.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Chapters Way Analysis of Variance - Completely Randomized Design.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
Copyright © 2014, 2011 Pearson Education, Inc. 1 Chapter 26 Analysis of Variance.
Chapter 12 Tests of Hypotheses Means 12.1 Tests of Hypotheses 12.2 Significance of Tests 12.3 Tests concerning Means 12.4 Tests concerning Means(unknown.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Class Seven Turn In: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 For Class Eight: Chapter 20: 18, 20, 24 Chapter 22: 34, 36 Read Chapters 23 &
Posthoc Comparisons finding the differences. Statistical Significance What does a statistically significant F statistic, in a Oneway ANOVA, tell us? What.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Interactions and Factorial ANOVA
Presentation transcript:

Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions

One-factor Analysis of variance Categorical Explanatory variable Quantitative Response variable p categories (groups) H 0 : All population means equal Normal conditional distributions Equal variances

Dummy Variables You have seen – Indicator dummy variables with intercept – Effect coding (with intercept) Cell means coding is also useful at times.

A common error Categorical explanatory variable with p categories p dummy variables (rather than p-1) And an intercept There are p population means represented by p+1 regression coefficients - not unique

But suppose you leave off the intercept Now there are p regression coefficients and p population means The correspondence is unique, and the model can be handy -- less algebra Called cell means coding

Cell means coding: p indicators and no intercept

Add a covariate: x 4

Contrasts

Overall F-test is a test of p-1 contrasts

In a one-factor design Mostly, what you want are tests of contrasts, Or collections of contrasts. You could do it with any dummy variable coding scheme. Cell means coding is often most convenient. With β=μ, test H 0 : Lβ=h And in fact you know how to get a confidence interval for any single contrast.

Multiple Comparisons Most hypothesis tests are designed to be carried out in isolation But if you do a lot of tests and all the null hypotheses are true, the chance of rejecting at least one of them can be a lot more than α. This is inflation of the Type I error rate. Otherwise known as the curse of a thousand t-tests. Multiple comparisons (sometimes called follow-up tests, post hoc tests, probing) try to offer a solution.

Multiple Comparisons Protect a family of tests against Type I error at some joint significance level α If all the null hypotheses are true, the probability of rejecting at least one is no more than α

Multiple comparison tests of contrasts in a one-factor design Usual null hypothesis is μ 1 = … = μ p. Usually do them after rejecting the initial null hypothesis with an ordinary F test. The big three are – Bonferroni – Tukey – Scheffé

Bonferroni Based on Bonferroni’s inequality Applies to any collection of k tests Assume all k null hypotheses are true Event A j is that null hypothesis j is rejected. Do the tests as usual Reject each H 0 if p < 0.05/k Or, adjust the p-values. Multiply them by k, and reject if pk < 0.05

Bonferroni Advantage: Flexible – Applies to any collection of hypothesis tests. Advantage: Easy to do. Disadvantage: Must know what all the tests are before seeing the data. Disadvantage: A little conservative; the true joint significance level is less than α.

Tukey (HSD) Based on the distribution of the largest mean minus the smallest. Applies only to pairwise comparisons of means If sample sizes are equal, it’s most powerful, period If sample sizes are not equal, it’s a bit conservative

Scheffé Find the usual critical value for the initial test. Multiply by p-1. This is the Scheffé critical value. Family includes all contrasts: Infinitely many! You don’t need to specify them in advance. Based on the union-intersection principle.

Aside Consider a 2-sided test, say of H 0 : β 3 =0 Reject if t>t α/2 or t<-t α/2 If you reject H 0, is there a formal basis for deciding whether β 3 >0 or β 3 <0? YES!

A family of 2 tests First do the initial 2-sided test. If H 0 is rejected, follow up with 2 one-sided tests: One with H 0 : β 3 ≤ 0, reject if if t>t α/2 The other with H 0 : β 3 ≥ 0, reject if if t<-t α/2 H 0 will be rejected with one follow-up if and only if the initial test rejects H 0 And you can draw a directional conclusion. This argument is valuable because it allows you to use common sense.

General principle The union of the critical regions is the critical region of the overall test The intersection of the null hypothesis regions is the null hypothesis region of the overall test. So if all the null hypotheses in the family are true, the parameter is in the null hypothesis region of the overall test. And the probability of rejecting at least one of the family null hypotheses is α, the significance level of the overall test.

Actually all you need is The null hypothesis region of the overall test must be contained in the intersection of the null hypothesis regions of the tests in the family, so that if the null hypothesis of the overall test is true, then all the null hypotheses in the family are true. The union of critical regions of tests in the family must be contained in the critical region of the overall (initial) test, so if any test in the family rejects H 0, the overall test does too. In this case the probability that at least one test in the family will wrongly reject H 0 is ≤ α.

Scheffé Follow-up tests cannot reject H 0 if the initial F-test does not. Not quite true of Bonferroni and Tukey. If the initial test (of p-1 contrasts) rejects H 0, there is a contrast for which the Scheffé test will reject H 0 (not necessarily a pairwise comparison). Adjusted p-value is the tail area beyond F/(p-1)

Which method should you use? If the sample sizes are nearly equal and you are only interested in pairwise comparisons, use Tukey because it's most powerful If the sample sizes are not close to equal and you are only interested in pairwise comparisons, there is (amazingly) no harm in applying all three methods and picking the one that gives you the greatest number of significant results. (It’s okay because this choice could be determined in advance based on number of treatments, α and the sample sizes.)

If you are interested in contrasts that go beyond pairwise comparisons and you can specify all of them before seeing the data, Bonferroni is almost always more powerful than Scheffé. (Tukey is out.) If you want lots of special contrasts but you don't know in advance exactly what they all are, Scheffé is the only honest way to go, unless you have a separate replication data set.

Interactions Interaction between independent variables means “It depends.” Relationship between one explanatory variable and the response variable depends on the value of the other explanatory variable. Can have – Quantitative by quantitative – Quantitative by categorical – Categorical by categorical

Quantitative by Quantitative For fixed x 2 Both slope and intercept depend on value of x 2 And for fixed x 1, slope and intercept relating x 2 to E(Y) depend on the value of x 1

Quantitative by Categorical One regression line for each category. Interaction means slopes are not equal Form a product of quantitative variable by each dummy variable for the categorical variable For example, three treatments and one covariate: x 1 is the covariate and x 2, x 3 are dummy variables

General principle Interaction between A and B means – Relationship of A to Y depends on value of B – Relationship of B to Y depends on value of A The two statements are formally equivalent

Make a table

What null hypothesis would you test for Equal slopes Comparing slopes for group one vs three Comparing slopes for group one vs two Equal regressions Interaction between group and x 1

What to do if H 0 : β 4 =β 5 =0 is rejected How do you test Group “controlling” for x 1 ? A good choice is to set x 1 to its sample mean, and compare treatments at that point. How about setting x 1 to sample mean of the group (3 different values)? With random assignment to Group, all three means just estimate E(X 1 ), and the mean of all the x 1 values is a better estimate.

Categorical by Categorical Soon But first, an example of multiple comparisons.