What Is Hypothesis Testing?

Slides:



Advertisements
Similar presentations
Chapter 12: Testing hypotheses about single means (z and t) Example: Suppose you have the hypothesis that UW undergrads have higher than the average IQ.
Advertisements

Statistics.  Statistically significant– When the P-value falls below the alpha level, we say that the tests is “statistically significant” at the alpha.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Econ 488 Lecture 5 – Hypothesis Testing Cameron Kaplan.
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved.
1 1 Slide © 2009 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
1 1 Slide STATISTICS FOR BUSINESS AND ECONOMICS Seventh Edition AndersonSweeneyWilliams Slides Prepared by John Loucks © 1999 ITP/South-Western College.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 9 Hypothesis Testing Developing Null and Alternative Hypotheses Developing Null and.
Introduction and Overview
4.3 Confidence Intervals -Using our CLM assumptions, we can construct CONFIDENCE INTERVALS or CONFIDENCE INTERVAL ESTIMATES of the form: -Given a significance.
The Use and Interpretation of the Constant Term
Copyright ©2011 Brooks/Cole, Cengage Learning Testing Hypotheses about Means Chapter 13.
Significance Testing Chapter 13 Victor Katch Kinesiology.
4.1 All rights reserved by Dr.Bill Wan Sing Hung - HKBU Lecture #4 Studenmund (2006): Chapter 5 Review of hypothesis testing Confidence Interval and estimation.
1 More Regression Information. 2 3 On the previous slide I have an Excel regression output. The example is the pizza sales we saw before. The first thing.
Pengujian Hipotesis Nilai Tengah Pertemuan 19 Matakuliah: I0134/Metode Statistika Tahun: 2007.
Topic 2: Statistical Concepts and Market Returns
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
BCOR 1020 Business Statistics Lecture 21 – April 8, 2008.
Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 7 th Edition Chapter 9 Hypothesis Testing: Single.
Inferences About Process Quality
Chapter 9 Hypothesis Testing.
Chapter 8 Introduction to Hypothesis Testing
1 Confidence Interval for Population Mean The case when the population standard deviation is unknown (the more common case).
Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 8 Tests of Hypotheses Based on a Single Sample.
Example 10.1 Experimenting with a New Pizza Style at the Pepperoni Pizza Restaurant Concepts in Hypothesis Testing.
AM Recitation 2/10/11.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Lecture Slides Elementary Statistics Twelfth Edition
Overview Definition Hypothesis
Confidence Intervals and Hypothesis Testing - II
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
Fundamentals of Hypothesis Testing: One-Sample Tests
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 2 – Slide 1 of 25 Chapter 11 Section 2 Inference about Two Means: Independent.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
1 1 Slide Slides Prepared by JOHN S. LOUCKS St. Edward’s University © 2002 South-Western/Thomson Learning.
LECTURE 19 THURSDAY, 14 April STA 291 Spring
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
1 1 Slide IS 310 – Business Statistics IS 310 Business Statistics CSU Long Beach.
Chapter 10 Hetero- skedasticity Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Pure Serial Correlation
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
1 Chapter 9 Hypothesis Testing. 2 Chapter Outline  Developing Null and Alternative Hypothesis  Type I and Type II Errors  Population Mean: Known 
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
Lecture 9 Chap 9-1 Chapter 2b Fundamentals of Hypothesis Testing: One-Sample Tests.
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Overview.
3-1 MGMG 522 : Session #3 Hypothesis Testing (Ch. 5)
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
2010, ECON Hypothesis Testing 1: Single Coefficient Review of hypothesis testing Testing single coefficient Interval estimation Objectives.
© Copyright McGraw-Hill 2004
Formulating the Hypothesis null hypothesis 4 The null hypothesis is a statement about the population value that will be tested. null hypothesis 4 The null.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Understanding Basic Statistics Fourth Edition By Brase and Brase Prepared by: Lynn Smith Gloucester County College Chapter Nine Hypothesis Testing.
Chapter 12 Tests of Hypotheses Means 12.1 Tests of Hypotheses 12.2 Significance of Tests 12.3 Tests concerning Means 12.4 Tests concerning Means(unknown.
Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 8 th Edition Chapter 9 Hypothesis Testing: Single.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Copyright © 2009 Pearson Education, Inc t LEARNING GOAL Understand when it is appropriate to use the Student t distribution rather than the normal.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Chapter 9 Introduction to the t Statistic
Chapter 9 Hypothesis Testing.
Presentation transcript:

What Is Hypothesis Testing? Hypothesis testing is used in a variety of settings The Food and Drug Administration (FDA), for example, tests new products before allowing their sale If the sample of people exposed to the new product shows some side effect significantly more frequently than would be expected to occur by chance, the FDA is likely to withhold approval of marketing that product Similarly, economists have been statistically testing various relationships, for example that between consumption and income Note here that while we cannot prove a given hypothesis (for example the existence of a given relationship), we often can reject a given hypothesis (again, for example, rejecting the existence of a given relationship) © 2011 Pearson Addison-Wesley. All rights reserved. 1

Classical Null and Alternative Hypotheses The researcher first states the hypotheses to be tested Here, we distinguish between the null and the alternative hypothesis: Null hypothesis (“H0”): the outcome that the researcher does not expect (almost always includes an equality sign) Alternative hypothesis (“HA”): the outcome the researcher does expect Example: H0: β ≤ 0 (the values you do not expect) HA: β > 0 (the values you do expect) © 2011 Pearson Addison-Wesley. All rights reserved. 2

Type I and Type II Errors Two types of errors possible in hypothesis testing: Type I: Rejecting a true null hypothesis Type II: Not rejecting a false null hypothesis Example: Suppose we have the following null and alternative hypotheses: H0: β ≤ 0 HA: β > 0 Even if the true β really is not positive, in any one sample we might still observe an estimate of β that is sufficiently positive to lead to the rejection of the null hypothesis This can be illustrated by Figure 5.1 © 2011 Pearson Addison-Wesley. All rights reserved. 3

Figure 5.1 Rejecting a True Null Hypothesis Is a Type I Error © 2011 Pearson Addison-Wesley. All rights reserved. 4

Type I and Type II Errors (cont.) Alternatively, it’s possible to obtain an estimate of β that is close enough to zero (or negative) to be considered “not significantly positive” Such a result may lead the researcher to “accept” the null hypothesis that β ≤ 0 when in truth β > 0 This is a Type II Error; we have failed to reject a false null hypothesis! This can be illustrated by Figure 5.2 © 2011 Pearson Addison-Wesley. All rights reserved. 5

Figure 5.2 Failure to Reject a False Null Hypothesis Is a Type II Error © 2011 Pearson Addison-Wesley. All rights reserved. 6

Decision Rules of Hypothesis Testing To test a hypothesis, we calculate a sample statistic that determines when the null hypothesis can be rejected depending on the magnitude of that sample statistic relative to a preselected critical value (which is found in a statistical table) This procedure is referred to as a decision rule The decision rule is formulated before regression estimates are obtained The range of possible values of the estimates is divided into two regions, an “acceptance” (really, non-rejection) region and a rejection region The critical value effectively separates the “acceptance”/non-rejection region from the rejection region when testing a null hypothesis Graphs of these “acceptance” and rejection regions are given in Figures 5.3 and 5.4 © 2011 Pearson Addison-Wesley. All rights reserved. 7

Figure 5.3 “Acceptance” and Rejection Regions for a One-Sided Test of β © 2011 Pearson Addison-Wesley. All rights reserved. 8

Figure 5.4 “Acceptance” and Rejection Regions for a Two-Sided Test of β © 2011 Pearson Addison-Wesley. All rights reserved. 9

The t-Test The t-test is the test that econometricians usually use to test hypotheses about individual regression slope coefficients Tests of more than one coefficient at a time (joint hypotheses) are typically done with the F-test, presented in Section 5.6 The appropriate test to use when the stochastic error term is normally distributed and when the variance of that distribution must be estimated Since these usually are the case, the use of the t-test for hypothesis testing has become standard practice in econometrics © 2011 Pearson Addison-Wesley. All rights reserved. 10

The t-Statistic For a typical multiple regression equation: (5.1) we can calculate t-values for each of the estimated coefficients Usually these are only calculated for the slope coefficients, though (see Section 7.1) Specifically, the t-statistic for the kth coefficient is: (5.2) © 2011 Pearson Addison-Wesley. All rights reserved. 11

The Critical t-Value and the t-Test Decision Rule To decide whether to reject or not to reject a null hypothesis based on a calculated t-value, we use a critical t-value A critical t-value is the value that distinguishes the “acceptance” region from the rejection region The critical t-value, tc, is selected from a t-table (see Statistical Table B-1 in the back of the book) depending on: whether the test is one-sided or two-sided, the level of Type I Error specified and the degrees of freedom (defined as the number of observations minus the number of coefficients estimated (including the constant) or N – K – 1) © 2011 Pearson Addison-Wesley. All rights reserved. 12

The Critical t-Value and the t-Test Decision Rule (cont.) The rule to apply when testing a single regression coefficient ends up being that you should: Reject H0 if |tk| > tc and if tk also has the sign implied by HA Do not reject H0 otherwise © 2011 Pearson Addison-Wesley. All rights reserved. 13

The Critical t-Value and the t-Test Decision Rule (cont.) Note that this decision rule works both for calculated t-values and critical t-values for one-sided hypotheses around zero (or another hypothesized value, S): H0: βk ≤ 0 H0: βk ≤ S HA: βk > 0 HA: βk > S H0: βk ≥ 0 H0: βk ≥ S HA: βk < 0 HA: βk < S © 2011 Pearson Addison-Wesley. All rights reserved. 14

The Critical t-Value and the t-Test Decision Rule (cont.) As well as for two-sided hypotheses around zero (or another hypothesized value, S): H0: βk = 0 H0: βk = S HA: βk ≠ 0 HA: βk ≠ S From Statistical Table B-1 the critical t-value for a one-tailed test at a given level of significance is exactly equal to the critical t-value for a two-tailed test at twice the level of significance of the one-tailed test—as also illustrated by Figure 5.5 © 2011 Pearson Addison-Wesley. All rights reserved. 15

Figure 5.5 One-Sided and Two-Sided t-Tests © 2011 Pearson Addison-Wesley. All rights reserved. 16

Choosing a Level of Significance The level of significance must be chosen before a critical value can be found, using Statistical Table B The level of significance indicates the probability of observing an estimated t-value greater than the critical t-value if the null hypothesis were correct It also measures the amount of Type I Error implied by a particular critical t-value Which level of significance is chosen? – 5 percent is recommended, unless you know something unusual about the relative costs of making Type I and Type II Errors © 2011 Pearson Addison-Wesley. All rights reserved. 17

Confidence Intervals A confidence interval is a range that contains the true value of an item a specified percentage of the time It is calculated using the estimated regression coefficient, the two-sided critical t-value and the standard error of the estimated coefficient as follows: (5.5) What’s the relationship between confidence intervals and two-sided hypothesis testing? If a hypothesized value fall within the confidence interval, then we cannot reject the null hypothesis © 2011 Pearson Addison-Wesley. All rights reserved. 18

p-Values This is an alternative to the t-test A p-value, or marginal significance level, is the probability of observing a t-score that size or larger (in absolute value) if the null hypothesis were true Graphically, it’s two times the area under the curve of the t-distribution between the absolute value of the actual t-score and infinity. In theory, we could find this by combing through pages and pages of statistical tables But we don’t have to, since we have EViews and Stata: these (and other) statistical software packages automatically give the p-values as part of the standard output! In light of all this, the p-value decision rule therefore is: Reject H0 if p-valueK < the level of significance and if has the sign implied by HA © 2011 Pearson Addison-Wesley. All rights reserved. 19

Examples of t-Tests: One-Sided The most common use of the one-sided t-test is to determine whether a regression coefficient is significantly different from zero (in the direction predicted by theory!) This involves four steps: 1. Set up the null and alternative hypothesis 2. Choose a level of significance and therefore a critical t-value 3. Run the regression and obtain an estimated t-value (or t-score) 4. Apply the decision rule by comparing calculated t-value with the critical t-value in order to reject or not reject the null hypothesis Let’s look at each step in more detail for a specific example: © 2011 Pearson Addison-Wesley. All rights reserved. 20

Examples of t-Tests: One-Sided (cont.) Consider the following simple model of the aggregate retail sales of new cars: (5.6) Where: Y = sales of new cars X1 = real disposable income X2 = average retail price of a new car adjusted by the consumer price index X3 = number of sports utility vehicles sold The four steps for this example then are as follows: © 2011 Pearson Addison-Wesley. All rights reserved. 21

Step 1: Set up the null and alternative hypotheses From equation 5.6, the one-sided hypotheses are set up as: H0: β1 ≤ 0 HA: β1 > 0 H0: β2 ≥ 0 HA: β2 < 0 H0: β3 ≥ 0 HA: β3 < 0 Remember that a t-test typically is not run on the estimate of the constant term β0 © 2011 Pearson Addison-Wesley. All rights reserved. 22

Step 2: Choose a level of significance and therefore a critical t-value Assume that you have considered the various costs involved in making Type I and Type II Errors and have chosen 5 percent as the level of significance There are 10 observations in the data set, and so there are 10 – 3 – 1 = 6 degrees of freedom At a 5-percent level of significance, the critical t-value, tc, can be found in Statistical Table B-1 to be 1.943 © 2011 Pearson Addison-Wesley. All rights reserved. 23

Step 3: Run the regression and obtain an estimated t-value Use the data (annual from 2000 to 2009) to run the regression on your OLS computer package Again, most statistical software packages automatically report the t-values Assume that in this case the t-values were 2.1, 5.6, and –0.1 for β1, β2, and β3, respectively © 2011 Pearson Addison-Wesley. All rights reserved. 24

Step 4: Apply the t–test decision rule As stated in Section 5.2, the decision rule for the t-test is to: Reject H0 if |tk| > tc and if tk also has the sign implied by HA In this example, this amounts to the following three conditions: For β1: Reject H0 if |2.1| > 1.943 and if 2.1 is positive. For β2: Reject H0 if |5.6| > 1.943 and if 5.6 is positive. For β3: Reject H0 if |–0.1| > 1.943 and if –0.1 is positive. Figure 5.6 illustrates all three of these outcomes © 2011 Pearson Addison-Wesley. All rights reserved. 25

Figure 5.6a One-Sided t-Tests of the Coefficients of the New Car Sales Model © 2011 Pearson Addison-Wesley. All rights reserved. 26

Figure 5.6b One-Sided t-Tests of the Coefficients of the New Car Sales Model © 2011 Pearson Addison-Wesley. All rights reserved. 27

Examples of t-Tests: Two-Sided The two-sided test is used when the hypotheses should be rejected if estimated coefficients are significantly different from zero, or a specific nonzero value, in either direction So, there are two cases: Two-sided tests of whether an estimated coefficient is significantly different from zero, and Two-sided tests of whether an estimated coefficient is significantly different from a specific nonzero value Let’s take an example to illustrate the first of these (the second case is merely a generalized case of this, see the textbook for details), using the Woody’s restaurant example in Chapter 3: © 2011 Pearson Addison-Wesley. All rights reserved. 28

Examples of t-Tests: Two-Sided (cont.) Again, in the Woody’s restaurant equation of Section 3.2, the impace of the average income of an area on the expected number of Woody’s customer’s in that area is ambiguous: A high-income neighborhood might have more total customers going out to dinner (positive sign), but those customers might decide to eat at a more formal restaurant that Woody’s (negative sign) The appropriate (two-sided) t-test therefore is: © 2011 Pearson Addison-Wesley. All rights reserved. 29

Figure 5.7 Two-Sided t-Test of the Coefficient of Income in the Woody’s Model © 2011 Pearson Addison-Wesley. All rights reserved. 30

Examples of t-Tests: Two-Sided (cont.) • The four steps are the same as in the one-sided case: Set up the null and alternative hypothesis H0: βk = 0 HA: βk ≠ 0 2. Choose a level of significance and therefore a critical t-value Keep the level at significance at 5 percent but this now must be distributed between two rejection regions for 29 degrees of freedom hence the correct critical t-value is 2.045 (found in Statistical Table B-1 for 29 degrees of freedom and a 5-percent, two-sided test) 3. Run the regression and obtain an estimated t-value: The t-value remains at 2.37 (from Equation 5.4) Apply the decision rule: For the two-sided case, this simplifies to: Reject H0 if |2.37| > 2.045; so, reject H0 © 2011 Pearson Addison-Wesley. All rights reserved. 31

Limitations of the t-Test With the t-values being automatically printed out by computer regression packages, there is reason to caution against potential improper use of the t-test: 1. The t-Test Does Not Test Theoretical Validity: If you regress the consumer price index on rainfall in a time-series regression and find strong statistical significance does that also mean that the underlying theory is valid? Of course not! © 2011 Pearson Addison-Wesley. All rights reserved. 32

Limitations of the t-Test 2. The t-Test Does Not Test “Importance”: The fact that one coefficient is “more statistically significant” than another does not mean that it is also more important in explaining the dependent variable—but merely that we have more evidence of the sign of the coefficient in question 3. The t-Test Is Not Intended for Tests of the Entire Population: From the definition of the t-score, given by Equation 5.2, it is seen that as the sample size approaches the population (whereby the standard error will approach zero—since the standard error decreases as N increases), the t-score will approach infinity! © 2011 Pearson Addison-Wesley. All rights reserved. 33

Key Terms from Chapter 5 Null hypothesis Alternative hypothesis Type I Error Level of significance Two-sided test Decision rule Critical value t-statistic Confidence interval p-value © 2011 Pearson Addison-Wesley. All rights reserved. 34