More About Type I and Type II Errors. O.J. Simpson trial: the situation O.J. is assumed innocent. Evidence collected: size 12 Bruno Magli bloody footprint,

Slides:



Advertisements
Similar presentations
Statistics Hypothesis Testing.
Advertisements

Minimizing Chance of Type I and Type II Errors
Hypothesis Testing An introduction. Big picture Use a random sample to learn something about a larger population.
CHAPTER 21 More About Tests: “What Can Go Wrong?”.
Chapter 12 Tests of Hypotheses Means 12.1 Tests of Hypotheses 12.2 Significance of Tests 12.3 Tests concerning Means 12.4 Tests concerning Means(unknown.
Inference Sampling distributions Hypothesis testing.
Last Time (Sampling &) Estimation Confidence Intervals Started Hypothesis Testing.
Our goal is to assess the evidence provided by the data in favor of some claim about the population. Section 6.2Tests of Significance.
Chapter 10 Section 2 Hypothesis Tests for a Population Mean
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 11 Introduction to Hypothesis Testing.
Section 7.1 Hypothesis Testing: Hypothesis: Null Hypothesis (H 0 ): Alternative Hypothesis (H 1 ): a statistical analysis used to decide which of two competing.
Fundamentals of Hypothesis Testing. Identify the Population Assume the population mean TV sets is 3. (Null Hypothesis) REJECT Compute the Sample Mean.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Hypothesis Testing: Type II Error and Power.
1/55 EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2008 Chapter 10 Hypothesis Testing.
Lecture 2: Thu, Jan 16 Hypothesis Testing – Introduction (Ch 11)
BCOR 1020 Business Statistics Lecture 21 – April 8, 2008.
Chapter 9 Hypothesis Testing.
BCOR 1020 Business Statistics Lecture 20 – April 3, 2008.
Ch. 9 Fundamental of Hypothesis Testing
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 11 Introduction to Hypothesis Testing.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Hypothesis Testing A hypothesis is a conjecture about a population. Typically, these hypotheses will be stated in terms of a parameter such as  (mean)
Chapter 10 Hypothesis Testing
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 9 Introduction to Hypothesis Testing.
Fundamentals of Hypothesis Testing: One-Sample Tests
Section 9.1 Introduction to Statistical Tests 9.1 / 1 Hypothesis testing is used to make decisions concerning the value of a parameter.
Inference in practice BPS chapter 16 © 2006 W.H. Freeman and Company.
1 Power and Sample Size in Testing One Mean. 2 Type I & Type II Error Type I Error: reject the null hypothesis when it is true. The probability of a Type.
Statistical Inference Decision Making (Hypothesis Testing) Decision Making (Hypothesis Testing) A formal method for decision making in the presence of.
Hypothesis Testing: One Sample Cases. Outline: – The logic of hypothesis testing – The Five-Step Model – Hypothesis testing for single sample means (z.
Chapter 10 Hypothesis Testing
Chapter 4 Introduction to Hypothesis Testing Introduction to Hypothesis Testing.
LECTURE 19 THURSDAY, 14 April STA 291 Spring
Hypothesis Testing: Statistical Approach (6 basic steps) 1.Formulate Null (H o ) and Alternative (H A ) Hypotheses. 2.Determine test criteria – what test.
A Broad Overview of Key Statistical Concepts. An Overview of Our Review Populations and samples Parameters and statistics Confidence intervals Hypothesis.
Topic 7 - Hypothesis tests based on a single sample Sampling distribution of the sample mean - pages Basics of hypothesis testing -
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
Testing of Hypothesis Fundamentals of Hypothesis.
Lecture 16 Dustin Lueker.  Charlie claims that the average commute of his coworkers is 15 miles. Stu believes it is greater than that so he decides to.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Significance Test A claim is made. Is the claim true? Is the claim false?
9.1 – The Basics Ch 9 – Testing a Claim. Jack’s a candidate for mayor against 1 other person, so he must gain at least 50% of the votes. Based on a poll.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Lecture 18 Dustin Lueker.  A way of statistically testing a hypothesis by comparing the data to values predicted by the hypothesis ◦ Data that fall far.
Statistical Inference An introduction. Big picture Use a random sample to learn something about a larger population.
Copyright ©2013 Pearson Education, Inc. publishing as Prentice Hall 9-1 σ σ.
Lecture 17 Dustin Lueker.  A way of statistically testing a hypothesis by comparing the data to values predicted by the hypothesis ◦ Data that fall far.
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
1 Hypothesis Testing A criminal trial is an example of hypothesis testing. In a trial a jury must decide between two hypotheses. The null hypothesis is.
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
Math 4030 – 9a Introduction to Hypothesis Testing
Hypothesis Testing. “Not Guilty” In criminal proceedings in U.S. courts the defendant is presumed innocent until proven guilty and the prosecutor must.
A review of key statistical concepts. An overview of the review Populations and parameters Samples and statistics Confidence intervals Hypothesis testing.
Hypothesis Testing Errors. Hypothesis Testing Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean.
Section 9.1(re-visited)  Making Sense of Statistical Significance  Inference as Decision.
Statistical Inference Making decisions regarding the population base on a sample.
What is a Hypothesis? A hypothesis is a claim (assumption) about the population parameter Examples of parameters are population mean or proportion The.
Chapter 12 Tests of Hypotheses Means 12.1 Tests of Hypotheses 12.2 Significance of Tests 12.3 Tests concerning Means 12.4 Tests concerning Means(unknown.
6.2 Large Sample Significance Tests for a Mean “The reason students have trouble understanding hypothesis testing may be that they are trying to think.”
DSCI 346 Yamasaki Lecture 1 Hypothesis Tests for Single Population DSCI 346 Lecture 1 (22 pages)1.
Section Testing a Proportion
Hypothesis Testing: Hypotheses
P-value Approach for Test Conclusion
Chapter 9 Hypothesis Testing.
Daniela Stan Raicu School of CTI, DePaul University
Type I and Type II Errors
STA 291 Spring 2008 Lecture 17 Dustin Lueker.
Presentation transcript:

More About Type I and Type II Errors

O.J. Simpson trial: the situation O.J. is assumed innocent. Evidence collected: size 12 Bruno Magli bloody footprint, bloody glove, blood spots on white Ford Bronco, the knock on the wall, DNA evidence from above, motive(?), etc…

O.J. Simpson trial: jury decisions In criminal trial: The evidence does not warrant rejecting the assumption of innocence. Behave as if O.J. is innocent. In civil trial: The evidence warrants rejecting the assumption of innocence. Behave as if O.J. is guilty. Was an error made in either trial?

Errors in Trials If O.J. is innocent, then an error was made in the civil trial. If O.J. is guilty, then an error was made in the criminal trial.

Errors in Hypothesis Testing

Definitions: Types of Errors Type I error: The null hypothesis is rejected when it is true. Type II error: The null hypothesis is not rejected when it is false. There is always a chance of making one of these errors. We’ll want to minimize the chance of doing so!

Example: Grade inflation? Is there evidence to suggest the mean GPA of college undergraduate students exceeds 2.7? H 0 : μ = 2.7 H A : μ > 2.7 Random sample of students n = 36 s = 0.6 and Data Decision Rule Set significance level α = If p-value < 0.05, reject null hypothesis.

If Reject null since p-value is (just barely!) smaller then Let’s consider what our conclusion is based upon different observed sample means…

Reject null since p-value is smaller then If

Reject null since p-value is smaller then If

Alternative Decision Rule “Reject if p-value  0.05” is equivalent to “Reject if the sample average, X-bar, is larger than 2.865” is called “rejection region.”

Type I Error

Minimize chance of Type I error... … by making significance level  small. Common values are  = 0.01, 0.05, or “How small” depends on seriousness of Type I error. Decision is not a statistical one but a practical one.

P(Type I Error) in trials Criminal trials: “Beyond a reasonable doubt”. 12 of 12 jurors must unanimously vote guilty. Significance level  set at 0.001, say. Civil trials: “Preponderance of evidence.” 9 out of 12 jurors must vote guilty. Significance level  set at 0.10, say.

Example: Serious Type I Error New Drug A is supposed to reduce diastolic blood pressure by more than 15 mm Hg. H 0 : μ = 15 versus H A : μ > 15 Drug A can have serious side effects, so don’t want patients on it unless μ > 15. Implication of Type I error: Expose patients to serious side effects without other benefit. Set  = P(Type I error) to be small  0.01

Example: Not so serious Type I Error Grade inflation? H 0 : μ = 2.7 vs. H A : μ > 2.7 Type I error: claim average GPA is more than 2.7 when it really isn’t. Implication: Instructors grade harder. Students get unhappy. Set  = P(Type I error) at, say, 0.10.

Type II Error and Power Type II Error is made when we fail to reject the null when the alternative is true. Want to minimize P(Type II Error). Now, if alternative H A is true: –P(reject|H A is true) + P(not reject|H A is true) =1 –“Power” + P(Type II error) = 1 –“Power” = 1 - P(Type II error)

Type II Error and Power “Power” of a test is the probability of rejecting null when alternative is true. “Power” = 1 - P(Type II error) To minimize the P(Type II error), we equivalently want to maximize power. But power depends on the value under the alternative hypothesis...

Type II Error and Power (Alternative is true)

Power Power is probability, so number between 0 and 1. 0 is bad! 1 is good! Need to make power as high as possible.

Maximizing Power … The farther apart the actual mean is from the mean specified in the null, the higher the power. The higher the significance level , the higher the P(Type I error), the higher the power. The smaller the standard deviation, the higher the power. The larger the sample, the higher the power.

That is, factors affecting power... Difference between value under the null and the actual value P(Type I error) =  Standard deviation Sample size

Strategy for designing a good hypothesis test Use pilot study to estimate std. deviation. Specify . Typically 0.01 to Decide what a meaningful difference would be between the mean in the null and the actual mean. Decide power. Typically 0.80 to Use software to determine sample size.

Using JMP to Determine Sample Size – DOE > Sample Size and Power

Using JMP to Determine Sample Size – One Sample Mean P(Type I Error ) =  Error Std Dev = “guessimate” for standard deviation (s or  Enter values for one or two of the quantities: 1) Difference to detect  H o mean – H A mean| = |    2) Sample Size = n A true) =  3) Power = P(Reject H o |H A true) = 

Using JMP to Determine Sample Size – DOE > Sample Size and Power For  s =.60 and leaving Power and Sample Size empty we obtain a plot of Power vs. Sample Size (n). Here we can see:

Using JMP to Determine Sample Size – DOE > Sample Size and Power (JMP Demo)

If sample is too small... … the power can be too low to identify even large meaningful differences between the null and alternative values. –Determine sample size in advance of conducting study. –Don’t believe the “fail-to-reject-results” of a study based on a small sample.

If sample is really large... … the power can be extremely high for identifying even meaningless differences between the null and alternative values. –In addition to performing hypothesis tests, use a confidence interval to estimate the actual population value. –If a study reports a “reject result,” ask how much different?

The moral of the story as researcher Always determine how many measurements you need to take in order to have high enough power to achieve your study goals. If you don’t know how to determine sample size, ask a statistical consultant to help you.

The moral of the story as reviewer When interpreting the results of a study, always take into account the sample size.