Type I and Type II Errors

Slides:



Advertisements
Similar presentations
Introduction to Hypothesis Testing
Advertisements

Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Inference Sampling distributions Hypothesis testing.
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
HYPOTHESIS TESTING Four Steps Statistical Significance Outcomes Sampling Distributions.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
9-1 Hypothesis Testing Statistical Hypotheses Statistical hypothesis testing and confidence interval estimation of parameters are the fundamental.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
Introduction to Hypothesis Testing CJ 526 Statistical Analysis in Criminal Justice.
© 2013 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Introductory Statistics: Exploring the World through.
Introduction to Hypothesis Testing CJ 526 Statistical Analysis in Criminal Justice.
Chapter 9 Hypothesis Testing.
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
Major Points Formal Tests of Mean Differences Review of Concepts: Means, Standard Deviations, Standard Errors, Type I errors New Concepts: One and Two.
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
INFERENTIAL STATISTICS – Samples are only estimates of the population – Sample statistics will be slightly off from the true values of its population’s.
Choosing Statistical Procedures
AM Recitation 2/10/11.
Testing Hypotheses I Lesson 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics n Inferential Statistics.
Hypothesis testing is used to make decisions concerning the value of a parameter.
Introduction to Hypothesis Testing for μ Research Problem: Infant Touch Intervention Designed to increase child growth/weight Weight at age 2: Known population:
Sampling Distributions and Hypothesis Testing. 2 Major Points An example An example Sampling distribution Sampling distribution Hypothesis testing Hypothesis.
Section 9.1 Introduction to Statistical Tests 9.1 / 1 Hypothesis testing is used to make decisions concerning the value of a parameter.
Chapter 8 Hypothesis Testing I. Chapter Outline  An Overview of Hypothesis Testing  The Five-Step Model for Hypothesis Testing  One-Tailed and Two-Tailed.
Chapter 9 Hypothesis Testing II: two samples Test of significance for sample means (large samples) The difference between “statistical significance” and.
Chapter 8 Introduction to Hypothesis Testing
9-1 Hypothesis Testing Statistical Hypotheses Definition Statistical hypothesis testing and confidence interval estimation of parameters are.
Learning Objectives In this chapter you will learn about the t-test and its distribution t-test for related samples t-test for independent samples hypothesis.
Biostatistics Class 6 Hypothesis Testing: One-Sample Inference 2/29/2000.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Education 793 Class Notes Decisions, Error and Power Presentation 8.
© Copyright McGraw-Hill 2004
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Introduction to Testing a Hypothesis Testing a treatment Descriptive statistics cannot determine if differences are due to chance. Sampling error means.
Education 793 Class Notes Inference and Hypothesis Testing Using the Normal Distribution 8 October 2003.
1 Testing Statistical Hypothesis The One Sample t-Test Heibatollah Baghi, and Mastee Badii.
Hypothesis Testing. Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean μ = 120 and variance σ.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Chapter 9 Introduction to the t Statistic
Lecture Slides Elementary Statistics Twelfth Edition
Inference for a Single Population Proportion (p)
Chapter Nine Hypothesis Testing.
HYPOTHESIS TESTING.
Logic of Hypothesis Testing
More on Inference.
CHAPTER 9 Testing a Claim
Hypothesis Testing: One Sample Cases
Unit 5: Hypothesis Testing
Inference and Tests of Hypotheses
CHAPTER 9 Testing a Claim
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
© LOUIS COHEN, LAWRENCE MANION AND KEITH MORRISON
Hypothesis Testing: Hypotheses
Chapter 9 Hypothesis Testing.
More on Inference.
9 Tests of Hypotheses for a Single Sample CHAPTER OUTLINE
Chapter 9 Hypothesis Testing.
Chapter Nine Part 1 (Sections 9.1 & 9.2) Hypothesis Testing
Significance Tests: The Basics
Hypothesis Testing.
Chapter 3 Probability Sampling Theory Hypothesis Testing.
Psych 231: Research Methods in Psychology
CHAPTER 9 Testing a Claim
Chapter 9: Testing a Claim
Psych 231: Research Methods in Psychology
Testing Hypotheses I Lesson 9.
CHAPTER 9 Testing a Claim
CHAPTER 9 Testing a Claim
Presentation transcript:

Type I and Type II Errors PUBH 6125/8125 Week 8

Objectives Review the use of statistics to estimate population parameters Review hypothesis testing to determine the significance of statistics Discuss the potential for chance and chance error in estimation Identify the 2 types of error: Type I and Type II Associate the types of error with the calculation of power and sample size

Parameter vs. Statistic Based on the population An absolute number Example: Data in the form of birth certificates is collected on every person born in the U.S. Therefore, Natality Records are population level data and all information computed using that data are parameters. Based on a sample of the population An estimate of the parameter it represents Example: Data in National Surveys such as NHANES or BRFSS is collected from a random sampling of the U.S. population. All information computed using that data are statistics.

Hypothesis Testing You have seen how to use statistics to estimate the 95% confidence interval of a variable and also how to interpret this estimate to learn something about the population the sample was drawn from You have also seen that testing a hypothesis is another way to use sample data to learn about your population In both cases these cases, statistic is an estimate and so has a possibility of error built into it The null hypothesis is frequently represented by H0 and is the hypothesis you are testing The alternative hypothesis is HA or H1 You test a null hypothesis of this sort by drawing a random sample from the population in question and calculating an appropriate statistic on the variables of interest If you obtain a value of the statistic that would occur rarely through chance alone if the null hypothesis is really true, you would have reason to reject that hypothesis

The Null Hypothesis The null hypothesis is often quite different from the research hypothesis It is usually the theory that no association exists between the independent and dependent variables In the null hypothesis you make an assumption that the results observed are no different from those expected on the basis of chance alone You write the null hypothesis so that it is testable – you can reject or fail to reject the null hypothesis based on statistical significance

Null vs. Alternative Hypothesis Example: Null Hypothesis H0: Mean blood pressure on drug A = Mean blood pressure on drug B, as measured 6 hours after start of treatment Alternative Hypothesis H1: Mean blood pressure on drug A < Mean blood pressure on drug B, as measured 6 hours after start of treatment It would be a better hypothesis if you can specify how much lower mean blood pressure should be to be clinically significant, example: 5 mmHg

Failure to Reject When testing the null hypothesis we first assume it is true – so we either reject or fail to reject it based on the analysis. This is based on the philosophical idea that you cannot prove a negative – you can only disprove or fail to disprove it. We compare the outcome we observed to the outcome that could be expected by chance alone. Sampling only allows us to estimate parameters and this means there is a possibility of error. If the probability of obtaining, by chance, an association of two or more variables as strong as the one observed, is very small, we can reject the null hypothesis. If that probability is large, we fail to reject the null hypothesis.

Error and Chance Error Error is a false or mistaken result obtained in a study or experiment. There are two main types of error: Random error: just as it says this error is random and not connected in any way to the methodology of the study or the way the population was sampled. This error is generally regarded as due to CHANCE. Systematic error: this type of error often has a recognizable source, either related to the method or to the sampling. This is also often called a bias. An example is selection bias which is a systematic error that results because of how you sampled your population. Chance error can be further described as either Type I or Type II, based on what effect the error has on your interpretation of the statistical results.

Types of Chance Error Type I Type II The chance or probability that you will reject a null hypothesis that should not have been rejected. This will result in you deciding two groups are different or two variables are related when they really are not. The probability of a Type I error is called alpha (α). The chance or probability that you will not reject a null hypothesis when it should have been rejected. This will result in you deciding two groups are not different or two variables are not related when they really are. The probability of a Type II error is called beta (β).

Examples Null Hypothesis is True Null Hypothesis is False Reject the Null Hypothesis Type I Error (α) Correct Decision Do Not Reject the Null Hypothesis Type II Error (β) You sample two populations and test the null hypothesis that population A has the same rate of bladder cancer than population B. You compare your samples and reject the null hypothesis and state that the rates are different. In reality the rates are the same but you made a Type I error due to the way you sampled the two populations. You sample two populations and test the null hypothesis that population A has the same rate of bladder cancer than population B. You compare your samples and fail to reject the null hypothesis and state that the rates are not different. In reality the rates are different but you made a Type II error because your samples were not large enough.

Type II Error and Power In the previous slide’s second example you learned that you can have a Type II error if your sample size is not large enough. Power, which is the ability to detect differences if they exist, is the inverse function of Type II error or β. Power = 1 – β. A typical minimum for power is 80%, so the largest preset for β is 0.20. As with α, smaller is better as it allows less opportunity for chance error. In the next slide you will see how your preset values for α and β affect your required sample size. Sample size is very important because with an underpowered study you could fail to reject a null hypothesis that should be rejected.

Type I and II Error and Sample Size For a Continuous (Normal) Distribution: 2n = 4(Z+ Z)2 2 2 Where Z is the value of Z at the preset  Z is the value of Z at the preset  2 is the pooled variance of the samples  is the clinically significant effect size you wish to observe (difference between the means) You do not need to know the above formula or use it during this class to calculate your sample size. What you do need to know is that both the preset α and the preset β are directly related to the sample size. As you set those lower, the Z value associated with them gets higher (look in the book at the Z tables if you doubt this) and therefore the sample size required gets larger as well. The more you want to protect against the possibility of error, the larger your sample will need to be.