Chapter 8 Introduction to Statistical Inferences.

Slides:



Advertisements
Similar presentations
Statistics.  Statistically significant– When the P-value falls below the alpha level, we say that the tests is “statistically significant” at the alpha.
Advertisements

Chapter 9 Hypothesis Testing Understandable Statistics Ninth Edition
Chapter 12 Tests of Hypotheses Means 12.1 Tests of Hypotheses 12.2 Significance of Tests 12.3 Tests concerning Means 12.4 Tests concerning Means(unknown.
Testing Hypotheses About Proportions Chapter 20. Hypotheses Hypotheses are working models that we adopt temporarily. Our starting hypothesis is called.
Statistical Techniques I EXST7005 Lets go Power and Types of Errors.
Significance Testing Chapter 13 Victor Katch Kinesiology.
Cal State Northridge  320 Ainsworth Sampling Distributions and Hypothesis Testing.
1/55 EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2008 Chapter 10 Hypothesis Testing.
Introduction to Hypothesis Testing
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 8-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
8-2 Basics of Hypothesis Testing
Inferences About Process Quality
Ch. 9 Fundamental of Hypothesis Testing
Chapter 8 Introduction to Hypothesis Testing
BCOR 1020 Business Statistics
Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
Hypothesis Testing:.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Overview of Statistical Hypothesis Testing: The z-Test
Chapter 10 Hypothesis Testing
Overview Definition Hypothesis
Confidence Intervals and Hypothesis Testing - II
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 9 Introduction to Hypothesis Testing.
Chapter 8 Hypothesis testing 1. ▪Along with estimation, hypothesis testing is one of the major fields of statistical inference ▪In estimation, we: –don’t.
Fundamentals of Hypothesis Testing: One-Sample Tests
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap th Lesson Introduction to Hypothesis Testing.
More About Significance Tests
Chapter 8 Hypothesis Tests. Hypothesis Testing We now begin the phase of this course that discusses the highest achievement of statistics. Statistics,
Lecture 7 Introduction to Hypothesis Testing. Lecture Goals After completing this lecture, you should be able to: Formulate null and alternative hypotheses.
Introduction to Hypothesis Testing: One Population Value Chapter 8 Handout.
6.1 - One Sample One Sample  Mean μ, Variance σ 2, Proportion π Two Samples Two Samples  Means, Variances, Proportions μ 1 vs. μ 2.
Essential Statistics Chapter 131 Introduction to Inference.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
Chapter 20 Testing hypotheses about proportions
Lecture 16 Dustin Lueker.  Charlie claims that the average commute of his coworkers is 15 miles. Stu believes it is greater than that so he decides to.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 20 Testing Hypotheses About Proportions.
Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
Statistical Inference Statistical Inference involves estimating a population parameter (mean) from a sample that is taken from the population. Inference.
Introduction to Inferece BPS chapter 14 © 2010 W.H. Freeman and Company.
Copyright © 2010, 2007, 2004 Pearson Education, Inc Section 8-2 Basics of Hypothesis Testing.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Fundamentals of Hypothesis Testing: One-Sample Tests Statistics.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
Chap 8-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 8 Introduction to Hypothesis.
Chapter 221 What Is a Test of Significance?. Chapter 222 Thought Question 1 The defendant in a court case is either guilty or innocent. Which of these.
Lecture 9 Chap 9-1 Chapter 2b Fundamentals of Hypothesis Testing: One-Sample Tests.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Chapter 20 Testing Hypothesis about proportions
Lecture 18 Dustin Lueker.  A way of statistically testing a hypothesis by comparing the data to values predicted by the hypothesis ◦ Data that fall far.
Ch 10 – Intro To Inference 10.1: Estimating with Confidence 10.2 Tests of Significance 10.3 Making Sense of Statistical Significance 10.4 Inference as.
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Overview.
Copyright ©2013 Pearson Education, Inc. publishing as Prentice Hall 9-1 σ σ.
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
1 Hypothesis Testing A criminal trial is an example of hypothesis testing. In a trial a jury must decide between two hypotheses. The null hypothesis is.
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
© Copyright McGraw-Hill 2004
Statistical Techniques
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Understanding Basic Statistics Fourth Edition By Brase and Brase Prepared by: Lynn Smith Gloucester County College Chapter Nine Hypothesis Testing.
1 Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
One Sample Inf-1 In statistical testing, we use deductive reasoning to specify what should happen if the conjecture or null hypothesis is true. A study.
Chapter 12 Tests of Hypotheses Means 12.1 Tests of Hypotheses 12.2 Significance of Tests 12.3 Tests concerning Means 12.4 Tests concerning Means(unknown.
Slide 20-1 Copyright © 2004 Pearson Education, Inc.
Hypothesis Tests for 1-Proportion Presentation 9.
Copyright © 2009 Pearson Education, Inc t LEARNING GOAL Understand when it is appropriate to use the Student t distribution rather than the normal.
Chapter 9 Hypothesis Testing Understanding Basic Statistics Fifth Edition By Brase and Brase Prepared by Jon Booze.
Chapter Nine Hypothesis Testing.
Presentation transcript:

Chapter 8 Introduction to Statistical Inferences

Parameters and Statistics A parameter is a numeric characteristic of a population or distribution, usually symbolized by a Greek letter, such as μ, the population mean. Inferential Statistics uses sample information to estimate parameters. A Statistic is a number calculated from data. There are usually statistics that do the same job for samples that the parameters do for populations, such as, the sample mean.

Using Samples for Estimation μ Population (unknown parameter) Sample (known statistic) estimate

The Idea of Estimation We want to find a way to estimate the population parameters. We only have information from a sample, available in the form of statistics. The sample mean,, is an estimator of the population mean, μ. This is called a “point estimate” because it is one point, or a single value.

Interval Estimation There is variation in, since it is a random variable calculated from data. A point estimate doesn’t reveal anything about how much the estimate varies. An interval estimate gives a range of values that is likely to contain the parameter. Intervals are often reported in polls, such as “56% ±4% favor candidate A.” This suggests we are not sure it is exactly 56%, but we are quite sure that it is between 52% and 60%. 56% is the point estimate, whereas (52%, 60%) is the interval estimate.

The Confidence Interval A confidence interval is a special interval estimate involving a percent, called the confidence level. The confidence level tells how often, if samples were repeatedly taken, the interval estimate would surround the true parameter. We can use this notation: (L,U) or (LCL,UCL). L and U stand for Lower and Upper endpoints. The longer versions, LCL and UCL, stand for “Lower Confidence Limit” and “Upper Confidence Limit.” This interval is built around the point estimate.

Theory of Confidence Intervals Alpha (α) represents the probability that when the sample is taken, the calculated CI will miss the parameter. The confidence level is given by (1-α)×100%, and used to name the interval, so for example, we may have “a 90% CI for μ.” After sampling, we say that we are, for example, “90% confident that we have captured the true parameter.” (There is no probability at this point. Either we did or we didn’t, but we don’t know.)

How to Calculate CI’s There are many variations, but most CI’s have the following basic structure: P ± TS –Where P is the parameter estimate, –T is a “table” value equal to the number of standard deviations needed for the confidence level, –and S is the standard deviation of the estimate. The quantity TS is also called the “Error Bound” (E) or “Margin of Error.” The CI should be written as (L,U) where L= P-TS, and U= P+TS.

A Confidence Interval for μ If σ is known, and the population is normally distributed, or n>30 (so that we can say is approximately normally distiributed), gives the endpoints for a (1- α)100% CI for μ Note how this corresponds to the P ± TS formula given earlier.

Distribution Details What is ? –α is the significance level, P(CI will miss) –The subscript on z refers to the upper tail probability, that is, P(Z>z). –To find this value in the table, look up the z-value for a probability of.5-α/2.

Hypothesis Tests So far, we have discussed estimating parameters. For example, use a sample mean to estimate μ, giving both a point estimate and a CI. Now we take a different approach. Suppose we have an existing belief about the value of μ. This could come from previous research, or it could be a standard that needs to be met. Examples: –Previous corn hybrids have achieved 100 bu/acre. We want to show that our new hybrid does better. –Advertising claims have been made that there are 20 chips in every chocolate chip cookie. Support or refute this claim.

Framing the Test We start with a null hypothesis. This represents the status quo, or the conclusion if our test cannot prove anything. The null hypothesis is denoted by H 0 : μ=μ 0 where μ 0 corresponds to the current belief or status quo. (The equal sign could be replaced with an inequality if appropriate.) Example: –In the corn problem, if our hybrid is not better, it doesn’t beat the previous yield achievement of 100 bu/acre. Then we have H 0 : μ=100 or possibly H 0 : μ≤100. –In the cookie problem, if the advertising claims are correct, we have H 0 : μ=20 or possibly H 0 : μ≥20. Notice the choice of null hypothesis is not based on what we hope to prove, but on what is currently accepted.

Framing the Test The alternative hypothesis is the result that you will get if your research proves something is different from status quo or from what is expected. It is denoted by H a : μ≠μ 0. Sometimes there is more than one alternative, so we can write H 1 : μ≠μ 0, H 2 : μ>μ 0, and H 3 : μ<μ 0. In the corn problem, if our yield is more than 100 we have proved that our hybrid is better, so the alternative H a : μ>100 is appropriate.

Framing the Test For the cookie example, if there are less than 20 chips per cookie, the advertisers are wrong and possibly guilty of false advertising, so we want to prove H a : μ<20. A jar of peanut butter is supposed to have 16 oz in it. If there is too much, the cost goes up, while if it is too little, consumers will complain. Therefore we have H 0 : μ=16 and H a : μ≠16.

Difference v. Confidence Intervals A hypothesis test makes use of an estimate, such as the sample mean, but is not directly concerned with estimation. The point is to determine if a proposed value of the parameter is contradicted by the data. A hypothesis test resembles the legal concept of “innocent until proven guilty.” The null hypothesis is innocence. If there is not enough evidence to reject that claim, it stands.

Accept vs. Reject In scientific studies, the null hypothesis is based on the current theory, which will continue to be believed unless there is strong evidence to reject it. However, the failure to reject the null hypothesis does not mean it is true, just as the guilty sometimes do go free because of lack of evidence. Thus, statisticians resist saying “accept H 0.” When there is enough evidence, we reject H 0, and replace it with H a. H 0 is never accepted as a result of the test, since it was assumed to begin with. Therefore, we will use the terms “Reject H 0 ” and “Do Not Reject H 0 ” (DNR) to describe the results of the test.

Hypothesis Tests of the Mean The null hypothesis is initially assumed true. It states that the mean has a particular value, μ 0. Therefore, it follows that the distribution of has the same mean, μ 0. We reason as follows: If we take a sample, we get a particular sample mean,. If the null hypothesis is true, is not likely to be “far away” from μ 0. It could happen, but it’s not likely. Therefore, if is “too far away,” we will suspect something is wrong, and reject the null hypothesis. The next slide shows this graphically.

Comments on the Graph What we see in the previous graph is the idea that lots of sample means will fall close to the true mean. About 68% fall within one standard deviation. There is still a 32% chance of getting a sample mean farther away than that. So, if a mean occurs more than one standard deviation away, we may still consider it quite possible that this is a random fluctuation, rather than a sign that something is wrong with the null hypothesis.

More Comments If we go to two standard deviations, about 95% of observed means would be included. There is only a 5% chance of getting a sample mean farther away than that. So, if a far-away mean occurs (more than two standard deviations out), we think it is more likely that it comes from a different distribution, rather than the one specified in the null hypothesis.

Choosing a Significance Level The next graph shows what it means to choose a 5% significance level. If the null hypothesis is true, there is only a 5% chance that the standardized sample mean will be above 1.96 or below These values will serve as a cutoff for the test.

Decision Time We have already shown that we can use a standardized value instead of to decide when to reject. We will call this value Z*, the standard normal test statistic. The criterion by which we decide when to reject the null hypothesis is called a “decision rule.” We establish a cutoff value, beyond which is the rejection region. If Z* falls into that region, we will reject H o. The next slide shows this for α=.05.

Steps in Hypothesis Testing 1.State the null and alternative hypotheses 2.Determine the appropriate type of test 3.State the decision rule (Define the rejection region) 4.Calculate the test statistic 5.State the decision and the conclusion in terms of the original problem

Example A jar of peanut butter is supposed to have 16 oz in it. If there is too much, the cost goes up, while if it is too little, consumers will complain. Assume the amount filled is normally distributed with a standard deviation of ½ oz. In a random sample of 20 jars, the mean amount of peanut butter is oz. Conduct a test to see if the jars are properly filled, using α=.05. Step 1: Hypotheses: H 0 : μ=16 and H a : μ≠16. Step 2: Type of test: The population is normal and standard deviation is given, use Z-test. Step 3: Decision Rule: Reject H 0 if Z*>1.96 or Z*< Step 4: Test Statistic: Step 5: Conclusion: Do not reject H 0 and conclude the jars may be properly filled.

One-tailed Tests Our graphs so far have shown tests with two tails. We have also seen that the alternative hypothesis could be of the form H 2 : μ>μ 0, or H 3 : μ<μ 0. These are one-tailed tests. The rejection region only goes to one side, and all of α goes into one tail (it doesn’t split).

Example Advertising claims have been made that there are 20 chips in every chocolate chip cookie. A sample of 30 cookies gives an average of 18.5 chips per cookie. Assume the standard deviation is 1.5 and conduct an appropriate test using α=.05. Step 1: Hypotheses: H 0 : μ=20 and H a : μ<20. Step 2: Type of Test: Sample is 30 and standard deviation known, use Z-test. Step 3: Decision Rule: Reject H 0 if Z*< Step 4: Test statistic: Step 5: Reject H 0 and conclude the cookies contain less than 20 chips per cookie on average.

Making Mistakes Hypothesis testing is a statistical process involving random events. As a result, we could make the wrong decision. A Type I Error occurs if we reject H 0 when it is true. The probability of this is known as α, the level of significance. A Type II Error occurs when we fail to reject a false null hypothesis. The probability of this is known as β. The Power of a test is 1-β. This is the probability of rejecting the null hypothesis when it is false.

Classification of Errors Actual Decision H o TrueH o False Reject Type I Err P(Error)= α Type B Correct Do Not Reject Type A Correct Type II Err P(Error)=β

Two numbers describe a test The significance level of a test is α, the probability of rejecting H o if it is true. The power of a test is 1-β, the probability of rejecting H o if it is false. There is a kind of trade-off between significance and power. We want significance small and power large, but they tend to increase or decrease together.

p-Value Testing Say you are reporting some research in biology and in your paper you state that you have rejected the null hypothesis at the.10 level. Someone reviewing the paper may say, “What if you used a.05 level? Would you still have rejected?” To avoid this kind of question, researchers began reporting the p-value, which is actually the smallest α that would result in a rejection. It’s kind of like coming at the problem from behind. Instead looking at α to determine a critical region, we let the estimate show us the critical region that would “work.”

How p-Values Work To simplify the explanation, let’s look at a right-tailed means test. We assume a distribution with mean μ 0 and we calculate a sample mean. What if our sample mean fell right on the boundary of the critical region? This is just at the point where we would reject H 0. So if we calculate the probability of a value greater than, this corresponds to the smallest α that results in a rejection. If the test is two tailed, we have to double the probability, because marks one part of the rejection region, but its negative marks the other part, on the other side (other tail).

Using a p-Value Using a p-Value couldn’t be easier. If p<α, we reject H 0. That’s it. p-Values tell us something about the “strength” of a rejection. If p is really small, we can be very confident in the decision. In real world problems, many p-Values turn out to be like.001 or even less. We can feel very good about a rejection in this case. However, if p is around.05 or.1, we might be a little nervous. When Fischer originally proposed these ideas early in the last century, he suggested three categories of decision: –p <.05  Reject H 0 –.05 ≤ p ≤.20  more research needed –p >.20  Accept H 0