1 Difference Between the Means of Two Populations.

Slides:



Advertisements
Similar presentations
Introductory Mathematics & Statistics for Business
Advertisements

Lecture 2 ANALYSIS OF VARIANCE: AN INTRODUCTION
HYPOTHESIS TESTING. Purpose The purpose of hypothesis testing is to help the researcher or administrator in reaching a decision concerning a population.
INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE
1 T-test for the Mean of a Population: Unknown population standard deviation Here we will focus on two methods of hypothesis testing: the critical value.
1 Confidence Interval for Population Mean The case when the population standard deviation is unknown (the more common case).
Chapter 4 Inference About Process Quality
Comparing Two Groups’ Means or Proportions: Independent Samples t-tests.
“Students” t-test.
Statistics Review – Part II Topics: – Hypothesis Testing – Paired Tests – Tests of variability 1.
1 1 Slide IS 310 – Business Statistics IS 310 Business Statistics CSU Long Beach.
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 Matched Samples The paired t test. 2 Sometimes in a statistical setting we will have information about the same person at different points in time.
1 Test a hypothesis about a mean Formulate hypothesis about mean, e.g., mean starting income for graduates from WSU is $25,000. Get random sample, say.
1 Analysis of Variance This technique is designed to test the null hypothesis that three or more group means are equal.
1 Difference Between the Means of Two Populations.
PSY 307 – Statistics for the Behavioral Sciences
1 More Regression Information. 2 3 On the previous slide I have an Excel regression output. The example is the pizza sales we saw before. The first thing.
DATA ANALYSIS I MKT525. Plan of analysis What decision must be made? What are research objectives? What do you have to know to reach those objectives?
An Inference Procedure
1/55 EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2008 Chapter 10 Hypothesis Testing.
1 The Basics of Regression Regression is a statistical technique that can ultimately be used for forecasting.
1 Hypothesis Testing In this section I want to review a few things and then introduce hypothesis testing.
More Simple Linear Regression 1. Variation 2 Remember to calculate the standard deviation of a variable we take each value and subtract off the mean and.
Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 7 th Edition Chapter 9 Hypothesis Testing: Single.
Chapter 11: Inference for Distributions
1 T-test for the Mean of a Population: Unknown population standard deviation Here we will focus on two methods of hypothesis testing: the critical value.
Chapter 9 Hypothesis Testing.
1 (Student’s) T Distribution. 2 Z vs. T Many applications involve making conclusions about an unknown mean . Because a second unknown, , is present,
1 Confidence Interval for Population Mean The case when the population standard deviation is unknown (the more common case).
An Inference Procedure
Fundamentals of Hypothesis Testing: One-Sample Tests
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Statistical Inferences Based on Two Samples Chapter 9.
Chapter 9 Hypothesis Testing and Estimation for Two Population Parameters.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 11 Inferences About Population Variances n Inference about a Population Variance n.
6.1 - One Sample One Sample  Mean μ, Variance σ 2, Proportion π Two Samples Two Samples  Means, Variances, Proportions μ 1 vs. μ 2.
QMS 6351 Statistics and Research Methods Regression Analysis: Testing for Significance Chapter 14 ( ) Chapter 15 (15.5) Prof. Vera Adamchik.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Section Inference about Two Means: Independent Samples 11.3.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
10.1: Confidence Intervals Falls under the topic of “Inference.” Inference means we are attempting to answer the question, “How good is our answer?” Mathematically:
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
© Copyright McGraw-Hill 2000
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 26 Chapter 11 Section 1 Inference about Two Means: Dependent Samples.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
10.1 – Estimating with Confidence. Recall: The Law of Large Numbers says the sample mean from a large SRS will be close to the unknown population mean.
Chapter 7 Inference Concerning Populations (Numeric Responses)
 List the characteristics of the F distribution.  Conduct a test of hypothesis to determine whether the variances of two populations are equal.  Discuss.
1 1 Slide IS 310 – Business Statistics IS 310 Business Statistics CSU Long Beach.
Chapters 22, 24, 25 Inference for Two-Samples. Confidence Intervals for 2 Proportions.
Chapter 9 Hypothesis Testing.
Inference for Two-Samples
Inference about Two Means - Independent Samples
Two-Sample Hypothesis Testing
Chapter 4. Inference about Process Quality
CHAPTER 6 Statistical Inference & Hypothesis Testing
John Loucks St. Edward’s University . SLIDES . BY.
Chapter 8 Hypothesis Testing with Two Samples.
Hypothesis Tests for a Population Mean in Practice
Chapter 9 Hypothesis Testing.
Chapter 11 Inferences About Population Variances
Chapter 9 Hypothesis Testing.
STAT Z-Tests and Confidence Intervals for a
Lecture 10/24/ Tests of Significance
Intro to Confidence Intervals Introduction to Inference
Chapter 9 Hypothesis Testing: Single Population
Statistical Inference for the Mean: t-test
Presentation transcript:

1 Difference Between the Means of Two Populations

2 We have be studying inference methods for a single variable. When the variable was quantitative we had inference for the population mean. When the variable was qualitative we had inference for the population proportion. Now we want to study inference methods for two variables. Both variables could be quantitative, both qualitative or one of each. Depending on the which we have, we will look to certain techniques. At this stage of the game we will begin to look at these different methods. I want to start with 1 quantitative variable and one qualitative variable. In fact the qualitative variable is special here: the variable identifies membership in one of only two groups. Then on the quantitative variable we segment each observation into the appropriate group and think about the mean of the quantitative variable of each group.

3 Our context here is that we really want to know about the population of the two groups, but we will only take a sample from each group. We will look at both confidence intervals and hypothesis tests in this context. Some notation: μ i = the population mean of group i for i = 1, 2. σ i = the population standard deviation for group i for i = 1, 2. x i = the sample mean of group i for i = 1, 2. s i = the sample standard deviation for group i for i = 1, 2. Now for ease of typing I will call the population means mu1 or mu2, population standard deviations sigma1 or sigma2, sample means xbar1 or xbar2, and sample standard deviations s1 or s2. n1 is the sample size from population 1 and n2 has similar meaning.

4 Our context for inference is really the difference in means: mu1 minus mu2. So we are checking to see what difference there is in the means from the two groups. Our point estimator will be xbar1 minus xbar2. In a repeated sampling context the point estimator would vary from sample to sample. As an example say I want to check the average age of students in the economics program and the finance program. One sample from each group would yield one estimate and the estimate would likely be different when I get a different sample (from each major). Also note the sample obtained from group 1 is independent of the sample obtained from group 2. The sampling distribution of xbar1 minus xbar2 will be studied next.

5 The sampling distribution of xbar1 minus xbar2 Case 1 – we can use the normal distribution for the sampling distribution when sigma1 and sigma2 are known. This means we will use Z in our confidence intervals and hypothesis tests The center of the sampling distribution is mu1 minus mu2 and the standard error is the (note or digress x^2 means x squared ) square root[((sigma1^2)/n1)+((sigma2^2)/n2)]. Case 2 – we can use the t distribution for the sampling distribution when sigma1 and sigma2 are unknown. This means we will use t in our confidence intervals and hypothesis tests. The center of the sampling distribution is mu1 minus mu2 and the standard error is seen as the denominator of the equation on page 389. This is not pretty, but we must use it. Note that when using a t distribution one needs to have a degrees of freedom value. In our current context the value is n1 + n2 – 2.

6 Inference for case 1 Confidence interval We are C% confident the unknown population difference mu1 minus mu2 is in the interval (xbar1 minus xbar2) ± MOE, Where MOE = margin of error and this equals the appropriate Z times the standard error of the sampling distribution. Remember if C = 95 the Z = 1.96, and if C = 90 the Z = 1.645, and if C = 99 the Z = 2.58.

7 Hypothesis Test Recall from our past work that in an hypothesis test context we have a null and an alternative hypothesis. Plus the form of the alternative hypothesis will determine if we have a one or a two tailed test. Two tailed test When we study the difference in the means from two populations if we feel there is a difference of Do, but not concerned about the difference being positive or negative, then the null and alternative hypotheses are Ho: mu1 minus mu2 = Do, Ha: mu1 minus mu2 ≠ Do, and we have a two tailed test. Based on an alpha value (the probability of a type I error), we pick critical values of Z and if the calculated Z is more extreme than either critical value we reject the null and go with the alternative.

8 The calculated value of Z from the sample information = [(xbar1 minus xbar2) minus Do] divided by the standard error listed on slide 5 with case 1. Another way to think of the hypothesis test is with the use of the p-value for the calculated Z. If the p-value < alpha, reject the null. Otherwise you have to stick with the null. In practice with a two tailed test you will find the p-value as the area on one side of the distribution but you must double it to be on both sides. alpha/2 Upper Critical Z Alpha/2 lower Critical Z

9 One tailed test When the researcher has the feeling that the difference in mu1 and mu2 should be positive, then the alternative will reflect this feeling and we will have Ho: mu1 minus mu2 ≤ Do Ha: mu1 minus mu2 > Do. The signs are reversed when the researcher feels the difference should be negative. The test procedure proceeds in the same fashion as with the two- tailed test except the focus is just on one side of the distribution as directed by the alternative hypothesis. Note that Do is often zero. In that case we just want to see if the group means are different.

10 Common critical Z’s Two tailed testOne tailed test (negative if on left) Alpha = Alpha = Alpha =

11 Inference for case 2 Inference for case 2 is similar to case 1, except in how the standard error is calculated and that is shown on slide 5 and in using t. Let’s do problem pages 396/397 Here is an Excel printout To assist us. Note means Are incorrect in the book – Check it out! Fixed-Rate (%) Variable-Rate (%) t-Test: Two-Sample Assuming Equal Variances Fixed-Rate (%)Variable-Rate (%) Mean Variance Observations85 Pooled Variance Hypothesized Mean Difference0 df11 t Stat P(T<=t) one-tail t Critical one-tail P(T<=t) two-tail t Critical two-tail

12 a) Let’s say mu1= mean score for fixed-rate and mu2= mean score for variable-rate. Ho: mu1 = mu2 H1: mu1 ≠ mu2. (I have a not equal sign because it says “differ.”) b) With a two tail test the critical t’s (use t because population standard devs are Unknown – with df = = 11 ) with alpha =.10 we have and with alpha =.05 we have and with alpha =.01 we have and with alpha =.001 we have and The t stat from Excel = 3.743, so we reject the null and go with the alternative hypothesis at all the levels listed except when alpha =.001. This is very strong evidence! c) The two-tail p-value is This means p-value < alpha in all cases except when alpha =.001 and conclusion is the same as in part b).

13 d) The 95% confidence interval has us use the t value The interval is ( – 6.966) – and (square root(0.2744(1/8 + 1/5))) = – and gives the interval [ , ]. We can be 95% confident the difference exceeds.4 because the interval is everywhere above.4. e) Ho: mu1 - mu2 < or =.4 H1: mu1 - mu2 >.4. The t stat is ( )/ (square root(0.2744(1/8 + 1/5)) =.71775/ = With a one-tail test the critical t is Since the the t stat is > the critical t we reject the null and go with the alternative.

14 Note that in this problem parts b and c are similar ways of concluding the same thing. Parts d and e also are similar ways of concluding the same thing.