Comparing Means: t-tests Wednesday 22 February 2012/ Thursday 23 February 2012.

Slides:



Advertisements
Similar presentations
Inferential Statistics
Advertisements

Chapter 12: Testing hypotheses about single means (z and t) Example: Suppose you have the hypothesis that UW undergrads have higher than the average IQ.
Review of the Basic Logic of NHST Significance tests are used to accept or reject the null hypothesis. This is done by studying the sampling distribution.
University of Warwick, Department of Sociology, 2012/13 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard) Week 4 Regression Analysis: Some Preliminaries...
Hypothesis Testing IV Chi Square.
PSY 307 – Statistics for the Behavioral Sciences
The t-test:. Answers the question: is the difference between the two conditions in my experiment "real" or due to chance? Two versions: (a) “Dependent-means.
Overview of Lecture Parametric Analysis is used for
Statistics 101 Class 9. Overview Last class Last class Our FAVORATE 3 distributions Our FAVORATE 3 distributions The one sample Z-test The one sample.
T-Tests Lecture: Nov. 6, 2002.
DTC Quantitative Research Methods Statistical Inference II: Statistical Testing Thursday 7th November 2014  
Hypothesis Testing Using The One-Sample t-Test
Getting Started with Hypothesis Testing The Single Sample.
Chapter 9: Introduction to the t statistic
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
Inferential Statistics
Statistical Analysis. Purpose of Statistical Analysis Determines whether the results found in an experiment are meaningful. Answers the question: –Does.
Inferential Statistics
AM Recitation 2/10/11.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
Jeopardy Hypothesis Testing T-test Basics T for Indep. Samples Z-scores Probability $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
Hypothesis Testing II The Two-Sample Case.
Copyright © 2012 by Nelson Education Limited. Chapter 8 Hypothesis Testing II: The Two-Sample Case 8-1.
Hypothesis testing – mean differences between populations
Section #4 October 30 th Old: Review the Midterm & old concepts 1.New: Case II t-Tests (Chapter 11)
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Significance Tests …and their significance. Significance Tests Remember how a sampling distribution of means is created? Take a sample of size 500 from.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Education 793 Class Notes T-tests 29 October 2003.
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 2 – Slide 1 of 25 Chapter 11 Section 2 Inference about Two Means: Independent.
1 Level of Significance α is a predetermined value by convention usually 0.05 α = 0.05 corresponds to the 95% confidence level We are accepting the risk.
Chapter 9 Hypothesis Testing and Estimation for Two Population Parameters.
Chapter 9: Testing Hypotheses
Hypothesis Testing CSCE 587.
Hypothesis Testing Using the Two-Sample t-Test
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 10. Hypothesis Testing II: Single-Sample Hypothesis Tests: Establishing the Representativeness.
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
Copyright © Cengage Learning. All rights reserved. 14 Elements of Nonparametric Statistics.
Chapter 12 A Primer for Inferential Statistics What Does Statistically Significant Mean? It’s the probability that an observed difference or association.
Two sample problems:  compare the responses in two groups  each group is a sample from a distinct population  responses in each group are independent.
Confidence intervals and hypothesis testing Petter Mostad
Statistical Inference Friday 17 th February 2012.
Essential Question:  How do scientists use statistical analyses to draw meaningful conclusions from experimental results?
FPP 28 Chi-square test. More types of inference for nominal variables Nominal data is categorical with more than two categories Compare observed frequencies.
Statistical Inference Friday 15 th February 2013.
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Chapter 9 Three Tests of Significance Winston Jackson and Norine Verberg Methods: Doing Social Research, 4e.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
I271B The t distribution and the independent sample t-test.
Marketing Research Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides 1.
Chapter 8 Parameter Estimates and Hypothesis Testing.
KNR 445 Statistics t-tests Slide 1 Introduction to Hypothesis Testing The z-test.
Testing Differences between Means, continued Statistics for Political Science Levin and Fox Chapter Seven.
Chapter 9: Testing Hypotheses Overview Research and null hypotheses One and two-tailed tests Type I and II Errors Testing the difference between two means.
Inferential Statistics Friday 13 th February 2009.
- We have samples for each of two conditions. We provide an answer for “Are the two sample means significantly different from each other, or could both.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
IM911 DTC Quantitative Research Methods Statistical Inference I: Sampling distributions Thursday 4 th February 2016.
366_7. T-distribution T-test vs. Z-test Z assumes we know, or can calculate the standard error of the distribution of something in a population We never.
Learning Objectives After this section, you should be able to: The Practice of Statistics, 5 th Edition1 DESCRIBE the shape, center, and spread of the.
Inference for distributions: - Comparing two means.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
Lecture 7: Bivariate Statistics. 2 Properties of Standard Deviation Variance is just the square of the S.D. If a constant is added to all scores, it has.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Statistical Inferences for Variance Objectives: Learn to compare variance of a sample with variance of a population Learn to compare variance of a sample.
Chapter 10: The t Test For Two Independent Samples.
Chapter 9 Introduction to the t Statistic
Hypothesis testing Imagine that we know that the mean income of university graduates is £16,500. We then do a survey of 64 sociology graduates and find.
Presentation transcript:

Comparing Means: t-tests Wednesday 22 February 2012/ Thursday 23 February 2012

Judging whether differences occur by chance… How do we judge whether it is plausible that two population means are the same and that any difference between sample means simply reflect sampling error? Example: Household size of minority ethnic groups (HOH = Head of household; data adapted by Richard Lampard from 1991 Census) 1.The size of the difference between the two sample means Mean Indian HOH: 3.0 Bangladeshi HOH:5.0 Mean Indian HOH: 3.0 Pakistani HOH:4.0 The first difference is more ‘convincing’

2. The sample sizes of the two samples Mean Pakistani HOH: Bangladeshi HOH: Mean Pakistani HOH: Bangladeshi HOH: The second difference is more ‘convincing’ Judging whether differences occur by chance…

3. The amount of variation in each of the two groups (samples) Mean Pakistani HOH: Bangladeshi HOH: Mean Pakistani HOH: Bangladeshi HOH: The second difference is more ‘convincing’. Judging whether differences occur by chance…

Example: the impact of variability on the difference of means.

The logic of a statistical test… The two statistical tests that we have so far looked at: Testing the plausibility of a suggested population mean (via a t-test [or z-test]): Is the sample mean sufficiently different from the suggested population mean that it is implausible that the suggested population mean is correct? Chi-square test: Are the observed frequencies in a table sufficiently different from what one would have expected to have seen if there was no relationship in the population for the idea that there is no relationship in the population to be implausible? Both have asked whether the difference between the actual (observed) data and what one would have expected to have seen given a particular hypothesis is sufficiently large that the hypothesis is implausible.

…the same logic applies to comparing sample means. If the two samples came from populations with identical population means, then one would expect the difference between the sample means to be (close to) zero. Thus, The larger the difference between the two sample means, the more implausible is the idea that the two population means are identical. This is also affected by: Sample size The extent to which there is variation within each group (or sample). The logic of a statistical test…

t-tests Test the null hypothesis: H 0 :  1 =  2 or H 0 :  1 -  2 = 0 The alternative hypothesis is: H 1 :  1   2 or H 1 :  1 -  2  0

What does a t-test measure? Note: T = treatment group and C = control group (from experimental research). In most discussions these will just be shown as groups 1 and 2, indicating different groups.

Example We want to compare the average amount of television watched by Australian and by British children. We have a sample of Australian and a sample of British children. We could say that what we have is something like this: Population of Australian children Population of British children Sample of Australian children Sample of British children inference Want to compare

t distribution critical values For a more detailed view of statistics go all the way to Australia: SurfStatSurfStat Example contd. Here the dependent variable is hours of TV And the independent variable is nationality. When we are comparing means SPSS calls the independent variable the grouping variable and the dependent variable the test variable.

Example contd. If the null hypothesis, of no difference between the two groups, is correct (and children watch the same amount of television in Australia and Britain) we would assume that if we took repeated samples from the two groups the difference in means between them would generally be small or zero. However it is likely that the difference between any two particular samples will be greater than zero. Therefore we build up a sampling distribution of the difference between the two sample means. We use this distribution to determine the probability of getting an observed difference between two sample means from populations with no difference.

If we take a large number of random samples and calculate the difference between each pair of sample means, we will end up with a sampling distribution that has the following properties: It will be a t-distribution (Under the null hypothesis) the mean of the difference between sample means will be zero Mean M 1 - M 2 = 0 The spread of scores around this mean of zero (the standard error) will be defined by the formula: This is called the pooled variance estimate

Back to example… Descriptive statistic Australian sample British sample Mean166 minutes187 minutes Standard deviation29 minutes30 minutes Sample size20 When we are choosing the test of significance it is important that: 1.We are making an inference from TWO samples (of Australian and of British children). Therefore we need a two-sample test 2.The two samples are being compared in terms of an interval- ratio variable – hours of TV watched. Therefore the relevant descriptive statistic is the mean.  These facts lead us to select the two sample t-test for the equality of means as the relevant test of significance. Table 1. Descriptive statistics for the samples

Descriptive statistic Australian sample British sample Mean166 minutes187 minutes Standard deviation 29 minutes30 minutes Sample size20 SDM = (20-1) (20-1) = – 2 20 x 20     t sample = 166 – 187 = – Calculating the t-score

Obtaining a p-value for a t-score To obtain the p-value for this t-score we need to consult the table for critical values for the t-distribution (see Appendix A.2. in Field) The number of degrees of freedom we refer to in the table in the combined sample size minus two: df = N 1 + N 2 – 2 Here that is – 2 = 38 The table doesn’t have a row of probabilities for 38. In that case we refer to the row for the nearest reported number of degrees of freedom below the desired number. Here that is 35. With 38 degrees of freedom on a two-tail test, t sample falls between the two stated t-scores of 2.03 and The p-value, which falls between the significance levels for these scores is therefore between 0.01 and 0.05 Therefore the p-value is statistically significant at a 0.05 level but not at a 0.01 level.

Reporting the results We can say that: The mean number of minutes of TV watched by the sample of 20 British children is 187 minutes, which is 21 minutes higher than the sample of 20 Australian children, and this difference is statistically significant at the 0.05 level (t(38)= - 2.3, p = 0.03, two-tail). Based on these results we can reject the hypothesis that British and Australian children watch the same average amount of television every night.

Calculating the effect size To discover whether the effect is substantive we want to know the size of it. You can convert t-values into an r-value (a PRE statistic) with the following equation: r = t 2 = = 0.34 t 2 + df This is a medium sized effect. 