1 Psych 5500/6500 t Test for Two Independent Means Fall, 2008.

Slides:



Advertisements
Similar presentations
Psych 5500/6500 t Test for Two Independent Groups: Power Fall, 2008.
Advertisements

Chapter 12: Testing hypotheses about single means (z and t) Example: Suppose you have the hypothesis that UW undergrads have higher than the average IQ.
Confidence Interval and Hypothesis Testing for:
Comparing Two Population Means The Two-Sample T-Test and T-Interval.
1. Estimation ESTIMATION.
The Two Sample t Review significance testing Review t distribution
Review: What influences confidence intervals?
PSY 307 – Statistics for the Behavioral Sciences
t scores and confidence intervals using the t distribution
Behavioural Science II Week 1, Semester 2, 2002
Lecture 8 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
Comparing Means: Independent-samples t-test Lesson 14 Population APopulation B Sample 1Sample 2 OR.
Hypothesis Testing Comparisons Among Two Samples.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 10: Hypothesis Tests for Two Means: Related & Independent Samples.
Inferences About Means of Two Independent Samples Chapter 11 Homework: 1, 2, 4, 6, 7.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 5): Outliers Fall, 2008.
Chapter 3 Hypothesis Testing. Curriculum Object Specified the problem based the form of hypothesis Student can arrange for hypothesis step Analyze a problem.
Copyright © 2010 Pearson Education, Inc. Chapter 24 Comparing Means.
Lecture 7 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
PY 427 Statistics 1Fall 2006 Kin Ching Kong, Ph.D Lecture 6 Chicago School of Professional Psychology.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 2): p Values One-Tail Tests Assumptions Fall, 2008.
Today Concepts underlying inferential statistics
Hypothesis Testing Using The One-Sample t-Test
Chapter 9: Introduction to the t statistic
1 Psych 5500/6500 Statistics and Parameters Fall, 2008.
AM Recitation 2/10/11.
Hypothesis Testing:.
1 Psych 5500/6500 Confounding Variables Fall, 2008.
Hypothesis Testing II The Two-Sample Case.
Tests of significance & hypothesis testing Dr. Omar Al Jadaan Assistant Professor – Computer Science & Mathematics.
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 23, Slide 1 Chapter 23 Comparing Means.
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
T tests comparing two means t tests comparing two means.
RMTD 404 Lecture 8. 2 Power Recall what you learned about statistical errors in Chapter 4: Type I Error: Finding a difference when there is no true difference.
Psych 5500/6500 ANOVA: Single-Factor Independent Means Fall, 2008.
Chapter 8 Introduction to Hypothesis Testing
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 24 Comparing Means.
Hypothesis Testing Using the Two-Sample t-Test
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
Stats Lunch: Day 4 Intro to the General Linear Model and Its Many, Many Wonders, Including: T-Tests.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 4): Power Fall, 2008.
Hypotheses tests for means
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 1): Two-tail Tests & Confidence Intervals Fall, 2008.
1 Psych 5500/6500 t Test for Dependent Groups (aka ‘Paired Samples’ Design) Fall, 2008.
1 Psych 5500/6500 Introduction to the F Statistic (Segue to ANOVA) Fall, 2008.
DIRECTIONAL HYPOTHESIS The 1-tailed test: –Instead of dividing alpha by 2, you are looking for unlikely outcomes on only 1 side of the distribution –No.
1 Chapter 8 Introduction to Hypothesis Testing. 2 Name of the game… Hypothesis testing Statistical method that uses sample data to evaluate a hypothesis.
Inference for 2 Proportions Mean and Standard Deviation.
STA 2023 Module 11 Inferences for Two Population Means.
AP Statistics Chapter 24 Comparing Means.
Review - Confidence Interval Most variables used in social science research (e.g., age, officer cynicism) are normally distributed, meaning that their.
Stats Lunch: Day 3 The Basis of Hypothesis Testing w/ Parametric Statistics.
Review I A student researcher obtains a random sample of UMD students and finds that 55% report using an illegally obtained stimulant to study in the past.
Chapter 10 The t Test for Two Independent Samples
Independent Samples T-Test. Outline of Today’s Discussion 1.About T-Tests 2.The One-Sample T-Test 3.Independent Samples T-Tests 4.Two Tails or One? 5.Independent.
Comparing Means Chapter 24. Plot the Data The natural display for comparing two groups is boxplots of the data for the two groups, placed side-by-side.
Sampling Distribution (a.k.a. “Distribution of Sample Outcomes”) – Based on the laws of probability – “OUTCOMES” = proportions, means, test statistics.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Copyright © 2009 Pearson Education, Inc t LEARNING GOAL Understand when it is appropriate to use the Student t distribution rather than the normal.
Chapter 10: The t Test For Two Independent Samples.
Chapter 9 Introduction to the t Statistic
Chapter 23 Comparing Means.
Issues in Inferential Statistics
Review: What influences confidence intervals?
Chapter 23 Comparing Means.
Elementary Statistics
Chapter 24 Comparing Means Copyright © 2009 Pearson Education, Inc.
Chapter Outline Inferences About the Difference Between Two Population Means: s 1 and s 2 Known.
Presentation transcript:

1 Psych 5500/6500 t Test for Two Independent Means Fall, 2008

2 Cause and Effect Relationships Does Variable X (cause/influence/affect) Variable Y? Variable X would be the independent variable (the cause). Variable Y would be the dependent variable (the effect).

3 t Test for 2 Independent Means Used to compare two independent samples, specifically we are comparing the means of the samples to see whether or not we can infer that the means of the two populations they represent are different. Usually we are testing a theory which proposes that an independent variable has made the population means different.

4 Example 1 Question: Does gender influence respiration rate? Design: randomly sample 20 males and 20 females from some population and measure their respiration rate while at rest.

5 Example 1 I.V.= D.V.= H0: μ female = μ male (or) μ female -μ male =0 HA: μ female  μ male (or) μ female -μ male  0

6 Hypotheses Note that the hypotheses are always about populations. H0: μ female = μ male HA: μ female  μ male NOT about samples: We want to be able to generalize the results to the populations from which we sampled, not just to the specific participants who are in our sample.

7 Example 2 Question: Is maze running ability affected by the presence or absence of some specific drug? Design: Select 11 rats and randomly divide them into two groups. The rats in Group 1 are given the drug, the rats in Group 2 are not. The number of wrong turns each rat makes before reaching the end of the maze is recorded.

8 Example 2 I.V.= D.V.= H0: μ drug = μ no_drug (or) μ drug -μ no_drug = 0 HA: μ drug  μ no_drug (or) μ drug -μ no_drug  0

9 Statistic The statistic upon which we will base our decision about H0 is: Challenge: if H0 is true then the two population means are equal, and we expect the two sample means to be similar, but even if H0 is true the two sample means will not exactly equal each other due to chance (i.e. the samples have random bias). So if the two sample means differ is that due to differences in the population means or due to chance?

10 The Solution Set up the ‘Sampling Distribution of assuming H0 is true’ and label it as such. Then: The mean of that sampling distribution is based upon H0. The standard deviation of that sampling distribution is estimated from the data from the two samples

11 Sampling Distribution of the statistic if H0 is true

12 Sampling Distribution (cont.) (based upon H0 being true) Computational formula for the standard error of the difference:

13 Standard Error of the Difference Let’s take a conceptual look at the standard deviation of the test statistic This t test assumes that both populations have the same variance, and uses this to ‘pool’ the two estimates of this population variance into one good estimate using the following formula, which weights the estimate based upon the size of each sample was.: Conceptual formula for the standard error of the difference:

14 Drug Example Data Group 1 (Drug) Group 2 (No Drug )

15 Group Statistics

16 t Computations

17

18 Setting up the Rejection Regions

19

20 t obt

21

22 Decision We ‘reject H0’. We can conclude that the two samples represent populations that have different means. We would like to then go on and state that the independent variable (drug vs. no drug) must have made the two population means different from each other, but at this point we can’t, first we have to show that there is no other reason for why the two populations might have different means. This takes us to the topic of ‘confounding variables’, which we will cover in the next lecture.

23 Confidence Intervals Confidence interval for the true difference between μ 1 and μ 2 : 95% confidence interval: 0.67  (μ 1 - μ 2 )  5.33 If a hypothesis states a value for μ 1 - μ 2 that is outside of that confidence interval (e.g. H0: μ 1 - μ 2 = 0) then you can reject it

24 Reporting Results A standard way of reporting the results of t tests is to use the following format: t(df)=t obt, p=? From last example: t(9)=2.91, p=... We can use the ‘t distribution tool’ in ‘Oakley Stat Tools’ to get the exact value of p, or we can use SPSS to do the t test for us. t(9) = 2.91, p=.017

25 One-Tailed Tests So far the examples have involved non-directional theories (which predict the two population means will be different but don’t predict one which will be greater than the other). This is done with a 2-tail test. It is also possible to test theories which are directional (predict specifically which population mean should be greater). This is done with a 1-tail test, which will influence how we write our hypotheses, and will lead to just one rejection region in our sampling distribution.

26 One-tailed Tests Again, express the prediction made by the theory when you write HA, H0 is then everything else. For testing a theory which predicts that the mean of population one (drugged rats) should be greater than the mean of population two (undrugged rats): H0: μ 1  μ 2 (or equivalently) μ 1 -μ 2  0 HA: μ 1 > μ 2 (or equivalently) μ 1 -μ 2 > 0

27 Rejection Region

28 one-tailed tests For testing a theory which predicts that the mean of population one (drugged rats) should be less than the mean of population two (undrugged rats): H0: μ 1  μ 2 (or equivalently) μ 1 -μ 2  0 HA: μ 1 < μ 2 (or equivalently) μ 1 -μ 2 < 0

29 Rejection Region

30 Null Hypothesis I have given two ways of expressing H0. H0: μ drug = μ no_drug (or) μ drug - μ no_drug = 0 The first way is somewhat conceptually easier, but the second has an advantage as well. I have mentioned before that the null hypothesis is usually, but not always, the hypothesis of ‘no difference’. Let me now give an example where that wouldn’t be true…

31 Let us say that the difference in population means between rats on the drug and those not on the drug has been previously established as equaling 1.5. In this experiment, however, we will have the rats swim through the maze and we are testing a theory which predicts that the difference between the two groups should be influenced by swimming (i.e. it will no longer equal 1.5) The null hypothesis is that performance won’t be influenced by swimming. Now our hypotheses would look like this: H0: μ drug -μ no_drug = 1.5 HA: μ drug -μ no_drug  1.5 Note that conceptually H0 can still be thought of as ‘no difference’, in this case we are saying that the difference between the drug and no drug groups will not be different when the rats are swimming.

32 Sampling Distribution if H0 is true

33 t obtained The t obtained score is still simply a standard score on the curve: i.e. the test statistic minus the mean of the curve divided by the standard deviation (standard error) of the curve.

34 Assumptions underlying this t test 1.Independence of scores (within and between groups) 2.Both populations normally distributed 3.The two populations have identical variances We will examine the assumption of identical variances in a latter lecture.

35 Effect Size Again the measures of effect size include: 1.Simply reporting the ‘raw’ effect size. 2.Reporting a standardized effect size. 3.Reporting the strength of association (which we will cover next semester).

36 ‘Raw’ Effect Size This would be simply reporting the mean of the two groups and the difference between those means. In our experiment the mean number of wrong turns made by the ‘drug’ group was 7, the mean number for the ‘no drug’ group was 4, which is a difference of 3 between the two means.

37 Standardized Effect Size We will take a look at three standardized measures of effect size: 1.Cohen’s d 2.Hedges’s g 3.Glass’s Δ

38 Cohen’s d For the effect size in the population (exact, not estimated): Remember that an assumption underlying this t test is that both populations have the same variance (σ Y ).

39 Cohen’s d For the effect size in the sample: If you examine the formula for S pooled you can see that it is simply the standard deviation of all the scores lumped into one group.

40 Hedges’s g Hedges’s g gives us an estimate of the effect size in the population form which we sampled. est. σ pooled is a ‘pooled’ estimate of the standard deviation of Y, the formula combines the estimates of σ Y from the two groups, weighting each estimate based upon how many scores were in the group.

41 Glass’s Δ Glass’s delta is similar to Hedges’s g, but instead of using the scores from both groups to estimate the standard deviation of Y, delta just uses the data from the control group to estimate it.

42 Glass’s Δ In an experiment designed to test the effect of some treatment it is common to include a group that is handled exactly like the group that gets the treatment except without the treatment, this is called the ‘Control Group’ and the group that gets the treatment is called the ‘Treatment Group’. The reason the control group is named as it is will be covered in the lecture on confounding variables. In our example the rats given the drug constitute the ‘treatment group’ and those not given the drug are the ‘control group’.

43 Glass’s Δ Glass’s delta is used when there is reason to believe that the treatment applied to the treatment group might have affected not only the mean of the group but the variance of that group as well. If this is the case then the assumption of equal variances is not met (we will see how to analyze the data anyway later) and it makes no sense to use the variance of both groups to estimate the variance of Y, as they would actually be measuring two different variances, not one.

44 Using these Formulas

45 Using these Formulas

46 Using these Formulas

47 Using these Formulas

48 Useful Conversions ‘prop’ means ‘proportion’, so if 14 of the 20 scores were in group 1, then p1 would be 14/20 while p2 would be 6/20.

49 Useful Conversions

50 Useful Conversions