Comparing Two Means Dependent and Independent T-Tests Class 14
Logic of Inferential Stats Detective Althype: “Tony 'Trout Eyes' Nullhype was at the murder scene.” Tony “Trout Eyes” Nullhype: “No fuggin way! I was at duh church rummage sale!” Dataville Witness Reports Witness 1: Saw Tony at scene Witness 2: Saw Tony at scene Witness 3: Not sure Dataville Witness Reports Witness 1: Saw Tony at scene Witness 2: Saw Tony at scene Witness 3: Not sure Witness 4: Not sure Witness 5: Not sure Witness 6: Not sure Witness 7: Not sure Error
Logic of Inferential Stats Degree of Certainty All Observations 2 witnesses ID’d Tony = 0.66 confirmation rate 3 witnesses total 2 witness ID’d Tony = 0.29 confirmation rate 7 witnesses total
Generating Anxiety—Photos vs. Reality: Within Subjects and Between Subjects Designs Problem Statement: Are people as aroused by photos of threatening things as by the physical presence of threatening things? Hypothesis: Physical presence will arouse more anxiety than pictures. Expt’l Hypothesis: Seeing a real tarantula will arouse more anxiety than will spider photos.
Spider Photos
WUNDT!!!!
WITHIN SUBJECTS DESIGN 1.All subjects see both spider pictures and real tarantula 2.Counter-balanced the order of presentation. Why? 3.DV: Anxiety after picture and after real tarantula Data (from spiderRM.sav) Subject Picture (anx. score) Real T (anx. score)
Results: Anxiety Due to Pictures vs. Real Tarantula Do the means LOOK different? Are they SIGNIFICANTLY DIFFERENT? Yes Need t-test
WHY MUST WE LEARN FORMULAS? Don’t computers make stat formulas unnecessary 1. SPSS conducts most computations, error free 2. In the old days—team of 3-4 work all night to complete stat that SPSS does in.05 seconds. Fundamental formulas explain the logic of stats 1. Gives you more conceptual control over your work 2. Gives you more integrity as a researcher 3. Makes you more comfortable in psych forums
) + (+ ( X ( 5 ) X (365 X 3 y ) = TODDLER FORMULA Point: Knowing the formula without understanding concepts leads to impoverished understanding.
Logic of Testing Null Hypothesis Inferential Stats test the null hypothesis ("null hyp.") This means that test is designed to CONFIRM that the null hyp is true. In WITHIN GROUPS t-test (AKA "dependent" t-test) null hyp. is that responses in Cond. A and in Cond. B come from same population of responses. Null hyp.: Cond A and Cond B DON'T differ. In BETWEEN GROUPS t-test (AKA "independent" t-test) null hyp. is that responses from Group A and from Group B DON’T differ. If tests do not confirm the null hyp, then must accept ALT. HYPE. Alt. hyp. within-groups: Cond A differs from Cond B Alt. hyp. between-groupsGroup A differs from Group B
Null Hyp. and Alt. Hyp in Pictures vs. Reality Study Within groups design : Cond. A (all subjs. see photos), then Cond. B (all subs. see actual tarantula) Null hyp? No differences between seeing photos (Cond A) and seeing real T (Cond B) Anxiety ratings Alt. hyp? There is a difference between seeing photos (Cond A) and seeing real T (Cond B)
QUIZ 2 POSTPONED TO NOV. 12 MID-TERM GRADE ADJUSTMENT 2 PTS. ADDED TO ALL SCORES (i.e., 84 originally now = 86)
Key to T-Test is: Central Tendency (i.e. Mean) Relative to Random Distribution (i.e., SD or SE) Diffs Btwn Means Distribution
T-Test as Measure of Difference Between Two Means 1. Two data samples—do means of each sample differ significantly? 2. Do samples represent same underlying population (null hyp: small diffs) or two distinct populations (alt. hyp: big diffs)? 3. Compare diff. between sample means to diff. we’d expect if null hyp is true 4. Use Standard Error (SE) to gauge variability btwn means. a. If SE small & null hyp. true, sample diffs should be smaller b. If SE big & null hyp. true, sample diffs. can be larger 5. If sample means differ much more than SE, then either: a. Diff. reflects improbable but true random difference w/n true pop. b. Diff. indicates that samples reflect two distinct true populations. 6. Larger diffs. Between sample means, relative to SE, support alt. hyp. 7. All these points relate to both Dependent and Independent t-tests
Logic of T-Test observed difference between sample means expected difference between population means (0 if null hyp. is true) t = − SE of difference between sample means Note: Logic the same for Dependent and Independent t-tests. However, the specific formulas differ.
If Difference Between Means Relative to SE (overlap) is Small: Null Hyp. Supported If Difference Between Means Relative to SE (overlap) is Large: Alternative Hyp. Supported
S D : The Standard Error of Differences Between Means Sampling Distribution: The spread of many sample means around a true mean. SE: The average amount that sample means vary around the true mean. SE = Std. Deviation of sample means. Formula for SE : SE = s/√n, when n > 30 If sample N > 30 the sampling distribution should be normal. Mean of sampling distribution = true mean. S D = Average amount Var. 1 mean differs from Var. 2 mean in Sample 1, then in Sample 2, then in Sample 3, ---- then in Sample N Note: S D is differently computed in Between-subs. designs.
S D : The Standard Error of Differences Between Means TARANTULA PICTURE D MEAN MEAN (T mean – P mean) Study Study Study Study Ave. 2.25
S D : The Standard Error of Differences Between Means TARANT. PICT. D D - D (D-D) 2 Sub Sub Sub Sub X Tarant = 5 X Pic = 2.75 D = 2.25 Σ (D-D) 2 =.77 S D 2 = Sum (D -D) 2 / N - 1; =.77 / 3 =.26 SD = √SD 2 = √.26 =.51 SE of D = σD = SD / √N =.51 / √4 =.51 / 2 =.255 t = D / SE of D = 2.25 /.255 = 8.823
Small S D indicates that average difference between pairs of variable means should be large or small, if null hyp true? Small S D will therefore increase or decrease our chance of confirming experimental prediction, if actual difference is real? Small Increase it. Understanding S D and Experiment Power Power of Experiment: Ability of expt. to detect actual differences.
Assumptions of Dependent T-Test 1. Samples are normally distributed 2. Data measured at interval level (not ordinal or categorical)
Conceptual Formula for Dependent Samples T-Test t = D − μ D s D / √N D = Average difference between mean Var. 1 – mean Var. 2. It represents systematic variation, aka experimental effect. μD = Expected difference in true population = 0 It represents random variation, aka the the null effect. s D / √N = Estimated standard error of differences between all potential sample means. It represents the likely random variation between means. = Experimental Effect Random Variation
Dependent (w/n subs) T-Test SPSS Output t = expt. effect / error t = X / SE t = -7 / 2.83 = SE = SD / √n 2.83 = / √12 Note: Mean = mean diff pic anx - real anx. = = - 7
Independent (between-subjects) t-test 1.Subjects see either spider pictures OR real tarantula 2.Counter-balancing less critical (but still important). Why? 3.DV: Anxiety after picture OR after real tarantula Data (from spiderBG.sav) Subject Condition Anxiety
Assumptions of Independent T-Test DEPENDENT T-TEST 1. Samples are normally distributed 2. Data measured at least at interval level (not ordinal or categorical) INDEPENDENT T-TESTS ALSO ASSUME 3. Homogeneity of variance 4. Scores are independent (b/c come from diff. people).
Logic of Independent Samples T-Test (Same as Dependent T-Test) observed difference between sample means expected difference between population means (if null hyp. is true) t = − SE of difference between sample means Note: SE of difference of sample means in independent t test differs from SE in dependent samples t-test
Conceptual Formula for Independent Samples T-Test t = ( X 1 − X 2 ) − (μ 1 − μ 2 ) Est. of SE (X 1 − X 2 ) = Diffs. btwn. samples It represents systematic variation, aka experimental effect. (μ 1 − μ 2 ) = Expected difference in true populations = 0 It represents random variation, aka the the null effect. Estimated standard error of differences between all potential sample means. It represents the likely random variation between means. = Experimental Effect Random Variation
Computational Formulas for Independent Samples T-Tests t = X 1 − X 2 2 N1N1 N2N2 () s 1 s √ When N 1 = N 2 t = X 1 − X 2 s p s p 2 + √ 2 n1n1 n2n2 When N 1 ≠ N 2 spsp 2 = (n 1 -1)s 1 + (n 2 -1)s n 1 + n 2 − 2 Weighted average of each groups SE =
Independent (between subjects) T-Test SPSS Output t = expt. effect / error t = (X 1 − X 2 ) / SE t = -7 / 4.16 =
Dependent (within subjects) T-Test SPSS Output t = expt. effect / error t = X / SE t = -7 / 2.83 = SE = SD / √n 2.83 = / √12 Note: Mean = mean diff pic anx - real anx. = = - 7
Dependent T-Test is Significant; Independent T-Test Not Significant. A Tale of Two Variances Dependent T-Test Independent T -Test SE = 2.83 SE = 4.16