R. G. Bias | School of Information | SZB 562BB | Phone: 512 471 7046 | i 1 INF397C Introduction to Research in Information Studies.

Slides:



Advertisements
Similar presentations
1 COMM 301: Empirical Research in Communication Lecture 15 – Hypothesis Testing Kwan M Lee.
Advertisements

Chapter 12: Testing hypotheses about single means (z and t) Example: Suppose you have the hypothesis that UW undergrads have higher than the average IQ.
Statistics and Quantitative Analysis U4320
Statistical Issues in Research Planning and Evaluation
Inference1 Data Analysis Inferential Statistics Research Methods Gail Johnson.
T-tests Computing a t-test  the t statistic  the t distribution Measures of Effect Size  Confidence Intervals  Cohen’s d.
Behavioural Science II Week 1, Semester 2, 2002
Inference about a Mean Part II
Today Concepts underlying inferential statistics
Hypothesis Testing Using The One-Sample t-Test
The t-test Inferences about Population Means when population SD is unknown.
Inferential Statistics
Statistical Analysis. Purpose of Statistical Analysis Determines whether the results found in an experiment are meaningful. Answers the question: –Does.
Chapter 5For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 Suppose we wish to know whether children who grow up in homes without access to.
Chapter Ten Introduction to Hypothesis Testing. Copyright © Houghton Mifflin Company. All rights reserved.Chapter New Statistical Notation The.
AM Recitation 2/10/11.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
Hypothesis Testing:.
Overview of Statistical Hypothesis Testing: The z-Test
Overview Definition Hypothesis
Chapter 8 Introduction to Hypothesis Testing
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
1 GE5 Lecture 6 rules of engagement no computer or no power → no lesson no SPSS → no lesson no homework done → no lesson.
Section #4 October 30 th Old: Review the Midterm & old concepts 1.New: Case II t-Tests (Chapter 11)
Statistical Analysis Statistical Analysis
Significance Tests …and their significance. Significance Tests Remember how a sampling distribution of means is created? Take a sample of size 500 from.
Fall 2013 Lecture 5: Chapter 5 Statistical Analysis of Data …yes the “S” word.
Statistics Primer ORC Staff: Xin Xin (Cindy) Ryan Glaman Brett Kellerstedt 1.
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
1 Statistical Inference. 2 The larger the sample size (n) the more confident you can be that your sample mean is a good representation of the population.
Making decisions about distributions: Introduction to the Null Hypothesis 47:269: Research Methods I Dr. Leonard April 14, 2010.
Exam Exam starts two weeks from today. Amusing Statistics Use what you know about normal distributions to evaluate this finding: The study, published.
Inference We want to know how often students in a medium-size college go to the mall in a given year. We interview an SRS of n = 10. If we interviewed.
Lecture 5: Chapter 5: Part I: pg Statistical Analysis of Data …yes the “S” word.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 1): Two-tail Tests & Confidence Intervals Fall, 2008.
Introduction to Inferential Statistics Statistical analyses are initially divided into: Descriptive Statistics or Inferential Statistics. Descriptive Statistics.
Essential Question:  How do scientists use statistical analyses to draw meaningful conclusions from experimental results?
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Review Hints for Final. Descriptive Statistics: Describing a data set.
1 Chapter 8 Introduction to Hypothesis Testing. 2 Name of the game… Hypothesis testing Statistical method that uses sample data to evaluate a hypothesis.
Physics 270 – Experimental Physics. Let say we are given a functional relationship between several measured variables Q(x, y, …) x ±  x and x ±  y What.
R. G. Bias | School of Information | SZB 562BB | Phone: | i 1 INF 397C Fall, 2003 Days 12 & 13.
Chapter 8 Parameter Estimates and Hypothesis Testing.
KNR 445 Statistics t-tests Slide 1 Introduction to Hypothesis Testing The z-test.
Stats Lunch: Day 3 The Basis of Hypothesis Testing w/ Parametric Statistics.
Chapter 10 The t Test for Two Independent Samples
Chapter Eight: Using Statistics to Answer Questions.
© Copyright McGraw-Hill 2004
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Please hand in homework on Law of Large Numbers Dan Gilbert “Stumbling on Happiness”
Hypothesis test flow chart
Chapter 13 Understanding research results: statistical inference.
R. G. Bias | School of Information | SZB 562BB | Phone: | i 1 INF397C Introduction to Research in Information Studies.
Hypothesis Testing and Statistical Significance
Chapter 7: Hypothesis Testing. Learning Objectives Describe the process of hypothesis testing Correctly state hypotheses Distinguish between one-tailed.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Statistical principles: the normal distribution and methods of testing Or, “Explaining the arrangement of things”
Chapter 10: The t Test For Two Independent Samples.
Chapter 9 Introduction to the t Statistic
The Single-Sample t Test Chapter 9. t distributions >Sometimes, we do not have the population standard deviation, σ. Very common! >So what can we do?
R. G. Bias | School of Information | SZB 562BB | Phone: | i 1 INF397C Introduction to Research in Information Studies.
INF397C Introduction to Research in Information Studies Spring, Day 12
Central Limit Theorem, z-tests, & t-tests
Inferential Statistics
Psych 231: Research Methods in Psychology
Presentation transcript:

R. G. Bias | School of Information | SZB 562BB | Phone: | i 1 INF397C Introduction to Research in Information Studies Fall, 2005 Days 10 & 11

R. G. Bias | School of Information | SZB 562BB | Phone: | i 2 Context Where we’ve been: –Descriptive statistics Frequency distributions Graphs Types of scales Probability Measures of central tendency and spread z scores –Experimental design The scientific method Operational definitions IV, DV, controls, counterbalancing, confounds Validity, reliability Within- and between-subject designs –Qualitative research Gracy, Rice Lively

R. G. Bias | School of Information | SZB 562BB | Phone: | i 3 Context (cont’d.) Where we’re going: –More descriptive statistics Correlation –Inferential statistics Confidence intervals Hypothesis testing, Type I and II errors, significance level t-tests Anova –Which method when? –Cumulative final

R. G. Bias | School of Information | SZB 562BB | Phone: | i 4 First, correcting a lie Parameters (for populations) Statistics (for samples) Meanµ = ΣX/N M = ΣX/N Standard deviation σ = SQRT of (ΣX 2 –(ΣX) 2 /N)/ N s = SQRT of (ΣX 2 –(ΣX) 2 /N)/ N-1

R. G. Bias | School of Information | SZB 562BB | Phone: | i 5 Degrees of Freedom Demo

R. G. Bias | School of Information | SZB 562BB | Phone: | i 6 Standard Error of the Mean So far, we’ve computed a sample mean (M, X bar), and used it to estimate the population mean (µ). One thing we’ve gotten convinced of (I hope) is... larger sample sizes are better. –Think about it – what if I asked ONE of you, what School are you a student in? Versus asking 10 of you?

R. G. Bias | School of Information | SZB 562BB | Phone: | i 7 Standard Error (cont’d.) Well, instead of picking ONE sample, and using that mean to estimate the population mean, what if we sampled a BUNCH of samples? If we sampled ALL possible samples, the mean of the means would equal the population mean. (“µ M ”) Here are some other things we know: –As we get more samples, the mean of the sample means gets closer to the population mean. –Distribution of sample means tends to be normal. –We can use the z table to find the probability of a mean of a certain value. –And most importantly...

R. G. Bias | School of Information | SZB 562BB | Phone: | i 8 Standard Error (cont’d.) We can easily work out the standard deviation of the distribution of sample means: SE = S M = S/SQRT(N) So, the standard error of the mean is the standard distance that a sample mean is from the population mean. Thus, the SE tells us how good an estimate our sample mean is of the population mean. Note, as N gets larger, the SE gets smaller, and the better the sample mean estimates the population mean. Hold on – we’ll use SE later.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 9 Four Questions 1.Does an iSchool IT-provided online tutorial lead to better learning than a face-to-face class?

R. G. Bias | School of Information | SZB 562BB | Phone: | i 10 Two methods of making statistical inferences Null hypothesis testing –Assume IV has no effect on DV; differences we obtain are just by chance (error variance) –If the difference is unlikely enough to happen by chance (and “enough” tends to be p <.05), then we say there’s a true difference. Confidence intervals –We compute a confidence interval for the “true” population mean, from sample data. (95% level, usually.) –If two groups’ confidence intervals don’t overlap, we say (we INFER) there’s a true difference.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 11 Remember... Earlier I said that there are two ways for us to be confident that something is true: –Statistical inference –Replicability Now I’m saying there are two avenues of statistical inference: –Hypothesis testing –Confidence intervals

R. G. Bias | School of Information | SZB 562BB | Phone: | i 12 t-tests Remember the z scores: –z = (X - µ)/σ –It is often the case that we want to know “What percentage of the scores are above (or below) a certain other score”? –Asked another way, “What is the area under the curve, beyond a certain point”? –THIS is why we calculate a z score, and the way we do it is with the z table, on p. 306 of Hinton. Problem: We RARELY truly know µ or σ.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 13 t-tests (cont’d.) So, typically what we do is use M to estimate µ and s to estimate σ. (Duh.) (Note: When we estimate σ with s, we divide by N-1, which is degrees of freedom.) Then, instead of z, we calculate t. Hinton’s example on p. 64 is for a t-test when you have a null hypothesis population mean (µ 0 ). (That is, you want to test if your observed sample mean is different from some value.) Hinton then offers examples in Chapter 8 of related (dependent, within-subjects) and independent (unrelated, between-subjects) t-tests. S, Z, & Z’s example on p. 409 is for a t-test to compare independent means.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 14 Formulae -For a single mean(compared with µ 0 ): -t = (M - µ)/(s/SQRTn) -For related (within-subjects) groups: -t = (M 1 – M 2 )/s M1 – M2 -Where s M1 – M2 = (s x1 – x2 )/SQRTn -See Hinton, p. 83 -For independent groups: -From S, Z, & Z, p. 409, and Hinton, p. 87 -t = (M 1 – M 2 )/s M1 – M2 –Where s M1 – M2 = SQRT [(S 1 2 /n 1 ) + (S 2 2 /n 2 )] –See Hinton, p. 87 »Will’s correction.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 15 Steps For a t test for a single sample –Restate the question as a research hypothesis and a null hypothesis about the populations. –Determine the characteristics of the comparison distribution. The mean is the known population mean. Compute the standard deviation by: –Calculate the estimated population variance (S 2 = SS/df) –Calculate the variance of the distribution of means (S 2 /n) –Take the square root, to get SE. –Note, we’re calculating t with N-1 df. Determine the cutoff sample score on the comparison distribution at which the null hypothesis should be rejected. –Decide on an alpha and one-tailed vs. two-tailed –Look up the critical value in the table –Determine your sample’s t score: t = m- µ / SE –Decide whether to reject or not reject the null hypothesis. (If the observed value of t exceeds the table value, reject.)

R. G. Bias | School of Information | SZB 562BB | Phone: | i 16 Steps For a t test for dependent means –Restate the question as a research hypothesis and a null hypothesis about the populations. –Determine the characteristics of the comparison distribution. Make each person’s score into a difference score. From here on out, use difference scores. Compute the mean of the difference scores. Assume a population mean of 0: µ = 0. Compute the standard deviation of the difference scores: –Calculate the estimated population variance (S 2 = SS/df) –Calculate the variance of the distribution of means (S 2 /n) –Take the square root, to get SE. –Note, we’re calculating t with N-1 df. Determine the cutoff sample score on the comparison distribution at which the null hypothesis should be rejected. –Decide on an alpha, and one-tailed vs. two-tailed –Look up the critical value in the table –Determine your sample’s t score: t = m - µ / SE –Decide whether to reject or not reject the null hypothesis. (If the observed value of t exceeds the table value, reject.)

R. G. Bias | School of Information | SZB 562BB | Phone: | i 17 Steps For a t test for independent means –Same as for dependent means, except the value for SE is that squirrely formula on Hinton, p. 87. –Basically, here’s the point. When you’re comparing DEPENDENT (within-subject, related) means, you can assume both sets of scores come from the same distribution, thus have the same standard deviation. But when you’re comparing independent (between- subject, unrelated) means, you gotta basically average the variability of each of the two distributions.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 18 Three points df –Four people, take your choice of candy. –One df used up calculating the mean. One or two tails –Must be VERY careful, choosing to do a one-tailed test. Comparing the z and t tables –Check out the.05 t table values for infinity df (1.96 for two-tailed test, for one-tailed). –Now find the commensurate values in the z table.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 19 Let’s work some examples

R. G. Bias | School of Information | SZB 562BB | Phone: | i 20 Confidence Intervals We calculate a confidence interval for a population parameter. The mean of a random sample from a population is a point estimate of the population mean. But there’s variability! (SE tells us how much.) What is the range of scores between which we’re 95% confident that the population mean falls? Think about it – the larger the interval we select, the larger the likelihood it will “capture” the true (population) mean. CI = M +/- (t.05 )(SE) See Box 12.2 on “margin of error.” NOTE: In the box they arrive at a 95% confidence that the poll has a margin of error of 5%. It is just coincidence that these two numbers add up to 100%.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 21 CI about a mean -- example CI = M +/- (t.05 )(SE) Establish the level of α (two-tailed) for the CI. (.05) M=15.0 s=5.0 N=25 Use Table A.2 to find the critical value associated with the df. –t.05 (24) = CI = / (5.0/SQRT 25) = / = – “The odds are 95 out of 100 that the population mean falls between and ” (NOTE: This is NOT the same as “95% of the scores fall within this range!!!)

R. G. Bias | School of Information | SZB 562BB | Phone: | i 22 Another CI example Hinton, p. 89. t-test not sig. What if we did this via confidence intervals?

R. G. Bias | School of Information | SZB 562BB | Phone: | i 23 Significance Level Remember, two ways to test statistical significance – hypothesis tests and confidence intervals. With confidence intervals, if two groups yield data and their confidence intervals don’t overlap, then we conclude a significant difference. In hypothesis testing, if the probability of finding our differences by chance is smaller than our chosen alpha, then we say we have a significant difference. We select alpha (α), by tradition. Statistical significance isn’t the same thing as practical significance.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 24 Type I and Type II Errors World  Our decision Null hypothesis is false Null hypothesis is true Reject the null hypothesis Correct decision Type I error (α) Fail to reject the null hypothesis Type II error (β) Correct decision (1-β)

R. G. Bias | School of Information | SZB 562BB | Phone: | i 25 Power of the Test The power of a statistical test refers to its ability to find a difference in distributions when there really is one there. Things that influence the power of a test: –Size of the effect. –Sample size. –Variability. –Alpha level.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 26 Correlation With correlation, we return to DESCRIPTIVE statistics. (This is counterintuitive. To me.) We are describing the strength and direction of the relationship between two variables. And how much one variable predicts the other.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 27 Homework Continue reading. Optional: For next Monday, send me a link to an on- line research paper that describes (or mentions!) a t test, a confidence interval, an anova, or a correlation.

R. G. Bias | School of Information | SZB 562BB | Phone: | i 28 Effect Size How big of an effect does the IV have on the DV? Remember, two things that make it hard to find a difference are: –There’s a small actual difference. –There’s a lot of within-group variability (error variance). –(Demonstrate with moving distributions.)

R. G. Bias | School of Information | SZB 562BB | Phone: | i 29 Effect Size (cont’d.) From S, Z, & Z: “To be able to observe the effect of the IV, given large within-group variability, the difference between the two group means must be large.” Cohen’s d = (µ 1 – µ 2 )/ σ “Because effect sizes are presented in standard deviation units, they can be used to make meaningful comparisons of effect sizes across experiments using different DVs.”

R. G. Bias | School of Information | SZB 562BB | Phone: | i 30 Effect Size (cont’d.) When σ isn’t known, it’s obtained by pooling the within-group variability across groups and dividing by the total number of scores in both groups. σ = SQRT {[(n 1 -1)S (n 2 -1) S 2 2 ]/N} And, by convention: –d of.20 is considered small –d of.50 is considered medium –d of.80 is considered large

R. G. Bias | School of Information | SZB 562BB | Phone: | i 31 Effect Size example Let’s look at the heights of men and women. Just for grins, intuitively, what you say – small, medium, or large difference? µ women = 64.6 in. µ men = 69.8 in. σ = 2.8 in. d = (µ 1 – µ 2 )/ σ = (69.8 – 64.6)/2.8 = 1.86 So, very large difference. Indeed, one that everyone is aware of.