Chapter 5: Introduction to Statistical Inference

Slides:



Advertisements
Similar presentations
Introduction to Hypothesis Testing
Advertisements

Inference Sampling distributions Hypothesis testing.
Inferential Statistics & Hypothesis Testing
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
Review: What influences confidence intervals?
HYPOTHESIS TESTING Four Steps Statistical Significance Outcomes Sampling Distributions.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
PY 427 Statistics 1Fall 2006 Kin Ching Kong, Ph.D Lecture 5 Chicago School of Professional Psychology.
Chapter 9 Hypothesis Testing.
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
Probability Population:
Chapter 5For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 Suppose we wish to know whether children who grow up in homes without access to.
Chapter Ten Introduction to Hypothesis Testing. Copyright © Houghton Mifflin Company. All rights reserved.Chapter New Statistical Notation The.
AM Recitation 2/10/11.
Hypothesis Testing:.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Overview of Statistical Hypothesis Testing: The z-Test
Overview Definition Hypothesis
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 9. Hypothesis Testing I: The Six Steps of Statistical Inference.
Chapter 8 Introduction to Hypothesis Testing
Chapter 11Prepared by Samantha Gaies, M.A.1 –Power is based on the Alternative Hypothesis Distribution (AHD) –Usually, the Null Hypothesis Distribution.
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Chapter 9 Hypothesis Testing II: two samples Test of significance for sample means (large samples) The difference between “statistical significance” and.
Chapter 8 Introduction to Hypothesis Testing
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 1): Two-tail Tests & Confidence Intervals Fall, 2008.
1.State your research hypothesis in the form of a relation between two variables. 2. Find a statistic to summarize your sample data and convert the above.
Copyright © 2010, 2007, 2004 Pearson Education, Inc Section 8-2 Basics of Hypothesis Testing.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
1 Chapter 8 Introduction to Hypothesis Testing. 2 Name of the game… Hypothesis testing Statistical method that uses sample data to evaluate a hypothesis.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Overview.
PSY 307 – Statistics for the Behavioral Sciences Chapter 9 – Sampling Distribution of the Mean.
Copyright ©2013 Pearson Education, Inc. publishing as Prentice Hall 9-1 σ σ.
Stats Lunch: Day 3 The Basis of Hypothesis Testing w/ Parametric Statistics.
Review I A student researcher obtains a random sample of UMD students and finds that 55% report using an illegally obtained stimulant to study in the past.
1 URBDP 591 A Lecture 12: Statistical Inference Objectives Sampling Distribution Principles of Hypothesis Testing Statistical Significance.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Education 793 Class Notes Inference and Hypothesis Testing Using the Normal Distribution 8 October 2003.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Chapter Nine Hypothesis Testing.
Tutorial 11: Hypothesis Testing
Logic of Hypothesis Testing
9.3 Hypothesis Tests for Population Proportions
Chapter 9 Hypothesis Testing.
Hypothesis Testing: One Sample Cases
INF397C Introduction to Research in Information Studies Spring, Day 12
Hypothesis Testing I The One-sample Case
Assumptions For testing a claim about the mean of a single population
Is this quarter fair?. Is this quarter fair? Is this quarter fair? How could you determine this? You assume that flipping the coin a large number of.
Inference and Tests of Hypotheses
Hypothesis Testing and Confidence Intervals (Part 2): Cohen’s d, Logic of Testing, and Confidence Intervals Lecture 9 Justin Kern October 17 and 19, 2017.
CHAPTER 6 Statistical Inference & Hypothesis Testing
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Hypothesis Testing: Hypotheses
INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Test Review: Ch. 7-9
Chapter 9 Hypothesis Testing.
Testing Hypotheses about Proportions
Review: What influences confidence intervals?
Chapter Nine Part 1 (Sections 9.1 & 9.2) Hypothesis Testing
Hypothesis Testing.
Hypothesis Testing.
What are their purposes? What kinds?
Intro to Confidence Intervals Introduction to Inference
Hypothesis Testing and Confidence Intervals (Part 2): Cohen’s d, Logic of Testing, and Confidence Intervals Lecture 9 Justin Kern April 9, 2018.
Chapter Nine: Using Statistics to Answer Questions
1 Chapter 8: Introduction to Hypothesis Testing. 2 Hypothesis Testing The general goal of a hypothesis test is to rule out chance (sampling error) as.
Type I and Type II Errors
Presentation transcript:

Chapter 5: Introduction to Statistical Inference Suppose we wish to know whether children who grow up in homes without access to television have higher IQs than children in the general population. Assume that IQ is normally distributed in the general population, with μ = 100 and σ = 15 points. We draw a random sample of (N =) 25 children from homes without television (no more than one child per household) and measure each child’s IQ. The mean IQ for our sample turns out to be 103.5. Can we conclude that children without TV are indeed smarter? It is possible that our sample mean exceeds 100 due entirely to chance factors involved in drawing our random sample. We will return to this problem after we have developed some tools for statistical inference. Chapter 5 Prepared by Samantha Gaies, M.A.

Sampling Distributions Population distributions are composed of individual scores; however, psychologists commonly perform their studies on groups. Therefore, we need to understand distribu- tions that are composed of statistics from groups (all of which are the same size). Such distributions are called sampling distributions. If we are looking specifically at the mean of each sample, the distribution is called the sampling distribution of the mean (SDM). What will the SDM look like? The Central Limit Theorem tells us that as the size of the samples increases, the SDM becomes closer in shape to the normal distribution (e.g., less skewed), regardless of the shape of the original population distribution. Chapter 5 Prepared by Samantha Gaies, M.A.

The Sampling Distribution of the Mean The mean of the SDM is the same as the mean of the population from which the samples are being randomly drawn. The standard deviation of the SDM is called the Standard Error of the Mean (SEM). It is found from the standard deviation of the population and the sample size, according to this formula: The SEM: Is larger when the standard deviation of the population distribution is larger. Is smaller than the standard deviation of the population distribution. The larger the samples, the smaller the SEM (i.e., as N increases, the SEM decreases. Chapter 5 Prepared by Samantha Gaies, M.A.

Prepared by Samantha Gaies, M.A. The z Score for Sample Means In summary, the sampling distribution of the mean can, in most cases, be assumed to be a normal distribution, with a mean equal to µ (the mean of the population being sampled), and a standard deviation equal to σ divided by the square root of N. To determine whether the mean of a particular sample is unusual, we can use the methods of the previous chapter and calculate a z score for a sample mean. The formula needed for this is just like the z score for individuals, except that the raw score is a sample mean, and the standard deviation is the SEM: Chapter 5 Prepared by Samantha Gaies, M.A.

The z Score and p value for the TV/IQ Example Let us now find the z score for the example in the first slide to determine whether drawing a random sample of 25 children would frequently yield a sample mean as far from 100 as the mean in our example. First, note that the SEM for our example is: Therefore, the z score for our sample mean is: The (one-tailed) p value corresponding to this z score is the area beyond z = 1.3, which is (from Table A) 50.00 – 40.32 = 9.68/100 = .0968. Chapter 5 Prepared by Samantha Gaies, M.A.

Null Hypothesis Testing Finding the p value for the non-TV sample of children is a major step toward using null hypothesis testing (NHT) to decide whether we can conclude that an entire population of children without TV would be any smarter than the current population of children with nearly universal access to TV. The major steps of NHT can be summarized as follows: Step 1: Assume that the worst-case scenario—i.e., the null hypothesis (H0)—is true. In this example, assume that the mean of the non-TV population is exactly the same as for the TV population—that is, µ0 = µ = 100. Step 2: Set an alpha (α) level such that if p is less than α you will reject (H0) as implausible. If a two-tailed test is decided upon, double your p value before comparing it to α. Step 3: Find the p value with respect to the null hypothesis distribution (NHD). For the one-sample case, the NHD is just the sampling distribution of the mean. Chapter 5 Prepared by Samantha Gaies, M.A.

The Logic of Null Hypothesis Testing It is only for the null hypothesis that we can easily find the distribution of possible results and then determine the probability of obtaining results as extreme as ours when H0 is really true. We hope that our results will look unlikely enough when H0 is true that p will be less than α, and we can therefore reject (H0) and declare our results to be statistically significant. The alternative hypothesis (HA) is simply the complement of H0; if H0 is that µ = 100, then HA is that µ ≠ 100 (anything but 100). NHT involves what is called an indirect proof. By casting doubt on the hypothesis we hope is not true (i.e., H0), we lend some support to the opposite hypothesis, HA, which is more consistent with our research hypothesis. Chapter 5 Prepared by Samantha Gaies, M.A.

Statistical Decisions Type I and Type II Errors If your p value is greater than your alpha level (normally .05), you need not “accept” the null hypothesis, but you cannot reject it. Some researchers say that they have to “retain” H0, whereas others prefer to say that they have “failed to reject” H0, or that they have insufficient evidence for rejecting it. Type I and Type II Errors Type I error – rejecting the null hypothesis when it is in fact true. This is the error rate that we control directly by choosing the level for alpha. Type II error – failing to reject the null hypothesis when it is in fact false. This error rate (beta) is not directly controlled, but it does increase if alpha is made smaller, and all else stays the same. Chapter 5 Prepared by Samantha Gaies, M.A.

More about Type I and Type II Errors A Type I error is like a false alarm, because you are saying that the mean of your sample is not consistent with the mean of the larger population, when in reality your sample mean is unusual due to chance factors alone. Alpha, the Type I error rate, can be defined as the percentage of null (ineffective) experiments that nonetheless attain statistical significance. A Type II error is a “miss,” because you are saying that the mean of your sample is consistent with the mean of the larger population, when in reality it is not, but the difference you found happened not to be large enough to attain statistical significance. You could reduce your Type I error rate by using a smaller alpha level (e.g., .01), but that would increase the rate of Type II errors (all else equal). You could reduce your Type II error rate by using a one-tailed rather than a two-tailed test, but you would have to convince your audience that there is no chance you would have tested the other tail. Chapter 5 Prepared by Samantha Gaies, M.A.

Prepared by Samantha Gaies, M.A. Try this example… National data shows that AP scores are normally distributed, with μ = 40 and σ = 5. For the past semester, N = 9 students were privately tutored for their AP psychology tests. The mean perfor- mance of the 9 students was 44. Is the mean of tutored students different from the mean of all students? State the hypotheses. Select the significance (alpha) level and the number of tails for the test. Calculate the test statistic (z score for groups, in this case). Find the p value corresponding to your test statistic and compare to alpha. Make a statistical decision with respect to the null hypothesis. Chapter 5 Prepared by Samantha Gaies, M.A.

Assumptions Underlying the One-Sample z Test The dependent variable was measured on an interval or ratio scale. The sample was drawn randomly from the population of interest. The dependent variable follows a normal distribution in the population. Because of the Central Limit Theorem, this assumption is not critical when the size of the sample is about 30 or more, unless the population distribution is nearly the opposite of the normal distribution. The standard deviation for the sampled population is the same as that of the comparison population. (This allows the SEM to be found from the known population standard deviation.) Chapter 5 Prepared by Samantha Gaies, M.A.