Parametric Inferential Statistics. Types of Inference Estimation: On the basis of information in a sample of scores, we estimate the value of a population.

Slides:



Advertisements
Similar presentations
Chapter 7 Hypothesis Testing
Advertisements

Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Statistics.  Statistically significant– When the P-value falls below the alpha level, we say that the tests is “statistically significant” at the alpha.
Hypothesis Testing A hypothesis is a claim or statement about a property of a population (in our case, about the mean or a proportion of the population)
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 9 Hypothesis Testing Developing Null and Alternative Hypotheses Developing Null and.
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
1. Estimation ESTIMATION.
Review: What influences confidence intervals?
HYPOTHESIS TESTING Four Steps Statistical Significance Outcomes Sampling Distributions.
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
Statistics for the Social Sciences
Behavioural Science II Week 1, Semester 2, 2002
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Cal State Northridge  320 Ainsworth Sampling Distributions and Hypothesis Testing.
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
Statistics for the Social Sciences Psychology 340 Fall 2006 Hypothesis testing.
Chapter Sampling Distributions and Hypothesis Testing.
Hypothesis Tests for Means The context “Statistical significance” Hypothesis tests and confidence intervals The steps Hypothesis Test statistic Distribution.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Overview of Lecture Independent and Dependent Variables Between and Within Designs.
Statistics for the Social Sciences Psychology 340 Spring 2005 Hypothesis testing.
Statistics for the Social Sciences Psychology 340 Fall 2006 Hypothesis testing.
8-2 Basics of Hypothesis Testing
Inferences About Process Quality
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Overview Central Limit Theorem The Normal Distribution The Standardised Normal.
PSY 307 – Statistics for the Behavioral Sciences
Chapter 5For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 Suppose we wish to know whether children who grow up in homes without access to.
INFERENTIAL STATISTICS – Samples are only estimates of the population – Sample statistics will be slightly off from the true values of its population’s.
AM Recitation 2/10/11.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
Overview of Statistical Hypothesis Testing: The z-Test
Testing Hypotheses I Lesson 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics n Inferential Statistics.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Chapter 8 Hypothesis Testing 8-1 Review and Preview 8-2 Basics of Hypothesis.
Lecture Slides Elementary Statistics Twelfth Edition
Business Statistics - QBM117 Introduction to hypothesis testing.
Overview Definition Hypothesis
Sections 8-1 and 8-2 Review and Preview and Basics of Hypothesis Testing.
Tuesday, September 10, 2013 Introduction to hypothesis testing.
Chapter 8 Introduction to Hypothesis Testing
1/2555 สมศักดิ์ ศิวดำรงพงศ์
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
Overview Basics of Hypothesis Testing
Jan 17,  Hypothesis, Null hypothesis Research question Null is the hypothesis of “no relationship”  Normal Distribution Bell curve Standard normal.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 22 Using Inferential Statistics to Test Hypotheses.
The Argument for Using Statistics Weighing the Evidence Statistical Inference: An Overview Applying Statistical Inference: An Example Going Beyond Testing.
Chapter 8 Introduction to Hypothesis Testing
Education Research 250:205 Writing Chapter 3. Objectives Subjects Instrumentation Procedures Experimental Design Statistical Analysis  Displaying data.
Biostatistics Class 6 Hypothesis Testing: One-Sample Inference 2/29/2000.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Section Inference about Two Means: Independent Samples 11.3.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Chapter 8 Introduction to Hypothesis Testing ©. Chapter 8 - Chapter Outcomes After studying the material in this chapter, you should be able to: 4 Formulate.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
1 Chapter 9 Hypothesis Testing. 2 Chapter Outline  Developing Null and Alternative Hypothesis  Type I and Type II Errors  Population Mean: Known 
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Overview.
Chapter 8 Parameter Estimates and Hypothesis Testing.
Ex St 801 Statistical Methods Inference about a Single Population Mean.
Review I A student researcher obtains a random sample of UMD students and finds that 55% report using an illegally obtained stimulant to study in the past.
© Copyright McGraw-Hill 2004
Hypothesis Testing Introduction to Statistics Chapter 8 Feb 24-26, 2009 Classes #12-13.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Education 793 Class Notes Inference and Hypothesis Testing Using the Normal Distribution 8 October 2003.
Hypothesis Testing and Statistical Significance
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Chapter Nine Hypothesis Testing.
Hypothesis Testing: Hypotheses
Testing Hypotheses I Lesson 9.
Type I and Type II Errors
Presentation transcript:

Parametric Inferential Statistics

Types of Inference Estimation: On the basis of information in a sample of scores, we estimate the value of a population parameter. Hypothesis Testing: We determine how well the data in the sample fit with the hypothesis that in the population a parameter has a particular value.

Sampling Distribution The probability distribution of a statistic. Imagine this: –Draw an uncountably large number of samples, each with N scores. –For each sample, calculate a statistic (for example, the mean). –The distribution of these statistics is the sampling distribution

Desirable Properties of Estimators CURSE –Consistency –Unbiasedness –Resistance –Sufficiency –Efficiency

Unbiasedness Unbiasedness: The expected value (mean) of the statistic is exactly equal to the value of the parameter being estimated. The sample mean is an unbiased estimator of the population mean. The sample variance is an unbiased estimator of the population variance

The sample standard deviation is not an absolutely unbiased estimator of the population standard deviation. Consider this sampling distribution of variances: Sample VarianceProbabilityE(s 2 )= Ʃ P i s Population Variance =3 The population standard deviation must be SQRT(3) =

For these same samples, the sampling distribution of the standard deviations is: Oops, the expected value (1.707) is not equal to the value of the population parameter (1.732). Sample SDProbabilityE(s)= Ʃ P i s SQRT(2) = SQRT(4) = 2.51 Population Variance =1.707

Relative Efficiency The standard deviation of a sampling distribution is called its standard error. The smaller the standard error, the less error one is likely to make when using the statistic to estimate a parameter. Statistics with relatively low standard errors are known as efficient statistics.

We Play a Little Game You are allowed to drawn a sample of scores from a population whose mean Professor Karl knows. You then get to estimate that mean from the sample data. If your estimate is within one point of the true value of the parameter, you get an A in the course; otherwise you fail.

There are three different estimators available, X, Y, and Z. Each is absolutely unbiased. They differ greatly in dispersion. Which one will you use? Let’s look at their sampling distributions

Estimator X: SEM = 5 You have a 16% chance of getting that A.

Estimator Y: SEM = 1 You have a 68% chance of earning that A.

Estimator Z: SEM = 0.5 You have a 95% chance of earning that A.

Consistency With a consistent estimator, the standard error decreases as the sample size increases. For example, for the standard error of the mean:

Sufficiency A sufficient estimator uses all of the information in the sample. Consider the range. It uses information from the lowest score and the highest score. Consider the variance. It uses information from every score.

Resistance This is with respect to resistance to the effects of outliers. We have already seen that the median is much more resistant than is the mean.

Parameter Estimation There are two basic types –Point Estimation: We come up with a single value which is our best bet regarding what the value of the parameter is. –Interval Estimation: We come up with an interval of values which we are confident contains the true value of the parameter.

Confidence Coefficient CC is the subjective probability that the interval will include the true value of the parameter. Suppose CC =.95. Were we to construct an infinite number of confidence intervals, 95% of them would include the true value. α is the subjective probability that the interval will not contain the true value.

If Sampling Distribution is Normal The confidence interval will be Where theta-hat is the estimator Sigma theta-hat is the standard error And z is the how far you need go out from the mean of the sampling distribution to encompass the middle CC proportion of the distribution.

Example At Bozo College, IQ is normally distributed with a standard deviation of 15. We want to estimate the mean. A sample of 1 has mean = 110. A 95% CI for the mean is 110  1.96(15) = [80.6, 139.4].

Hypothesis Testing The null hypothesis states that some parameter has a particular value. Example:  = 100 The alternative hypothesis states that the null hypothesis is not correct. Example:   100 These are nondirectional hypotheses – the alternative does not predict the direction of difference from 100.

Directional Hypotheses The alternative does predict a direction. Example –H  is   100 –H 1 is  > 100 Another example –H  is   100 –H 1 is  < 100

What the Null Usually Is It is usually that the value of the correlation between two variables or sets of variables is zero. That is, the variables are not related. The researcher usually thinks that the null is incorrect.

What the Null Rarely Is The prediction from a mathematical model being tested. For example, mean weight loss during this treatment is pounds. In this case, if we find the null to be substantially incorrect, we need to reject or revise the mathematical model (aka “theory”).

Testing the Null Gather relevant data Determine how well the data fit with the null hypothesis, using a statistic called “p,” the level of significance. p is the probability of obtaining a sample as or more discrepant with the H  than is that which we did obtain, assuming that the H  is true.

The higher this p, the better the fit between the data and the H . If this p is low we have cast doubt upon the H . If p is very low, we reject the H . How low is very low? Very low is usually the decision rule most often used in Psychology is Reject the null if p .05.

Simple Example H  : mean IQ of my extended family is 145. H 1 : no it isn’t Sample of one score: IQ = 110. Assume SD = 15 and population normal. z = ( ) / 15 = ‑ 2.33 Is that z score unusual enough for us to reject the null hypothesis?

From the normal curve table, we determine that were the null true, we would get a z that far from zero only 1.98% of the time. That is, p = This is less than.05, so we reject the null hypothesis and conclude that the mean is not 145.

This Was a Two-Tailed Test

Do You Need a p Value? Journals act like it is not real science unless there is a p value, but all you really need is a CI. 95% CI = 110  1.96(15) = [80.6, 139.4] We are 95% confident that the mean is between 80.6 and That excludes 145, so we at least 95% confident that the mean is not 145  Reject the null.

A More Traditional Approach Think about the sampling distribution of the test statistic (z). The nonrejection region is the area with values that would not lead to rejection of the null. The rejection region is the area with values that would lead to rejection of the null.

CV is the “critical value,” the boundary between rejection and nonrejection regions

Decision Rule If |z| > 1.96, then p <.05 and we reject the null. Otherwise, we retain the null. We never figure out exactly what the value of p is. I strongly prefer that you use the modern approach, where you find the exact value of p.

“Confusion Matrix” The True Hypothesis Is DecisionThe H 1 The H  Reject H  Assert H 1 correct decision (power) Type I error (  ) Retain H  Do not assert H 1 Type II error (  ) correct decision (1-  )

Signal Detection Is the Signal Really There? PredictionYesNo Signal is there True Positive [hit] (power) False Positive (  ) Signal is not there False Negative [miss] (  ) True Negative (1-  )

Relative Seriousness of Type I and II Errors Tumor rate in rats is 10%. Treat them with suspect drug. –H  : rate  10%; drug is safe –H 1 : rate > 10%; drug is not safe Type I Error: The drug is safe, but you conclude it is not. Type II Error: The drug is not safe, but you conclude it is.

Testing experimental blood pressure drug. –H  : Drop in BP  0; drug does not work –H 1 : Drop in BP > 0; drug does work Type I Error: The drug does not lower BP, but you conclude it does. Type II Error: The drug does lower BP, but you conclude it does not.

Directional Hypotheses The alternative hypothesis predicts the direction of the difference between the actual value of the parameter and the hypothesized value. The rejection region will be in only one tail – the side to which the alternative hypothesis points. H  :   145 versus H 1 :  > 145

z = ( ) / 15 = ‑ 2.33 Our test statistic is not in the rejection region, we must retain the null. The p value will be P(z > -2.33), which is equal to The data fit very well with the null hypothesis that   145.

Change the Prediction H  :   145 versus H 1 :  < 145. The rejection region is now in the lower tail. If z  , we reject the null. Our z is still -2.33, we reject the null.

The p value is now P(z < -2.33), which is equal to We do not double the p value as we would with nondirectional hypotheses.

Pre- or Post-diction? If you correctly predicted (H 1 ) the direction, the p value will be half what it would have been with nondirectional hypotheses. That gives you more power. BUT others will suspect that you postdicted the direction of the difference.

Frequency of Type I Errors If we are in that universe where the null hypothesis is always true, using the.05 criterion of statistical significance, We should make Type I errors 5% of the time. There may be factors that inflate this percentage Failures to reject the null are not often published.

“Publish or Perish.” May produce unconscious bias against keeping the null, affecting data collection and analysis. May lead to fraud.

The File Drawer Problem 100 researchers test the same absolutely true null hypothesis. 5 get “significant” results. Joyfully, they publish, unaware that their conclusions are Type I errors. The other 95 just give up and put their “not significant” results in a folder in the filing cabinet.

Is an Exact Null Ever True? The null is usually that the correlation between two things is zero. Correlation coefficients are continuously distributed between -1 and +1. The probability of any exact value of a continuous variable (such as  = 0) is vanishingly small. But a null may be so close to true that it might as well be true.

Doggie Dance