1 Psych 5500/6500 Data Transformations Fall, 2008.

Slides:



Advertisements
Similar presentations
Psych 5500/6500 t Test for Two Independent Groups: Power Fall, 2008.
Advertisements

1 COMM 301: Empirical Research in Communication Lecture 15 – Hypothesis Testing Kwan M Lee.
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
PSY 307 – Statistics for the Behavioral Sciences Chapter 20 – Tests for Ranked Data, Choosing Statistical Tests.
5/15/2015Slide 1 SOLVING THE PROBLEM The one sample t-test compares two values for the population mean of a single variable. The two-sample test of a population.
Models with Discrete Dependent Variables
Significance Testing Chapter 13 Victor Katch Kinesiology.
MARE 250 Dr. Jason Turner Hypothesis Testing II To ASSUME is to make an… Four assumptions for t-test hypothesis testing: 1. Random Samples 2. Independent.
MARE 250 Dr. Jason Turner Hypothesis Testing II. To ASSUME is to make an… Four assumptions for t-test hypothesis testing:
1/55 EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2008 Chapter 10 Hypothesis Testing.
MARE 250 Dr. Jason Turner Hypothesis Testing III.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 5): Outliers Fall, 2008.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 13 Using Inferential Statistics.
Today Concepts underlying inferential statistics
The one sample t-test November 14, From Z to t… In a Z test, you compare your sample to a known population, with a known mean and standard deviation.
Getting Started with Hypothesis Testing The Single Sample.
Chapter 9: Introduction to the t statistic
Chapter 14 Inferential Data Analysis
Richard M. Jacobs, OSA, Ph.D.
Slide 1 Testing Multivariate Assumptions The multivariate statistical techniques which we will cover in this class require one or more the following assumptions.
Non-parametric statistics
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Inferential Statistics
Choosing Statistical Procedures
Analysis of Variance. ANOVA Probably the most popular analysis in psychology Why? Ease of implementation Allows for analysis of several groups at once.
AM Recitation 2/10/11.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
Jeopardy Hypothesis Testing T-test Basics T for Indep. Samples Z-scores Probability $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
Fundamentals of Hypothesis Testing: One-Sample Tests
Research Methods for Counselors COUN 597 University of Saint Joseph Class # 9 Copyright © 2014 by R. Halstead. All rights reserved.
Education 793 Class Notes T-tests 29 October 2003.
16-1 Copyright  2010 McGraw-Hill Australia Pty Ltd PowerPoint slides to accompany Croucher, Introductory Mathematics and Statistics, 5e Chapter 16 The.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Hypothesis Testing: One Sample Cases. Outline: – The logic of hypothesis testing – The Five-Step Model – Hypothesis testing for single sample means (z.
Measures of Variability In addition to knowing where the center of the distribution is, it is often helpful to know the degree to which individual values.
One-sample In the previous cases we had one sample and were comparing its mean to a hypothesized population mean However in many situations we will use.
Stats 95.
1 Psych 5500/6500 t Test for Two Independent Means Fall, 2008.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 4): Power Fall, 2008.
1 Psych 5500/6500 Standard Deviations, Standard Scores, and Areas Under the Normal Curve Fall, 2008.
Biostatistics, statistical software VII. Non-parametric tests: Wilcoxon’s signed rank test, Mann-Whitney U-test, Kruskal- Wallis test, Spearman’ rank correlation.
Hypothesis Testing A procedure for determining which of two (or more) mutually exclusive statements is more likely true We classify hypothesis tests in.
Inference and Inferential Statistics Methods of Educational Research EDU 660.
Chapter 6 USING PROBABILITY TO MAKE DECISIONS ABOUT DATA.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 1): Two-tail Tests & Confidence Intervals Fall, 2008.
1 Psych 5500/6500 Introduction to the F Statistic (Segue to ANOVA) Fall, 2008.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
© Copyright McGraw-Hill 2000
Central Tendency & Dispersion
1 Psych 5500/6500 Measures of Variability Fall, 2008.
KNR 445 Statistics t-tests Slide 1 Introduction to Hypothesis Testing The z-test.
DTC Quantitative Methods Bivariate Analysis: t-tests and Analysis of Variance (ANOVA) Thursday 14 th February 2013.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
 Assumptions are an essential part of statistics and the process of building and testing models.  There are many different assumptions across the range.
Handout Six: Sample Size, Effect Size, Power, and Assumptions of ANOVA EPSE 592 Experimental Designs and Analysis in Educational Research Instructor: Dr.
Non-parametric Approaches The Bootstrap. Non-parametric? Non-parametric or distribution-free tests have more lax and/or different assumptions Properties:
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
Hypothesis Tests u Structure of hypothesis tests 1. choose the appropriate test »based on: data characteristics, study objectives »parametric or nonparametric.
Chapter 7: The Distribution of Sample Means
Two-Sample-Means-1 Two Independent Populations (Chapter 6) Develop a confidence interval for the difference in means between two independent normal populations.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Bootstrapping and Randomization Techniques Q560: Experimental Methods in Cognitive Science Lecture 15.
STAT 312 Chapter 7 - Statistical Intervals Based on a Single Sample
I. ANOVA revisited & reviewed
Logic of Hypothesis Testing
Chapter 15 Strategies When Population Distributions are Not Normal:
What are their purposes? What kinds?
Presentation transcript:

1 Psych 5500/6500 Data Transformations Fall, 2008

2 Data Transformations We are now going to examine an option we have if we find that our data appear to violate the assumption that the populations are normally distributed or the assumption that the populations have the same variance. This option is to transform our data to better fit the assumptions. First we will look at some options on how to do this, and then we will turn to general issues and concerns about transforming data.

3 You Might Not Need to Worry About the Assumptions Remember that violation of the assumption of normality grows less serious as the N of our samples increase; and that violation of the assumption of homogeneity of variances is not important when the N’s of our groups are roughly equal. Thus you may not need to turn to transformations if the N’s of your groups are largish and approximately the same size.

4 Transformations to Take Care of Anticipated Problems Certain types of measurements routinely produce samples that violate one or both of the assumptions of normality and equal variances. The solutions are fairly well established and it is always better to anticipate a priori what transformation might be appropriate. Post hoc decisions to transform the data face two criticisms; 1) are you changing you data just to get the results you want?, and 2) is your population actually ok and you are changing your data (and thus the population it represents) to fix a problem that appeared in your sample just due to chance (i.e. the populations were actually ok)?

5 Reaction Times Reaction times are often positively skewed. Transformations that are recommended for reaction times (or any positively skewed data) are: Try both 1 and 2 and see which works best.

6 Counts If the variable involves counting something (e.g. occurrences of some behavior) then there could be a problem, particularly if low counts (around zero) occur. The floor effect of not being able to have a score below zero will effect the normality of the data, and if one of the groups has more of a floor effect than the other then the groups will have different variances. A recommended transformation in this case is:

7 Proportions Proportions as measures suffer from several problems, two of them are: 1.Proportions around.5 have greater variance than proportions close to 0 or 1 (due to floor and ceiling effects). Thus the variance of two groups will differ if one has proportions closer to.5 than the other. 2.Many people consider a difference in the mean proportion of two groups of.02 and.08 (a difference of.06 but also a quadrupling of the proportion) to be greater than a difference of.48 and.54 (also a difference of.06 but a much smaller change in terms of ratios).

8 Proportions (cont.) Stretching out both tails of the distribution can help with both of those problems. Transforms that can accomplish this are: arcsine transform: logit transform Try them both and see which works best.

9 Power Transformations The general formula of raising Y to some power (‘pow’), as in, Y transform =Y pow can be used to reshape the distribution in a variety of ways. We have already seen one use of this, the square root of Y, which is the same as Y 0.5 (in SPSS that would be Y**.5). Here we are venturing into the territory of post hoc transformations, where you try out various transformations until the distribution is the shape that you want.

10 Power Transformation Strategies Try various values for power, you might try the following to see which works best: pow = 3, 2, 0.5, ‘0’, -0.5, -1, -2. Obviously, Y 1 isn’t on the list as Y 1 = Y The ‘0’ is in quote marks because raising Y to the power of 0 will result in changing all of your scores to ‘1’. In place of using pow=0, substitute Y transform = log 10 Y.

11 Power Transformation Strategies Instead of pure trial and error, you can use the general strategy of trying a value of pow>1 if you want to bring in a long negative tail, and pow<1 if you want to bring in a long positive tail. The further from ‘1’ in either direction the more the tail will be pulled in.

12 Power Transformation Strategies Some statistical programs will run through possible values for the power of Y and then report which one best turns the data into a normal curve (SPSS apparently doesn’t offer that). It is important to note that whether you figure out a good value for p or the computer does, such a transformation would be purely post hoc.

13 Rank Transformation If you cannot transform you data in such a way that it becomes more normally distributed then an interesting option of last resort is to transform your data into rank scores. Each score is transformed into a rank score which indicates where it falls in a list of all the scores in the study for that variable. If observations are tied, then each observation receives the mean of the ranks they would have received if they weren’t tied.

14 For example, the (cardinal) data –Group 1: 3.4, 7.2, 5.1, 6.9 –Group 2: 7.2, 5.1, 7.2, 8.4 Would first be put into one list and ordered: –Y= 3.4, 5.1, 5.1, 6.9, 7.2, 7.2, 7.2, 8.4 Scores of 5.1 come in 2 nd and 3 rd on the list, so they each get a rank of (2+3)/2=2.5. Scores of 7.2 come in 5 th, 6 th, and 7 th, so they each get a rank of (5+6+7)/3=6. Y transform =1, 2.5, 2.5, 4, 6, 6, 6, 8 Transformed (to ranked) data: –Group 1: 1, 6, 2.5, 4 –Group 2: 6, 2.5, 6, 8

15 Upside of rank transformations: 1.Control the effect of outliers (because distance to extreme score is reduced). 2.While they don’t create normal data they generally reduce problems of thick tails. 3.While they don’t ensure homogeneity of variance they generally prevent very large differences in variance. Downside of rank scores: you are losing information when you move from cardinal to rank scores.

16 Parametric and Nonparametric Tests Statistical procedures that are based upon certain assumptions about the population being true are called parametric tests. The t tests--in fact every test we will look at this semester--are parametric tests. Statistical procedures that are not based upon certain assumptions about the population being true are called nonparametric tests. These tests are useful in that they can be applied to situations where assumptions certainly are not met, but nonparametric tests tend to have low power.

17 Parametric Tests of Rank Data Most nonparametric tests require rank data. To apply a parametric test to rank data is an interesting half- way step between parametric and non-parametric tests. By ranking the data you are not meeting the assumptions of the parametric tests but you are violating the assumptions in a way that tends to be ok (e.g. creating thin tails rather than fat tails). While it is still better not to have to transform cardinal data to rank data (because you lose so much information and probably lose power) if you do it may be better to use a parametric test than a non- parametric on the data.

18 Advantages The advantages of using a parametric test on rank data vs. using a nonparametric test are: 1.You don’t have to learn another whole class of statistical tests (nonparametric tests). 2.In some cases the parameter test of rank data may be just as good or even better than the nonparametric test (see Judd & McClelland, 1989, and Ruxton, 2006). Judd, C. M. & McClelland, G. H. (1989). Data analysis: A model- comparison approach. New York: Harcourt College Publishers. Ruxton, G. D. (2006). The unequal variance t-test is an underused alternative to Student's t-test and the Mann–Whitney U test. Behavioral Ecology, 17,

19 The Mann–Whitney U test is the nonparametric equivalent to a t test. The following table compares type 1 error rate of the U test to that of Welch’s t’ in a Monte Carlo study (Zimmerman and Zumbo, 1993, as adapted and cited in Ruxton, 2006) when both tests were applied to rank data. N1N1 N2N2 σ 1 / σ 2 Ut’

20 Issues and Concerns Regarding Transformations 1)You are moving the data further away from reflecting reality in an attempt to better make if fit your tool. 2)If your original measure is more relevant to your theory than the transform, then can you generalize the analysis of the transformed data to the theory?

21 Issues and Concerns (cont.) 3)Transforming your data to fit the tool may or may not be justified, but transforming your data to make the results fit your theory is definitely not justified. Comment: this is why transforms that are selected a priori due to known characteristics of the measure have an advantage over post hoc transforms.

22 Issues and Concerns (cont.) 4)Post hoc transforms might end up fixing problems that are a fluke, that only appear in this particular sample. This shows one of the advantages of replicating studies. If the data in the first study suggests a particular transform would be useful, you might want to use it, but then withhold final judgment on its appropriateness until a second study confirms its utility.