Comparing Groups.

Slides:



Advertisements
Similar presentations
Dr. AJIT SAHAI Director – Professor Biometrics JIPMER, Pondicherry
Advertisements

1 Chapter 20: Statistical Tests for Ordinal Data.
More on ANOVA. Overview ANOVA as Regression Comparison Methods.
Research methods and statistics
PSY 1950 Nonparametric Statistics November 24, 2008.
Inferential Statistics
Review I volunteer in my son’s 2nd grade class on library day. Each kid gets to check out one book. Here are the types of books they picked this week:
Hypothesis Testing Charity I. Mulig. Variable A variable is any property or quantity that can take on different values. Variables may take on discrete.
Non-Parametric Methods Professor of Epidemiology and Biostatistics
Non-parametric Tests. With histograms like these, there really isn’t a need to perform the Shapiro-Wilk tests!
Parametric & Non-parametric Parametric Non-Parametric  A parameter to compare Mean, S.D.  Normal Distribution & Homogeneity  No parameter is compared.
Biostatistics, statistical software VII. Non-parametric tests: Wilcoxon’s signed rank test, Mann-Whitney U-test, Kruskal- Wallis test, Spearman’ rank correlation.
Ordinally Scale Variables
Comparing Samples. Last Time I talked about what could go wrong in an experiment where you compared a sample mean against a population with a known population.
Experimental Psychology PSY 433 Appendix B Statistics.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Quantitative methods and R – (2) LING115 December 2, 2009.
Nonparametric statistics. Four levels of measurement Nominal Ordinal Interval Ratio  Nominal: the lowest level  Ordinal  Interval  Ratio: the highest.
Inferential Statistics Assoc. Prof. Dr. Şehnaz Şahinkarakaş.
I. ANOVA revisited & reviewed
One-Way Between-Subjects Design and Analysis of Variance
Inferential Statistics
Nonparametric Statistics
Step 1: Specify a null hypothesis
Causality, Null Hypothesis Testing, and Bivariate Analysis
Comparing Two Means Prof. Andy Field.
Non-parametric test ordinal data
DTC Quantitative Methods Bivariate Analysis: t-tests and Analysis of Variance (ANOVA) Thursday 20th February 2014  
Done by : Mohammad Da’as Special thanks to Dana Rida and her slides 
Nonparametric Tests IPS Chapter : The Wilcoxon Rank Sum Test
Analysis of variance ANOVA.
INF397C Introduction to Research in Information Studies Spring, Day 12
Introduction to Statistics: Probability and Types of Analysis
Non-Parametric Tests 12/1.
Comparing several means: ANOVA (GLM 1)
Non-Parametric Tests 12/1.
Understanding and conducting
Statistics.
Non-Parametric Tests 12/6.
Hypothesis testing. Chi-square test
University of Warwick, Department of Sociology, 2014/15 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard) Analysing Means II: Nonparametric techniques.
Non-Parametric Tests.
Planned Comparisons & Post Hoc Tests
Nonparametric Statistical Methods: Overview and Examples
Nonparametric Statistics
Introduction to Statistics
Comparing Several Means: ANOVA
Nonparametric Statistical Methods: Overview and Examples
Single-Factor Studies
Single-Factor Studies
Analysis of Variance (ANOVA)
Nonparametric Tests BPS 7e Chapter 28 © 2015 W. H. Freeman and Company.
十二、Nonparametric Methods (Chapter 12)
Some Nonparametric Methods
Statistics in SPSS Lecture 8
Nonparametric Statistical Methods: Overview and Examples
Reasoning in Psychology Using Statistics
I. Statistical Tests: Why do we use them? What do they involve?
Nonparametric Statistical Methods: Overview and Examples
Reasoning in Psychology Using Statistics
Non-parametric tests, part A:
Psych 231: Research Methods in Psychology
Statistics for the Behavioral Sciences
Analysis of Variance: repeated measures
Psych 231: Research Methods in Psychology
Chapter 10 Introduction to the Analysis of Variance
Tests of inference about 2 population means
InferentIal StatIstIcs
Type I and Type II Errors
Univariate analysis Önder Ergönül, MD, MPH June 2019.
Presentation transcript:

Comparing Groups

Non-parametric Tests You now know about comparing means using the z or t distributions. Those are both theoretical distributions that depend on your data being approximately normally distributed. You can lean on the CLT to help you get a normally distributed sampling distribution but there are times when your sample is small and you strongly suspect the data (and more importantly, the sampling distribution) are not normally distributed.

Wilcoxon’s Signed-rank It parallels the one sample t-test. It is frequently done on the difference between two scores and you want to see if the data is different from zero. Rank the absolute scores from lowest to highest. Call the lowest person 1, the 2nd lowest 2, etc. Add up the rank scores for the people below 0 (call that total R-) and add up the rank scores for the people above 0 (call that total R+).

How Signed Rank Works You use the ranks rather than the original values and get an approximate t score and test it against the t distribution.

Example Wilcoxon Signed Rank See nonParametric.egp

Assumptions The Signed Rank test makes the assumption that the observations are independent and the data’s distribution is approximately symmetrical. If you can’t say the data is symmetrical, there is another non-parametric alternative: the sign test. If the data really are normally distributed and you are interested in comparing means, the t-test is more powerful and should be used.

Wilcoxon Rank-sum (aka Mann-Whitney U) It is like the two sample t-test. It, like the t-test, assumes that observations are independent (e.g., different people) but it does not assume the data are approximately normally distributed. It is often applied to Likert scale data. Instead of using the original scaled scores, you convert the data to ranks (from low to high) and look to see if the average ranks are different between the two groups.

How it Works R1 sum of the ranks for the first group. R2 sum of the ranks for the second group. Expected rank if half were from each group Tweak this formula slightly if lots of ties

Other Nonparametric Methods These methods are considered non-parametric because they are playing with the rank orders and you are not directly appealing to the data being described by a couple of parameters (the mean and SD typically). There are truly nonparametric methods for testing for differences between groups.

Permutation Theory… If you collect scores (data) from two groups, at the end of the day you will have two columns. One column indicates group membership and the other has the scores. You can get the sum of the scores (or any other statistic) for the two groups. Then shuffle the scores around randomly (across both groups) and recalculate the sums (statistics). You can then do this a LOT of times (or until you have done all the possible reshufflings – every permutation). After doing this many times you then look to see how unusual your original total is relative to the possible scores you saw with the permutations. The percentile rank gives you a p-value to say how odd your data is.

In the beginning…. Back in the beginning of time, I talked about modeling. You want to make a prediction about a group’s outcome given one or more predictors. Think about group membership as a model. If you are in a baseline group, you get assigned one mean and if you are in another group, you add or subtract relative to the baseline.

T-Test as a Model Last week I talked about building a model where you are modeling the mean score for a group or a difference between groups. Predicting the means extends naturally to comparing more than 1 group.

T-test as a Model I am making a prediction about the outcome mean. Do I want to use the overall mean or will I do “better” if I call one level the baseline mean and then add something to it to represent the other mean? I will naturally test to see if that amount of difference could happen by chance alone.

T as a Model Score as 1 or 0 if you are in the other group.

ANOVA as a Model Exactly the same logic extends to comparing 3 or more groups. You have a baseline level and then you add (or subtract) something from the mean if you are not in the reference group. As I mentioned, the statistical method to see whether the amount of difference is meaningful is strange because it is based on variances.

Reduced Variance With the T or Z distributions you got excited if your sample mean was far from the proposed null population mean. Here, you get excited if the ratio of the two variances is far from 1. You need a distribution that can describe the ratio of two variances. That distribution is the F. It has a parameter to describe the number of subjects in the two halves of the fraction.

Decomposition of Variance The way the result of an ANOVA is typically shown is the variance within groups relative to the variance between groups.

One-way ANOVA If your model has one categorical predictor variable, it is called a one-way ANOVA. You will do an ANOVA to test the global hypothesis that the means of the groups are the same. A statistically significant ANOVA only tells you some mean is different from some other mean (a statistically significant ANOVA does mean that all means are different from each other – although that rarely happens). If this omnibus test says there is a difference between the means, you will then do tests to see which mean is different.

Before You do ANOVA You assume the subjects are randomly selected from (or representative of) a large population. The samples are obtained independently. If your subjects were not independent or if you measure the same person repeatedly, you will want to use ANOVAs specially designed for repeated measures or more complicated statistics. As with a t-test, you assume that the population is approximately normally distributed (or you have a large sample).

Why bother learning this? Why not just do a pile of t-tests first? Multiple comparisons: Your experiment-wise error is not .05, it is closer to p ≈ 1-.95numberOfComparisons Pooled standard errors: You want to make the best possible estimate of the variability in the population (SEs). Therefore you want to combine the information across the different subgroups. Lack of independence in the tests: If you know about group1 vs. group2 and if you also know about group1 vs. group3, you can make inferences (sort of) about group2 vs. group3.

Doing ANOVA See Enterprise Guide Project for ANOVA.

After You do ANOVA Do you have planned comparisons in mind? Planning on looking at a reference group vs. all the others (e.g., placebo vs. 3 dosages) --- Dunnett’s test. Planned comparisons on a couple of means: The Bonferroni test is very conservative. That is, its power is too low. It is reasonable for 3 or 4 comparisons. Compare all means: If you get 3 statisticians in a room and ask them for the best test to do post hoc group mean comparisons, you will get at least 4 answers…. Bonferroni - conservative Tukey HSD – middle of the road Student-Newman-Keuls - liberal

Non-Parametric ANOVA? The Kruskal-Wallis test works if you do not have repeated measures. The repeated measures non-parametric ANOVA analog is Friedman’s test. (It is like Wilcoxon rank-sum.)

Two-way ANOVA You can extend the same modeling ideas to having two predictors. You pick a baseline level in each predictor, then add something to change the baseline mean to represent the change in the mean associated with group membership for the other levels of the first predictor. You do the same process to represent the change from the baseline level in the other variable. But… what happens if there are special impacts on the means from some combinations of the levels over the two predictors?