16 Nonparametric Tests Why Use Nonparametric Tests?

Slides:



Advertisements
Similar presentations
Prepared by Lloyd R. Jaisingh
Advertisements

Hypothesis Testing Steps in Hypothesis Testing:
COMPLETE BUSINESS STATISTICS
Chapter 16 Introduction to Nonparametric Statistics
Economics 105: Statistics Go over GH 11 & 12 GH 13 & 14 due Thursday.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Nonparametric Methods Chapter 15.
© 2003 Pearson Prentice Hall Statistics for Business and Economics Nonparametric Statistics Chapter 14.
Chapter 12 Chi-Square Tests and Nonparametric Tests
Chapter 14 Analysis of Categorical Data
Chapter 12 Chi-Square Tests and Nonparametric Tests
Topic 2: Statistical Concepts and Market Returns
Final Review Session.
1 Pertemuan 11 Analisis Varians Data Nonparametrik Matakuliah: A0392 – Statistik Ekonomi Tahun: 2006.
Chapter 11: Inference for Distributions
Inferences About Process Quality
1 Pertemuan 25 Metode Non Parametrik-1 Matakuliah: A0064 / Statistik Ekonomi Tahun: 2005 Versi: 1/1.
15-1 Introduction Most of the hypothesis-testing and confidence interval procedures discussed in previous chapters are based on the assumption that.
Chapter 15 Nonparametric Statistics
AM Recitation 2/10/11.
Non-parametric Dr Azmi Mohd Tamil.
Marketing Research, 2 nd Edition Alan T. Shao Copyright © 2002 by South-Western PPT-1 CHAPTER 17 BIVARIATE STATISTICS: NONPARAMETRIC TESTS.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 12-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Chapter 14: Nonparametric Statistics
Hypothesis Testing Charity I. Mulig. Variable A variable is any property or quantity that can take on different values. Variables may take on discrete.
14 Elements of Nonparametric Statistics
Bivariate Regression (Part 1) Chapter1212 Visual Displays and Correlation Analysis Bivariate Regression Regression Terminology Ordinary Least Squares Formulas.
NONPARAMETRIC STATISTICS
1 1 Slide © 2005 Thomson/South-Western AK/ECON 3480 M & N WINTER 2006 n Power Point Presentation n Professor Ying Kong School of Analytic Studies and Information.
Chapter 14 Nonparametric Statistics. 2 Introduction: Distribution-Free Tests Distribution-free tests – statistical tests that don’t rely on assumptions.
Copyright © 2012 Pearson Education. Chapter 23 Nonparametric Methods.
Previous Lecture: Categorical Data Methods. Nonparametric Methods This Lecture Judy Zhong Ph.D.
Nonparametric Statistics aka, distribution-free statistics makes no assumption about the underlying distribution, other than that it is continuous the.
Chi-Square Tests Chi-Square Tests Chapter1414 Chi-Square Test for Independence Chi-Square Tests for Goodness-of-Fit Copyright © 2010 by The McGraw-Hill.
© 2000 Prentice-Hall, Inc. Statistics Nonparametric Statistics Chapter 14.
Chapter 16 The Chi-Square Statistic
© Copyright McGraw-Hill CHAPTER 13 Nonparametric Statistics.
1 1 Slide © 2003 South-Western/Thomson Learning™ Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Biostatistics, statistical software VII. Non-parametric tests: Wilcoxon’s signed rank test, Mann-Whitney U-test, Kruskal- Wallis test, Spearman’ rank correlation.
Ordinally Scale Variables
Nonparametric Statistics. In previous testing, we assumed that our samples were drawn from normally distributed populations. This chapter introduces some.
1 Nonparametric Statistical Techniques Chapter 17.
Nonparametric Statistics
Lesson 15 - R Chapter 15 Review. Objectives Summarize the chapter Define the vocabulary used Complete all objectives Successfully answer any of the review.
Chapter 13 CHI-SQUARE AND NONPARAMETRIC PROCEDURES.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 11-1 Chapter 11 Chi-Square Tests and Nonparametric Tests Statistics for.
Nonparametric Methods and Chi-Square Tests Session 5.
GG 313 Lecture 9 Nonparametric Tests 9/22/05. If we cannot assume that our data are at least approximately normally distributed - because there are a.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
One-Sample Hypothesis Tests Chapter99 Logic of Hypothesis Testing Statistical Hypothesis Testing Testing a Mean: Known Population Variance Testing a Mean:
Statistics in Applied Science and Technology Chapter14. Nonparametric Methods.
CD-ROM Chap 16-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition CD-ROM Chapter 16 Introduction.
NON-PARAMETRIC STATISTICS
Nonparametric Statistics
© Copyright McGraw-Hill 2004
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Biostatistics Nonparametric Statistics Class 8 March 14, 2000.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Chapter 14 Nonparametric Methods and Chi-Square Tests
Copyright © 2010, 2007, 2004 Pearson Education, Inc Lecture Slides Elementary Statistics Eleventh Edition and the Triola Statistics Series by.
Lesson Test to See if Samples Come From Same Population.
Nonparametric statistics. Four levels of measurement Nominal Ordinal Interval Ratio  Nominal: the lowest level  Ordinal  Interval  Ratio: the highest.
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Non-parametric Tests Research II MSW PT Class 8. Key Terms Power of a test refers to the probability of rejecting a false null hypothesis (or detect a.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. CHAPTER 14: Nonparametric Methods to accompany Introduction to Business Statistics fifth.
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
1 Pertemuan 26 Metode Non Parametrik-2 Matakuliah: A0064 / Statistik Ekonomi Tahun: 2005 Versi: 1/1.
Slide 1 Shakeel Nouman M.Phil Statistics Nonparametric Methods and Chi-Square Tests (1) Nonparametric Methods and Chi-Square Tests (1) By Shakeel Nouman.
Two-Sample Hypothesis Testing
Chapter 18: The Chi-Square Statistic
Presentation transcript:

16 Nonparametric Tests Why Use Nonparametric Tests? Chapter Why Use Nonparametric Tests? One-Sample Runs Test Wilcox on Signed-Rank Test Mann-Whitney Test Kruskal-Wallis Test for Independent Samples Friedman Test for Related Samples Spearman Rank Correlation Test McGraw-Hill/Irwin Copyright © 2009 by The McGraw-Hill Companies, Inc. All rights reserved.

Why Use Nonparametric Tests? Parametric hypothesis tests require the estimation of one or more unknown parameters (e.g., population mean or variance). Often, unrealistic assumptions are made about the normality of the underlying population. Large sample sizes are often required to invoke the Central Limit Theorem.

Why Use Nonparametric Tests? Nonparametric or distribution-free tests - usually focus on the sign or rank of the data rather than the exact numerical value. - do not specify the shape of the parent population. - can often be used in smaller samples. - can be used for ordinal data.

Why Use Nonparametric Tests? Advantages and Disadvantages of Nonparametric Tests Table 16.1

Why Use Nonparametric Tests? Some Common Nonparametric Tests Figure 16.1

One-Sample Runs Test Wald-Wolfowitz Runs Test The one-sample runs test (Wald-Wolfowitz test) detects non-randomness. Ask – Is each observation in a sequence of binary events independent of its predecessor? A nonrandom pattern suggests that the observations are not independent. The hypotheses are H0: Events follow a random pattern H1: Events do not follow a random pattern 16-6

One-Sample Runs Test Wald-Wolfowitz Runs Test To test the hypothesis, first count the number of outcomes of each type. n1 = number of outcomes of the first type n2 = number of outcomes of the second type n = total sample size = n1 + n2 A run is a series of consecutive outcomes of the same type, surrounded by a sequence of outcomes of the other type.

DAAAAAAADDDDAAAAAAAADDAAAAAAAADDDDAAAAAAAAAA One-Sample Runs Test Wald-Wolfowitz Runs Test For example, consider the following series representing 44 defective (D) or acceptable (A) computer chips: DAAAAAAADDDDAAAAAAAADDAAAAAAAADDDDAAAAAAAAAA The grouped sequences are: A run can be a single outcome if it is preceded and followed by outcomes of the other type.

One-Sample Runs Test Wald-Wolfowitz Runs Test There are 8 runs (R = 8). n1 = number of defective chips (D) = 11 n2 = number of acceptable chips (A) = 33 n = total sample size = n1 + n2 = 11 + 33 = 44 The hypotheses are: H0: Defects follow a random sequence H1: Defects follow a nonrandom sequence

One-Sample Runs Test Wald-Wolfowitz Runs Test When n1 > 10 and n2 > 10, then the number of runs R may be assumed to be normally distributed with mean mR and standard deviation sR. calc

One-Sample Runs Test Wald-Wolfowitz Runs Test The test statistic is: For a given level of significance a, find the critical value za for a two-tailed test. Reject the hypothesis of a random pattern if z < -za or if z > +za . calc

One-Sample Runs Test Wald-Wolfowitz Runs Test Decision rule for large-sample runs tests: Figure 16.2

Wilcox on Signed-Rank Test The Wilcox on signed-rank test compares a single sample median with a benchmark using only ranks of the data instead of the original observations. It is used to compare paired observations. Advantages are - freedom from the normality assumption, - robustness to outliers - applicability to ordinal data. The population should be roughly symmetric.

Wilcox on Signed-Rank Test To compare the sample median (M) with a benchmark median (M0), the hypotheses are: When evaluating the difference between paired observations, use the median difference (Md) and zero as the benchmark.

Wilcox on Signed-Rank Test Calculate the difference between the paired observations. Rank the differences from smallest to largest by absolute value. Add the ranks of the positive differences to obtain the rank sum W.

Wilcox on Signed-Rank Test For small samples, a special table is required to obtain critical values. For large samples (n > 20), the test statistic is approximately normal. Use Excel or Appendix C to get a p-value. Reject H0 if p-value < a. calc

Mann-Whitney Test The Mann-Whitney test is a nonparametric test that compares two populations. It does not assume normality. It is a test for the equality of medians, assuming - the populations differ only in centrality, - equal variances The hypotheses are H0: M1 = M2 (no difference in medians) H1: M1 ≠ M2 (medians differ)

Mann-Whitney Test Performing the Test Step 1: Sort the combined samples from lowest to highest. Step 2: Assign a rank to each value. If values are tied, the average of the ranks is assigned to each. Step 3: The ranks are summed for each column (e.g., T1, T2). Step 4: The sum of the ranks T1 + T2 must be equal to n(n + 1)/2, where n = n1 + n2.

Mann-Whitney Test Performing the Test Step 5: Calculate the mean rank sums T1 and T2. Step 6: For large samples (n1 < 10, n2 > 10), use a z test. calc Step 7: For a given a, reject H0 if z < -za or z > +za

Kruskal-Wallis Test for Independent Samples The Kruskal-Wallis (K-W) test compares c independent medians, assuming the populations differ only in centrality. The K-W test is a generalization of the Mann-Whitney test and is analogous to a one-factor ANOVA (completely randomized model). Groups can be of different sizes if each group has 5 or more observations. Populations must be of similar shape but normality is not a requirement.

Kruskal-Wallis Test for Independent Samples Performing the Test First, combine the samples and assign a rank to each observation in each group. For example: When a tie occurs, each observation is assigned the average of the ranks. Table 16.7

Kruskal-Wallis Test for Independent Samples Performing the Test Next, arrange the data by groups and sum the ranks to obtain the Tj’s. Remember, STj = n(n+1)/2. Table 16.8

Kruskal-Wallis Test for Independent Samples Performing the Test The hypotheses to be tested are: H0: All c population medians are the same H1: Not all the population medians are the same For a completely randomized design with c groups, the tests statistic is where n = n1 + n2 + … + nc nj = number of observations in group j Tj = sum of ranks for group j calc

Kruskal-Wallis Test for Independent Samples Performing the Test The H test statistic follows a chi-square distribution with n = c – 1 degrees of freedom. This is a right-tailed test, so reject H0 if H > c2a or if p-value < a.

Friedman Test for Related Samples The Friedman test determines if c treatments have the same central tendency (medians) when there is a second factor with r levels and the populations are assumed to be the same except for centrality. This test is analogous to a two-factor ANOVA without replication (randomized block design) with one observation per cell. The groups must be of the same size. Treatments should be randomly assigned within blocks. Data should be at least interval scale.

Friedman Test for Related Samples In addition to the c treatment levels that define the columns, the Friedman test also specifies r block factor levels to define each row of the observation matrix. The hypotheses to be tested are: H0: All c populations have the same median H1: Not all the populations have the same median Unlike the Kruskal-Wallis test, the Friedman ranks are computed within each block rather than within a pooled sample.

Friedman Test for Related Samples Performing the Test First, assign a rank to each observation within each row. For example, within each Trial: When a tie occurs, each observation is assigned the average of the ranks.

Friedman Test for Related Samples Performing the Test Compute the test statistic: where r = the number of blocks (rows) c = the number of treatments (columns) Tj = the sum of ranks for treatment j calc

Friedman Test for Related Samples Performing the Test The Friedman test statistic F, follows a chi-square distribution with n = c – 1 degrees of freedom. Reject H0 if F > c2a or if p-value < a.

Spearman Rank Correlation Test Spearman’s rank correlation coefficient (Spearman’s rho) is an overall nonparametric test that measures the strength of the association (if any) between two variables. This method does not assume interval measurement. The sample rank correlation coefficient rs ranges from -1 < rs < +1.

Spearman Rank Correlation Test The sign of rs indicates whether the relationship is direct – ranks tend to vary in the same direction, or inverse – ranks tend to vary in opposite directions The magnitude of rs indicated the degree of relationship. If rs is near 0 – there is little or no agreement between rankings rs is near +1 – there is strong direct agreement rs is near -1 – there is strong inverse agreement

Spearman Rank Correlation Test Performing the Test First, rank each variable. For example, If more than one value is the same, assign the average of the ranks. Table 16.11

Spearman Rank Correlation Test Performing the Test The sums of ranks within each column must always be n(n+1)/2. Next, compute the difference in ranks di for each observation. The rank differences should sum to zero.

Spearman Rank Correlation Test Performing the Test Calculate the sample rank correlation coefficient rs. where di = difference in ranks for case i n = sample size For a right-tailed test, the hypotheses to be tested are H0: True rank correlation is zero (rs < 0) H1: True rank correlation is positive (rs > 0)

Spearman Rank Correlation Test Performing the Test If n is large (at least 20 observations), then rs may be assumed to follow the Student’s t distribution with degrees of freedom n = n - 1 Reject H0 if t > ta or if p-value < a. calc

Correlation versus Causation Caution: correlation does not prove causation. Correlations may prove to be “significant” even when there is no causal relation between the two variables. However, causation is not ruled out. Multiple causes may be present.

Applied Statistics in Business & Economics End of Chapter 16 16-37