Relationship Mining Correlation Mining Week 5 Video 1.

Slides:



Advertisements
Similar presentations
Locating Variance: Post-Hoc Tests
Advertisements

Chapter 7 Hypothesis Testing
1 Health Warning! All may not be what it seems! These examples demonstrate both the importance of graphing data before analysing it and the effect of outliers.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 April 18, 2012.
Statistical Issues in Research Planning and Evaluation
Chapter 10 Section 2 Hypothesis Tests for a Population Mean
Statistical Techniques I EXST7005 Lets go Power and Types of Errors.
C82MST Statistical Methods 2 - Lecture 4 1 Overview of Lecture Last Week Per comparison and familywise error Post hoc comparisons Testing the assumptions.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Multiple testing adjustments European Molecular Biology Laboratory Predoc Bioinformatics Course 17 th Nov 2009 Tim Massingham,
Hypothesis testing Week 10 Lecture 2.
Locating Variance: Post-Hoc Tests Dr James Betts Developing Study Skills and Research Methods (HL20107)
HYPOTHESIS TESTING Four Steps Statistical Significance Outcomes Sampling Distributions.
07/01/15 MfD 2014 Xin You Tai & Misun Kim
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
PSY 1950 Post-hoc and Planned Comparisons October 6, 2008.
PSY 1950 Confidence and Power December, Requisite Quote “The picturing of data allows us to be sensitive not only to the multiple hypotheses that.
One-way Between Groups Analysis of Variance
© 2013 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Introductory Statistics: Exploring the World through.
Today Concepts underlying inferential statistics
Richard M. Jacobs, OSA, Ph.D.
Introduction to Testing a Hypothesis Testing a treatment Descriptive statistics cannot determine if differences are due to chance. A sampling error occurs.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
False Discovery Rate (FDR) = proportion of false positive results out of all positive results (positive result = statistically significant result) Ladislav.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
Multiple testing in high- throughput biology Petter Mostad.
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 9. Hypothesis Testing I: The Six Steps of Statistical Inference.
Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 1 False discovery rate: setting the probability.
Chapter 8 Introduction to Hypothesis Testing
Intermediate Applied Statistics STAT 460
Testing Theories: Three Reasons Why Data Might not Match the Theory Psych 437.
From last lecture (Sampling Distribution): –The first important bit we need to know about sampling distribution is…? –What is the mean of the sampling.
Individual values of X Frequency How many individuals   Distribution of a population.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Association between 2 variables
Research Process Parts of the research study Parts of the research study Aim: purpose of the study Aim: purpose of the study Target population: group whose.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 27, 2013.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
PSY2004 Research Methods PSY2005 Applied Research Methods Week Five.
Back to basics – Probability, Conditional Probability and Independence Probability of an outcome in an experiment is the proportion of times that.
Statistical Power The power of a test is the probability of detecting a difference or relationship if such a difference or relationship really exists.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Probability Lecture 2. Probability Why did we spend last class talking about probability? How do we use this?
Digression - Hypotheses Many research designs involve statistical tests – involve accepting or rejecting a hypothesis Null (statistical) hypotheses assume.
9.2 Linear Regression Key Concepts: –Residuals –Least Squares Criterion –Regression Line –Using a Regression Equation to Make Predictions.
DTC Quantitative Methods Bivariate Analysis: t-tests and Analysis of Variance (ANOVA) Thursday 14 th February 2013.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Statistical Techniques
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 6 –Multiple hypothesis testing Marshall University Genomics.
Outline of Today’s Discussion 1.Independent Samples ANOVA: A Conceptual Introduction 2.Introduction To Basic Ratios 3.Basic Ratios In Excel 4.Cumulative.
Independent Samples ANOVA. Outline of Today’s Discussion 1.Independent Samples ANOVA: A Conceptual Introduction 2.The Equal Variance Assumption 3.Cumulative.
Statistics (cont.) Psych 231: Research Methods in Psychology.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Posthoc Comparisons finding the differences. Statistical Significance What does a statistically significant F statistic, in a Oneway ANOVA, tell us? What.
Estimating the False Discovery Rate in Genome-wide Studies BMI/CS 576 Colin Dewey Fall 2008.
Do data match an expected ratio
Core Methods in Educational Data Mining
Setting significance levels at the correct level
Core Methods in Educational Data Mining
I. Statistical Tests: Why do we use them? What do they involve?
Chapter 12 Power Analysis.
HUDM4122 Probability and Statistical Inference
False discovery rate estimation
What determines Sex Ratio in Mammals?
Presentation transcript:

Relationship Mining Correlation Mining Week 5 Video 1

Relationship Mining  Discover relationships between variables in a data set with many variables  Many types of relationship mining

Correlation Mining  Perhaps the simplest form of relationship mining  Finding substantial linear correlations between variables  Remember this from earlier in the class?  In a large set of variables

Use Cases  You have 100 variables, and you want to know how each one correlates to a variable of interest  Not quite the same as building a prediction model  You have 100 variables, and you want to know how they correlate to each other

Many Uses…  Studying relationships between questionnaires on traditional motivational constructs (goal orientation, grit, interest) and student reasons for taking a MOOC  Correlating features of the design of mathematics problems to a range of outcome measures  Correlating features of schools to a range of outcome measures

The Problem  You run 100 correlations (or 10,000 correlations)  9 of them come up statistically significant  Which ones can you “trust”?

If you…  Set p=0.05  Then, assuming just random noise  5% of your correlations will still turn up statistically significant

The Problem  Comes from the paradigm of conducting a single statistical significance test

The Solution  Adjust for the probability that your results are due to chance, using a post-hoc control

Two paradigms  FWER – Familywise Error Rate  Control for the probability that any of your tests are falsely claimed to be significant (Type I Error)  FDR – False Discovery Rate  Control for the overall rate of false discoveries

Bonferroni Correction  The classic approach to FWER correction is the Bonferroni Correction

Bonferroni Correction  Ironically, derived by Miller rather than Bonferroni

Bonferroni Correction  Ironically, derived by Miller rather than Bonferroni  Also ironically, there appear to be no pictures of Miller on the internet

Bonferroni Correction  A classic example of Stigler’s Law of Eponomy  “No scientific discovery is named after its original discoverer”

Bonferroni Correction  A classic example of Stigler’s Law of Eponomy  “No scientific discovery is named after its original discoverer”  Stigler’s Law of Eponomy was proposed by Robert Merton

Bonferroni Correction  If you are conducting n different statistical tests on the same data set  Adjust your significance criterion  to be   / n  E.g. For 4 statistical tests, use statistical significance criterion of rather than 0.05

Bonferroni Correction: Example  Five tests  p=0.04, p=0.12, p=0.18, p=0.33, p=0.55  Five corrections  All p compared to  = 0.01  None significant anymore  p=0.04 seen as being due to chance

Bonferroni Correction: Example  Five tests  p=0.04, p=0.12, p=0.18, p=0.33, p=0.55  Five corrections  All p compared to  = 0.01  None significant anymore  p=0.04 seen as being due to chance  Does this seem right?

Bonferroni Correction: Example  Five tests  p=0.001, p=0.011, p=0.02, p=0.03, p=0.04  Five corrections  All p compared to  = 0.01  Only p=0.001 still significant

Bonferroni Correction: Example  Five tests  p=0.001, p=0.011, p=0.02, p=0.03, p=0.04  Five corrections  All p compared to  = 0.01  Only p=0.001 still significant  Does this seem right?

Quiz  If you run 100 tests, which of the following are statistically significant? A) 0.05 B) 0.01 C) D) E) All of the Above F) None of the Above

Bonferroni Correction Advantages You can be “certain” that an effect is real if it makes it through this correction Does not assume tests are independent In our “100 correlations with the same variable” case, they aren’t! Disadvantages Massively over-conservative Throws out everything if you run a lot of correlations

Often attacked these days  Arguments for rejecting the sequential Bonferroni in ecological studies. MD Moran - Oikos, JSTOR Arguments for rejecting the sequential Bonferroni in ecological studies  Beyond Bonferroni: less conservative analyses for conservation genetics. SR Narum - Conservation Genetics, 2006 – Springer Beyond Bonferroni: less conservative analyses for conservation genetics  What's wrong with Bonferroni adjustments. TV Perneger - Bmj, bmj.com What's wrong with Bonferroni adjustments  p Value fetishism and use of the Bonferroni adjustment. JF Morgan - Evidence Based Mental Health, 2007 p Value fetishism and use of the Bonferroni adjustment

There are FWER corrections that are a little less conservative…  Holm Correction/Holm’s Step-Down (Toothaker, 1991)  Tukey’s HSD (Honestly Significant Difference)  Sidak Correction  Still generally very conservative  Lead to discarding results that probably should not be discarded

FDR Correction  (Benjamini & Hochberg, 1991)

FDR Correction  Different paradigm, arguably a better match to the original conception of statistical significance

Statistical significance  p<0.05  A test is treated as rejecting the null hypothesis if there is a probability of under 5% that the results could have occurred if there were only random events going on  This paradigm accepts from the beginning that we will accept junk (e.g. Type I error) 5% of the time

FWER Correction  p<0.05  Each test is treated as rejecting the null hypothesis if there is a probability of under 5% divided by N that the results could have occurred if there were only random events going on  This paradigm accepts junk far less than 5% of the time

FDR Correction  p<0.05  Across tests, we will attempt to accept junk exactly 5% of the time  Same degree of conservatism as the original conception of statistical significance

FDR Procedure (Benjamini & Hochberg, 1991)  Order your n tests from most significant (lowest p) to least significant (highest p) Test your first test according to significance criterion  / n Test your second test according to significance criterion  / n Test your third test according to significance criterion  / n Quit as soon as a test is not significant

FDR Correction: Example  Five tests  p=0.001, p=0.011, p=0.02, p=0.03, p=0.04

FDR Correction: Example  Five tests  p=0.001, p=0.011, p=0.02, p=0.03, p=0.04  First correction  p = compared to  = 0.01  Still significant!

FDR Correction: Example  Five tests  p=0.001, p=0.011, p=0.02, p=0.03, p=0.04  Second correction  p = compared to  = 0.02  Still significant!

FDR Correction: Example  Five tests  p=0.001, p=0.011, p=0.02, p=0.03, p=0.04  Third correction  p = 0.02 compared to  = 0.03  Still significant!

FDR Correction: Example  Five tests  p=0.001, p=0.011, p=0.02, p=0.03, p=0.04  Fourth correction  p = 0.03 compared to  = 0.04  Still significant!

FDR Correction: Example  Five tests  p=0.001, p=0.011, p=0.02, p=0.03, p=0.04  Fifth correction  p = 0.04 compared to  = 0.05  Still significant!

FDR Correction: Example  Five tests  p=0.04, p=0.12, p=0.18, p=0.33, p=0.55

FDR Correction: Example  Five tests  p=0.04, p=0.12, p=0.18, p=0.33, p=0.55  First correction  p = 0.04 compared to  = 0.01  Not significant; stop

Conservatism  Much less conservative than Bonferroni Correction  Much more conservative than just accepting p<0.05, no matter how many tests are run

q value extension in FDR (Storey, 2002)

 p = probability that the results could have occurred if there were only random events going on  q = probability that the current test is a false discovery, given the post-hoc adjustment

q value extension in FDR (Storey, 2002)  q can actually be lower than p  In the relatively unusual case where there are many statistically significant results

Closing thought  Correlation mining can be a powerful way to see what factors are mathematically associated with each other  Important to get the right level of conservatism

Next lecture  Causal mining