Sequence comparison: Multiple testing correction

Slides:



Advertisements
Similar presentations
Chapter 7 Hypothesis Testing
Advertisements

Sequence comparison: Significance of similarity scores Genome 559: Introduction to Statistical and Computational Genomics Prof. James H. Thomas.
Sequence comparison: Significance of similarity scores Genome 559: Introduction to Statistical and Computational Genomics Prof. William Stafford Noble.
1 1 Slide STATISTICS FOR BUSINESS AND ECONOMICS Seventh Edition AndersonSweeneyWilliams Slides Prepared by John Loucks © 1999 ITP/South-Western College.
Chapter 9 Hypothesis Testing
BLAST, PSI-BLAST and position- specific scoring matrices Prof. William Stafford Noble Department of Genome Sciences Department of Computer Science and.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Overview of Lecture Independent and Dependent Variables Between and Within Designs.
On Comparing Classifiers: Pitfalls to Avoid and Recommended Approach Published by Steven L. Salzberg Presented by Prakash Tilwani MACS 598 April 25 th.
Sequence comparison: Significance of similarity scores Genome 559: Introduction to Statistical and Computational Genomics Prof. James H. Thomas.
Whole genome alignments Genome 559: Introduction to Statistical and Computational Genomics Prof. James H. Thomas
AM Recitation 2/10/11.
Multiple testing correction
Testing Hypotheses I Lesson 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics n Inferential Statistics.
Copyright © 2010, 2007, 2004 Pearson Education, Inc Lecture Slides Elementary Statistics Eleventh Edition and the Triola Statistics Series by.
Overview Definition Hypothesis
Sections 8-1 and 8-2 Review and Preview and Basics of Hypothesis Testing.
Overview Basics of Hypothesis Testing
Comp. Genomics Recitation 3 The statistics of database searching.
Multiple Testing Matthew Kowgier. Multiple Testing In statistics, the multiple comparisons/testing problem occurs when one considers a set of statistical.
Statistical significance of alignment scores Prof. William Stafford Noble Department of Genome Sciences Department of Computer Science and Engineering.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Chapter 8 Parameter Estimates and Hypothesis Testing.
INTRODUCTION TO HYPOTHESIS TESTING From R. B. McCall, Fundamental Statistics for Behavioral Sciences, 5th edition, Harcourt Brace Jovanovich Publishers,
Comp. Genomics Recitation 3 (week 4) 26/3/2009 Multiple Hypothesis Testing+Suffix Trees Based in part on slides by William Stafford Noble.
C HAPTER 2  Hypothesis Testing -Test for one means - Test for two means -Test for one and two proportions.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Comp. Genomics Recitation 2 (week 3) 19/3/09. Outline Finding repeats Branch & Bound for MSA Multiple hypotheses testing.
+ Homework 9.1:1-8, 21 & 22 Reading Guide 9.2 Section 9.1 Significance Tests: The Basics.
Lecture Slides Elementary Statistics Twelfth Edition
Chapter 8 Introducing Inferential Statistics.
Step 1: Specify a null hypothesis
Chapter 9: Testing a Claim
CHAPTER 9 Testing a Claim
Chapter 9: Testing a Claim
Review and Preview and Basics of Hypothesis Testing
Tests of Significance The reasoning of significance tests
CHAPTER 9 Testing a Claim
Understanding Results
Chapter 9: Testing a Claim
Chapters 20, 21 Hypothesis Testing-- Determining if a Result is Different from Expected.
© LOUIS COHEN, LAWRENCE MANION AND KEITH MORRISON
Elementary Statistics
Hypothesis Testing: Hypotheses
Transcription factor binding motifs
When we free ourselves of desire,
Sequence comparison: Significance of similarity scores
Motif p-values GENOME 559: Introduction to Statistical and Computational Genomics Prof. William Stafford Noble.
Sequence comparison: Multiple testing correction
I. Statistical Tests: Why do we use them? What do they involve?
INTRODUCTION TO HYPOTHESIS TESTING
Hypothesis Testing.
Chapter 9: Testing a Claim
Intro to Confidence Intervals Introduction to Inference
Sequence comparison: Significance of similarity scores
Chapter 9: Testing a Claim
Chapter 9: Testing a Claim
False discovery rate estimation
CHAPTER 9 Testing a Claim
CHAPTER 9 Testing a Claim
Transcription factor binding motifs
Chapter 9: Testing a Claim
Chapter 9: Testing a Claim
CHAPTER 16: Inference in Practice
Chapter 9: Testing a Claim
Chapter 9: Testing a Claim
Testing Hypotheses I Lesson 9.
Chapter 9: Significance Testing
Chapter 9: Testing a Claim
Unit 5: Hypothesis Testing
Chapter 9: Testing a Claim
Presentation transcript:

Sequence comparison: Multiple testing correction Genome 559: Introduction to Statistical and Computational Genomics Prof. William Stafford Noble

One-minute response Difficulty connecting conceptual part of course to programming. This will improve now that you know more programming constructs. Allocate more time to practice problems. Remind how can you exit an if test for a command line argument? sys.exit() will stop the program It might be helpful to have the previous question’s answer on the screen at the same time as the next question. I don’t know if I have space. More sample problems with solutions would be helpful. There are more sample problems linked from the last page of the slides. It would be helpful to do a brief recap at the beginning of class with Most Important Functions we’ve covered so far. Explanation of string formatting was helpful. Nice pace. Python is really fun! I liked the examples for the if statement problems. Love going through Python in class.

The show so far … Most statistical tests compare observed data to the expected result according to the null hypothesis. Sequence similarity scores follow an extreme value distribution, which is characterized by a larger tail. The p-value associated with a score is the area under the curve to the right of that score. The p-value for score x for an extreme value distribution with parameters μ and λ is

What p-value is significant?

What p-value is significant? The most common thresholds are 0.01 and 0.05. A threshold of 0.05 means you are 95% sure that the result is significant. Is 95% enough? It depends upon the cost associated with making a mistake. Examples of costs: Doing expensive wet lab validation. Making clinical treatment decisions. Misleading the scientific community. Most sequence analysis uses more stringent thresholds because the p- values are not very accurate.

Two types of errors In statistical testing, there are two types of errors: False positive (aka “Type I error”): incorrectly rejecting the null hypothesis. I.e., saying that two sequences are homologous when they are not. False negative (aka “Type II error”): incorrectly accepting the null hypothesis. I.e., saying that two sequences are not homologous when they are. We use p-values to control the false positive rate.

Multiple testing Say that you perform a statistical test with a 0.05 threshold, but you repeat the test on twenty different observations. Assume that all of the observations are explainable by the null hypothesis. What is the chance that at least one of the observations will receive a p-value less than 0.05?

Multiple testing Say that you perform a statistical test with a 0.05 threshold, but you repeat the test on twenty different observations. Assuming that all of the observations are explainable by the null hypothesis, what is the chance that at least one of the observations will receive a p-value less than 0.05? Pr(making a mistake) = 0.05 Pr(not making a mistake) = 0.95 Pr(not making any mistake) = 0.9520 = 0.358 Pr(making at least one mistake) = 1 - 0.358 = 0.642 There is a 64.2% chance of making at least one mistake.

Family-wise error rate For a single test, the Type I error rate is the probability of incorrectly rejecting the null hypothesis. For multiple tests, the family-wise error rate is the probability of incorrectly rejecting at least one null hypothesis.

Bonferroni correction Divide the desired p-value threshold by the number of tests performed. For the previous example, 0.05 / 20 = 0.0025. Pr(making a mistake) = 0.0025 Pr(not making a mistake) = 0.9975 Pr(not making any mistake) = 0.997520 = 0.9512 Pr(making at least one mistake) = 1 - 0.9512 = 0.0488

Proof: Bonferroni adjustment controls the family-wise error rate Note: Bonferroni adjustment does not require that the tests be independent. Boole’s inequality Definition of p-value m = number of hypotheses m0 = number of null hypotheses ⍺ = desired control pi = ith p-value Definition of m and m0

Database searching Say that you search the non-redundant protein database at NCBI, containing roughly one million sequences. What p-value threshold should you use?

Database searching Say that you search the non-redundant protein database at NCBI, containing roughly one million sequences. What p-value threshold should you use? Say that you want to use a conservative p-value of 0.001. Recall that you would observe such a p-value by chance approximately every 1000 times in a random database. A Bonferroni correction would suggest using a p-value threshold of 0.001 / 1,000,000 = 0.000000001 = 10-9.

E-values The E-value is the expected number of times that the given score would appear in a random database of the given size. One simple way to compute the E-value is to multiply the p-value times the size of the database. Thus, for a p-value of 0.001 and a database of 1,000,000 sequences, the corresponding E-value is 0.001 × 1,000,000 = 1,000. BLAST actually calculates E-values in a more complex way.

Summary Confidence thresholds of 0.01 and 0.05 are arbitrary. In practice, the threshold should be selected based on the costs associated with false positive and false negative conclusions. The Bonferroni correction divides the desired threshold by the number of tests performed. The E-value is equivalent to Bonferroni adjustment, but multiplies the p-value by the number of tests performed.