Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 1 False discovery rate: setting the probability.

Slides:



Advertisements
Similar presentations
Inference in the Simple Regression Model
Advertisements

Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Hypothesis Testing A hypothesis is a claim or statement about a property of a population (in our case, about the mean or a proportion of the population)
NASSP Masters 5003F - Computational Astronomy Lecture 5: source detection. Test the null hypothesis (NH). –The NH says: let’s suppose there is no.
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved.
1 1 Slide Chapter 9 Hypothesis Tests Developing Null and Alternative Hypotheses Developing Null and Alternative Hypotheses Type I and Type II Errors Type.
1 1 Slide © 2009 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Hypothesis Testing Developing Null and Alternative Hypotheses Developing Null and Alternative Hypotheses Type I and Type II Errors Type I and Type II Errors.
1 1 Slide STATISTICS FOR BUSINESS AND ECONOMICS Seventh Edition AndersonSweeneyWilliams Slides Prepared by John Loucks © 1999 ITP/South-Western College.
1 1 Slide MA4704Gerry Golding Developing Null and Alternative Hypotheses Hypothesis testing can be used to determine whether Hypothesis testing can be.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 9 Hypothesis Testing Developing Null and Alternative Hypotheses Developing Null and.
07/01/15 MfD 2014 Xin You Tai & Misun Kim
G.A.Prodi - INFN and Università di Trento, Italy International Gravitational Event Collaboration igec.lnl.infn.it ALLEGRO group:ALLEGRO (LSU)
Differentially expressed genes
Pengujian Hipotesis Nilai Tengah Pertemuan 19 Matakuliah: I0134/Metode Statistika Tahun: 2007.
Cbio course, spring 2005, Hebrew University (Alignment) Score Statistics.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Pearson Prentice-Hall, Inc.Chap 9-1 Statistics for Managers Using Microsoft® Excel 5th Edition.
Hypothesis Tests for Means The context “Statistical significance” Hypothesis tests and confidence intervals The steps Hypothesis Test statistic Distribution.
Ch. 9 Fundamental of Hypothesis Testing
Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 8 Tests of Hypotheses Based on a Single Sample.
Statistical hypothesis testing – Inferential statistics I.
Example 10.1 Experimenting with a New Pizza Style at the Pepperoni Pizza Restaurant Concepts in Hypothesis Testing.
False Discovery Rate (FDR) = proportion of false positive results out of all positive results (positive result = statistically significant result) Ladislav.
AM Recitation 2/10/11.
Multiple testing correction
1 1 Slide © 2005 Thomson/South-Western Chapter 9, Part A Hypothesis Tests Developing Null and Alternative Hypotheses Developing Null and Alternative Hypotheses.
© 2003 Prentice-Hall, Inc.Chap 9-1 Fundamentals of Hypothesis Testing: One-Sample Tests IE 340/440 PROCESS IMPROVEMENT THROUGH PLANNED EXPERIMENTATION.
Section 9.1 Introduction to Statistical Tests 9.1 / 1 Hypothesis testing is used to make decisions concerning the value of a parameter.
Inference in practice BPS chapter 16 © 2006 W.H. Freeman and Company.
Intermediate Applied Statistics STAT 460
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 9: Testing a Claim Section 9.3a Tests About a Population Mean.
+ Chapter 9 Summary. + Section 9.1 Significance Tests: The Basics After this section, you should be able to… STATE correct hypotheses for a significance.
Essential Statistics in Biology: Getting the Numbers Right
CHAPTER 16: Inference in Practice. Chapter 16 Concepts 2  Conditions for Inference in Practice  Cautions About Confidence Intervals  Cautions About.
Statistical problems in network data analysis: burst searches by narrowband detectors L.Baggio and G.A.Prodi ICRR TokyoUniv.Trento and INFN IGEC time coincidence.
Example 10.1 Experimenting with a New Pizza Style at the Pepperoni Pizza Restaurant Concepts in Hypothesis Testing.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Learning Objectives In this chapter you will learn about the t-test and its distribution t-test for related samples t-test for independent samples hypothesis.
1 1 Slide IS 310 – Business Statistics IS 310 Business Statistics CSU Long Beach.
Computational Biology Jianfeng Feng Warwick University.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
Silvia Poggi - GW burst detection strategy in non-homogeneus networks Detection strategies for bursts in networks of non-homogeneus gravitational waves.
Testing of Hypothesis Fundamentals of Hypothesis.
Back to basics – Probability, Conditional Probability and Independence Probability of an outcome in an experiment is the proportion of times that.
Frequentistic approaches in physics: Fisher, Neyman-Pearson and beyond Alessandro Palma Dottorato in Fisica XXII ciclo Corso di Probabilità e Incertezza.
Introduction to sample size and power calculations Afshin Ostovar Bushehr University of Medical Sciences.
GravStat 2005 and other Topics in Statistics for Gravitational Waves Detection Lucio Baggio Seminar to ICRR’s GW group.
Multiple Testing Matthew Kowgier. Multiple Testing In statistics, the multiple comparisons/testing problem occurs when one considers a set of statistical.
ILIAS WP1 – Cascina IGEC – First experience using the data of 5 bar detectors: ALLEGRO, AURIGA, EXPLORER NAUTILUS and NIOBE. – 1460.
Information Technology and Decision Making Information Technology and Decision Making Example 10.1 Experimenting with a New Pizza Style at the Pepperoni.
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Statistics for Decision Making Basic Inference QM Fall 2003 Instructor: John Seydel, Ph.D.
Simple examples of the Bayesian approach For proportions and means.
© 2004 Prentice-Hall, Inc.Chap 9-1 Basic Business Statistics (9 th Edition) Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests.
© Copyright McGraw-Hill 2004
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Lecture 8 Source detection NASSP Masters 5003S - Computational Astronomy
1 Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
BIOSTATISTICS Hypotheses testing and parameter estimation.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 6 –Multiple hypothesis testing Marshall University Genomics.
A Quantitative Overview to Gene Expression Profiling in Animal Genetics Armidale Animal Breeding Summer Course, UNE, Feb Analysis of (cDNA) Microarray.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
GWDAW11 – Potsdam Results by the IGEC2 collaboration on 2005 data Gabriele Vedovato for the IGEC2 collaboration.
Hypothesis Tests u Structure of hypothesis tests 1. choose the appropriate test »based on: data characteristics, study objectives »parametric or nonparametric.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Critical Appraisal Course for Emergency Medicine Trainees Module 2 Statistics.
Estimating the False Discovery Rate in Genome-wide Studies BMI/CS 576 Colin Dewey Fall 2008.
Slides by JOHN LOUCKS St. Edward’s University.
Presentation transcript:

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 1 False discovery rate: setting the probability of false claim of detection Lucio Baggio Italy, INFN and University of Trento

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 2  ” We made many different (almost independent) background and coincidence counts using different values for the target signal amplitude (thresholds)” Where does this FDR comes from?  I met prof. James T. Linnemann (MSU) at PHYSTAT2003, I explained him the problem we had in assessing frequentist probability for IGEC results…  ” At last, one of the 90% confidence intervals was not including the null hypothesis… but when one accounts for many trials, it is possible to compute that with 30% probability we had a chance that at least one of the tests falsely rejected the null hypothesis. So, no true discovery, after all.”  ” Perhaps next time we should use 99.99% confidence intervals, in order to make the probability of false claim still low after tens of trials…But I’m afraid that signals cannot any more emerge with such stringent a requirement.”  He pointed out that maybe what I’m looking for are false discovery rate methods… Thanks, Jim!!

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 3 Why FDR? When should I care of multiple test procedures?. All sky surveys: many source directions and polarizations are tried Template banks Wide-open-eyes searches: many analysis pipelines are tried altogether, with different amplitude thresholds, signal durations, and so on Periodic updates of results: every new science run is a chance for a “discovery”. “Maybe next one is the good one”. Many graphical representations or aggregations of the data: “If I change the binning, maybe the signal shows up better…

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 4 Preliminary (1) : hypothesis testing False discoveries (false positives) Detected signals (true positives) Reported signal candidates inefficiency Null Retained (can’t reject) Reject= Reject Null = Accept Accept Alternative Total Null (H o ) True Background (noise) U B Type I Error α = ε b momo Alternative True signal Type II Error β = 1- ε s T Sm1m1 m-R R = S+B m

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 5 Preliminary (2): p-level Assume you have a model for the noise that affects the measure x. However, for our purposes it is sufficient assuming that the signal can be distinguished from the noise, i.e. dP/dp  1. Typically, the measured values of p are biased toward 0. signal You derive a test statistics t(x) from x. F(t) is the distribution of t when x is sampled from noise only (off-signal). The p-level associated with t(x) is the value of the distribution of t in t(x): p = F(t) = P(t>t(x)) Example:  2 test  p is the “one-tail”  2 probability associated with n counts (assuming d degrees of freedom) Usually, the alternative hypothesis is not known. p-level 1 background pdf The distribution of p is always linearly raising in case of agreement of the noise with the model P(p)=p  dP/dp = 1

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 6 Usual multiple testing procedures For each hypothesis test, the condition {p<   reject null} leads to false positives with a probability  In case of multiple tests (need not to be the same test statistics, nor the same tested null hypothesis), let p={p 1, p 2, … p m } be the set of p-levels. m is the trial factor. We select “discoveries” using a threshold T(p): {p j <T(p)  reject null}. Uncorrected testing: T(p)=  –The probability that at least one rejection is wrong is P(B>0) = 1 – (1-  ) m ~ m  hence false discovery is guaranteed for m large enough Fixed total 1 st type errors (Bonferroni): T(p)=  /m –Controls familywise error rate in the most stringent manner: P(B>0) =  –This makes mistakes rare… –… but in the end efficiency (2 nd type errors) becomes negligible!!

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 7 p S pdf m0m0 Let us make a simple case when signals are easily separable (e.g. high SNR) Controlling false discovery fraction We desire to control (=bound) the ratio of false discoveries over the total number of claims: B/R = B/(B+S)  q. The level T(p) is then chosen accordingly. B m q p B S cumulative counts R

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 8 Benjamini & Hochberg FDR control procedure Among the procedures that accomplish this task, one simple recipe was proposed by Benjamini & Hochberg (JRSS-B (1995) 57: ) choose your desired FDR q (don’t ask too much!); define c(m)=1 if p-values are independent or positively correlated; otherwise c(m)=Sum j (1/j) compute p-values {p 1, p 2, … p m } for a set of tests, and sort them in creasing order; p m determine the threshold T(p)= p k by finding the index k such that p j k; reject H 0 q/c(m)

Lucio Baggio - Lucio Baggio - False discovery rate: setting the probability of false claim of detection 9 Summary In case of multiple tests one wants to control the false claim probability, but it is advisable to mitigate the strict requirement that we want NO false claim, which could end up in burying the signals also. Controlling FDR seems to be a wise suggestion. This talk was based mainly on Miller et. al. ApJ 122: Dec and Jim Linnemann’s talk However, there is a fairly wide literature aboute this, if one looks for references in biology, imaging, HEP, and recently also astrophysics, at last!