Statistics for Particle Physics: Limits Roger Barlow Karlsruhe: 12 October 2009.

Slides:



Advertisements
Similar presentations
27 th March CERN Higgs searches: CL s W. J. Murray RAL.
Advertisements

Practical Statistics for LHC Physicists Bayesian Inference Harrison B. Prosper Florida State University CERN Academic Training Lectures 9 April, 2015.
1 LIMITS Why limits? Methods for upper limits Desirable properties Dealing with systematics Feldman-Cousins Recommendations.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #22.
An Answer and a Question Limits: Combining 2 results Significance: Does  2 give  2 ? Roger Barlow BIRS meeting July 2006.
Chapter 19 Confidence Intervals for Proportions.
Setting Limits in the Presence of Nuisance Parameters Wolfgang A Rolke Angel M López Jan Conrad, CERN.
Statistics In HEP 2 Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Statistics for Particle Physics: Intervals Roger Barlow Karlsruhe: 12 October 2009.
Slide 1 Statistics for HEP Roger Barlow Manchester University Lecture 1: Probability.
Point estimation, interval estimation
Results from Small Numbers Roger Barlow Manchester 26 th September 2007.
Slide 1 Statistics for HEP Roger Barlow Manchester University Lecture 3: Estimation.
Probability Frequentist versus Bayesian and why it matters Roger Barlow Manchester University 11 th February 2008.
Credibility of confidence intervals Dean Karlen / Carleton University Advanced Statistical Techniques in Particle Physics Durham, March 2002.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
Hypothesis Tests for Means The context “Statistical significance” Hypothesis tests and confidence intervals The steps Hypothesis Test statistic Distribution.
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 41 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability.
Inferences About Process Quality
Sensitivity of searches for new signals and its optimization
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Role and Place of Statistical Data Analysis and very simple applications Simplified diagram of a scientific research When you know the system: Estimation.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Everything You Always Wanted To Know About Limits* Roger Barlow Manchester University YETI06 *But were afraid to ask.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination using shapes ATLAS Statistics Meeting CERN, 19 December, 2007 Glen Cowan.
Business Statistics: Communicating with Numbers
Statistical Analysis of Systematic Errors and Small Signals Reinhard Schwienhorst University of Minnesota 10/26/99.
Statistical aspects of Higgs analyses W. Verkerke (NIKHEF)
Hypothesis Testing. Distribution of Estimator To see the impact of the sample on estimates, try different samples Plot histogram of answers –Is it “normal”
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
Statistical problems in network data analysis: burst searches by narrowband detectors L.Baggio and G.A.Prodi ICRR TokyoUniv.Trento and INFN IGEC time coincidence.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
1 Statistical Distribution Fitting Dr. Jason Merrick.
Practical Statistics for Particle Physicists Lecture 3 Harrison B. Prosper Florida State University European School of High-Energy Physics Anjou, France.
Confidence Intervals First ICFA Instrumentation School/Workshop At Morelia, Mexico, November 18-29, 2002 Harrison B. Prosper Florida State University.
Issues concerning the interpretation of statistical significance tests.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #25.
G. Cowan, RHUL Physics Discussion on significance page 1 Discussion on significance ATLAS Statistics Forum CERN/Phone, 2 December, 2009 Glen Cowan Physics.
Experience from Searches at the Tevatron Harrison B. Prosper Florida State University 18 January, 2011 PHYSTAT 2011 CERN.
Practical Statistics for Particle Physicists Lecture 2 Harrison B. Prosper Florida State University European School of High-Energy Physics Parádfürdő,
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Easy Limit Statistics Andreas Hoecker CAT Physics, Mar 25, 2011.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
A bin-free Extended Maximum Likelihood Fit + Feldman-Cousins error analysis Peter Litchfield  A bin free Extended Maximum Likelihood method of fitting.
DIJET STATUS Kazim Gumus 30 Aug Our signal is spread over many bins, and the background varies widely over the bins. If we were to simply sum up.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Maximum likelihood estimators Example: Random data X i drawn from a Poisson distribution with unknown  We want to determine  For any assumed value of.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
G. Cowan RHUL Physics Statistical Issues for Higgs Search page 1 Statistical Issues for Higgs Search ATLAS Statistics Forum CERN, 16 April, 2007 Glen Cowan.
Confidence Limits and Intervals 3: Various other topics Roger Barlow SLUO Lectures on Statistics August 2006.
Frequentist Confidence Limits and Intervals Roger Barlow SLUO Lectures on Statistics August 2006.
Discussion on significance
arXiv:physics/ v3 [physics.data-an]
BAYES and FREQUENTISM: The Return of an Old Controversy
Confidence Intervals and Limits
Lecture 4 1 Probability (90 min.)
PSY 626: Bayesian Statistics for Psychological Science
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Introduction to Statistics − Day 4
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Statistics for HEP Roger Barlow Manchester University
Everything You Always Wanted To Know About Limits*
Presentation transcript:

Statistics for Particle Physics: Limits Roger Barlow Karlsruhe: 12 October 2009

Limits Neyman constructions Poisson data Poisson Data with low statistics Poisson Data with low statistcs and backgrounds Feldman Cousins methid Bayesian limits Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits2

Intervals and Limits Quoting an interval gives choice – not just in probability/confidence level Usual ‘interval’ is symmetric. E.g. 68% inside, 16% above, 16% below. ‘Limit’ is 1-sided. E.g.Less than 68% Usually upper limits. Occasionally lower limits. Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits3

Generic Problem You are searching for some process (“Cut and count”. More sophisticated methods technically tougher but conceptually similar) Choose cuts (from Monte Carlo, sidebands, etc) Open the box (Blind analysis!) Observe N events Report N ± √N events (Poisson Statistics) Scale up to get Branching Ratio, Production Cross section, or more complicated things like particle mass What happens if N is small (So Poisson ≠ Gaussian) ? What happens if N =0? How do you handle background component? Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits4

Intervals (again) Non-Gaussian distributions need care. Confidence band (Neymann) Construct horizontally. Read vertically. Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits5 x a

Example Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits6 Open the box. See nothing! Cannot quote answer as μ= 0 ± 0 Poisson formula P(n;μ)=e -μ μ n /n! Plausibly μ ≅ 0.1. Run 100x longer and you’ll see something. Possibly μ ≈1 and you’re just unlucky. Find μ such that P(0; μ)=100-68% (or whatever) In particular P(0;3.0) = 5% p-values again: If μ=3, the probability of getting a result this bad (=small) is only 5% So: when you see nothing, say “With 95% confidence, the upper limit on the signal is 3” – and calculate BR, cross section etc from that

Small numbers See an event. Or a few events. Discovery? Perhaps. But maybe you expect a background… Repeat p-value calculation. See n, find μ such that Σ 0 n e -μ μ r /r! is 0.05 (or whatever). Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits7 n90%95%

Where it gets sticky… You expect 3.5 background events. Suppose you open the box and find 15 events. Measurement of signal 11.5±√15. Suppose you find 4 events. Upper limit on total Upper limit on signal 95% Suppose you find 0 events  -0.5 Or 1 event: 4.74  1.24 Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits8

Problem Technically these are correct. 5% of ‘95%CL’ statements are allowed to be wrong. But we don’t want to make crazy statements. Even though purists argue we should to avoid bias. Problem is that the background clearly has a downward fluctuation*. But we have no way of including this information in the formalism * Assuming you calculated it right. Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits9

Frequentist and Bayesian Probability Roger BarlowSlide 10 Similar problems Expected number of events must be non-negative Mass of an object must be non-negative Mass-squared of an object must be non-negative Higgs mass from EW fits must be bigger than LEP2 limit of 114 GeV 3 Solutions Publish a ‘clearly crazy’ result Use Feldman-Cousins technique Switch to Bayesian analysis

Frequentist and Bayesian Probability Roger BarlowSlide 11 Bayesian limits from small number counts P(r,  )=e -   r /r! With uniform prior this gives posterior for  Shown for various small r results Read off intervals... r=6 r=2 r=1 r=0 P(  ) 

Frequentist and Bayesian Probability Roger BarlowSlide 12 Upper limits Upper limit from n events  0  HI exp(-  )  n /n! d  = CL Repeated integration by parts:  0 n exp(-  HI )  HI r /r! = 1-CL Same as frequentist limit This is a coincidence! Lower Limit formula is not the same

Frequentist and Bayesian Probability Roger BarlowSlide 13 Result depends on Prior Example: 90% CL Limit from 0 events Prior flat in  X X = = Prior flat in 

Frequentist and Bayesian Probability Roger BarlowSlide 14 Aside: “Objective” Bayesian statistics Attempt to lay down rule for choice of prior ‘Uniform’ is not enough. Uniform in what? Suggestion (Jeffreys): uniform in a variable for which the expected Fisher information is minimum (statisticians call this a ‘flat prior’). Has not met with general agreement – different measurements of the same quantity have different objective priors

Frequentist and Bayesian Probability Roger BarlowSlide 15  =S+b for Bayesians No problem! Prior for  is uniform for S  b Multiply and normalise as before Posterior Likelihood Prior Read off Confidence Levels by integrating posterior = X

Frequentist and Bayesian Probability Roger BarlowSlide 16 Another Aside: Coverage Given P(x;  ) and an ensemble of possible measurements {x i } and some confidence level algorithm, coverage is how often ‘  LO   HI ’ is true. Isn’t that just the confidence level? Not quite. Discrete observables may mean the confidence belt is not exact – move on side of caution Other ‘nuisance’ parameters may need to be taken account of – again erring on side of caution Coverage depends on . For a frequentist it is never less than the CL (‘undercoverage’). It may be more (‘overcoverage’) – this is to be minimised but not crucial For a Bayesian coverage is technically irrelevant – but in practice useful

Frequentist and Bayesian Probability Roger BarlowSlide 17 Feldman Cousins Method Works by attacking what looks like a different problem... * by Feldman and Cousins, mostly

Frequentist and Bayesian Probability Roger BarlowSlide 18 Feldman Cousins:  =s+b This is called 'flip-flopping' and BAD because is wrecks the whole design of the Confidence Belt Suggested solution: 1) Construct belts at chosen CL as before (for s,N and given b) 2) Find new ranking strategy to determine what's inside and what's outside 1 sided 90% 2 sided 90%

Frequentist and Bayesian Probability Roger BarlowSlide 19 Feldman Cousins: Ranking First idea (almost right) Sum/integrate over range of values with highest probabilities. (advantage: this is the shortest interval) Glitch: Suppose N small. (low fluctuation) P(N;s+b) will be small for any s and never get counted Instead: compare to 'best' probability for this N, at s=N-b or s=0 and rank on that number Such a plot does an automatic ‘flip-flop’ N~b single sided limit (upper bound) for s N>>b 2 sided limits for s N s

Frequentist and Bayesian Probability Roger BarlowSlide 20 How it works Has to be computed for the appropriate value of background b. (Sounds complicated, but there is lots of software around) As N increases, flips from 1- sided to 2-sided limits – but in such a way that the probability of being in the belt is preserved s N Means that sensible 1-sided limits are quoted instead of nonsensical 2- sided limits!

Frequentist and Bayesian Probability Roger BarlowSlide 21 Arguments against using Feldman Cousins Argument 1 It takes control out of hands of physicist. You might want to quote a 2 sided limit for an expected process, an upper limit for something weird Counter argument: This is the virtue of the method. This control invalidates the conventional technique. The physicist can use their discretion over the CL. In rare cases it is permissible to say ”We set a 2 sided limit, but we're not claiming a signal”

Frequentist and Bayesian Probability Roger BarlowSlide 22 Feldman Cousins: Argument 2 Argument 2 If zero events are observed by two experiments, the one with the higher background b will quote the lower limit. This is unfair to hardworking physicists Counterargument An experiment with higher background has to be ‘lucky’ to get zero events. Luckier experiments will always quote better limits. Averaging over luck, lower values of b get lower limits to report. Example: you reward a good student with a lottery ticket which has a 10% chance of winning $10. A moderate student gets a ticket with a 1% chance of winning $20. They both win. Were you unfair? Example: you reward a good student with a lottery ticket which has a 10% chance of winning €10. A moderate student gets a ticket with a 1% chance of winning €20. They both win. Were you unfair?

Frequentist and Bayesian Probability Roger BarlowSlide 23 Including Systematic Errors  =aS+b  is predicted number of events S is (unknown) signal source strength. Probably a cross section or branching ratio or decay rate a is an acceptance/luminosity factor known with some (systematic) error b is the background rate, known with some (systematic) error

Frequentist and Bayesian Probability Roger BarlowSlide 24 1) Full Bayesian Assume priors for S (uniform?) For a (Gaussian?) For b (Poisson or Gaussian?) Write down the posterior P(S,a,b). Integrate over all a,b to get marginalised P(s) Read off desired limits by integration

Frequentist and Bayesian Probability Roger BarlowSlide 25 2) Hybrid Bayesian Assume priors For a (Gaussian?) For b (Poisson or Gaussian?) Integrate over all a,b to get marginalised P(r,S) Read off desired limits by  0 n P(r,S) =1-CL etc Done approximately for small errors (Cousins and Highland). Shows that limits pretty insensitive to  a,  b Numerically for general errors (RB: java applet on SLAC web page). Includes 3 priors (for a) that give slightly different results

Frequentist and Bayesian Probability Roger BarlowSlide 26 And more… Extend Feldman Cousins Profile Likelihood: Use P(S)=P(n,S,a max,b max ) where a max,b max give maximum for this S,n Empirical Bayes And more…

And another things… Using information as well as numbers can use χ 2 or likelihood P(x;s,a)=B(x)+s S(x) Can obtain intervals for s and limits on s Warning: these limits are not valid if S(x)=S(x,a). Even if you reduce the number of degrees of freedom of χ 2 Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits27

Example Suppose S(x) is a Breit-Wigner of known mass and small width. All OK. Minimisation neutralises the relevant bin. Big improvement means its doing something real Suppose S(x) is a Breit-Wigner but mass is allowed to float. Minimisation will neutralise the worst bin. Massive improvement anyway. Large Δ χ 2 but so what? Use a toy Monte Carlo!! Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits28

Frequentist and Bayesian Probability Roger BarlowSlide 29 Summary Straight Frequentist approach is objective and clean but sometimes gives ‘crazy’ results Bayesian approach is valuable but has problems. Check for robustness under choice of prior Feldman-Cousins deserves more widespread adoption Lots of work still going on This will all be needed at the LHC