1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24.

Slides:



Advertisements
Similar presentations
October 1999 Statistical Methods for Computer Science Marie desJardins CMSC 601 April 9, 2012 Material adapted.
Advertisements

Practical Statistics for LHC Physicists Bayesian Inference Harrison B. Prosper Florida State University CERN Academic Training Lectures 9 April, 2015.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
1 LIMITS Why limits? Methods for upper limits Desirable properties Dealing with systematics Feldman-Cousins Recommendations.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #22.
Visual Recognition Tutorial
Maximum likelihood (ML) and likelihood ratio (LR) test
Elementary hypothesis testing
Estimation Procedures Point Estimation Confidence Interval Estimation.
Slide 1 Statistics for HEP Roger Barlow Manchester University Lecture 3: Estimation.
Elementary hypothesis testing
Elementary hypothesis testing Purpose of hypothesis testing Type of hypotheses Type of errors Critical regions Significant levels Hypothesis vs intervals.
Statistical inference form observational data Parameter estimation: Method of moments Use the data you have to calculate first and second moment To fit.
Credibility of confidence intervals Dean Karlen / Carleton University Advanced Statistical Techniques in Particle Physics Durham, March 2002.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
Copyright (c) Bani Mallick1 Lecture 4 Stat 651. Copyright (c) Bani Mallick2 Topics in Lecture #4 Probability The bell-shaped (normal) curve Normal probability.
Hypothesis Tests for Means The context “Statistical significance” Hypothesis tests and confidence intervals The steps Hypothesis Test statistic Distribution.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Rao-Cramer-Frechet (RCF) bound of minimum variance (w/o proof) Variance of an estimator of single parameter is limited as: is called “efficient” when the.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Maximum likelihood (ML)
Sample Size Determination Ziad Taib March 7, 2014.
Statistical Analysis of Systematic Errors and Small Signals Reinhard Schwienhorst University of Minnesota 10/26/99.
Interval Estimates A point estimate gives a plausible single number estimate for a parameter. We may also be interested in a range of plausible values.
Statistical aspects of Higgs analyses W. Verkerke (NIKHEF)
Chapter 8 Hypothesis testing 1. ▪Along with estimation, hypothesis testing is one of the major fields of statistical inference ▪In estimation, we: –don’t.
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
Statistical problems in network data analysis: burst searches by narrowband detectors L.Baggio and G.A.Prodi ICRR TokyoUniv.Trento and INFN IGEC time coincidence.
1 Probability and Statistics  What is probability?  What is statistics?
Statistical Decision Theory
Overview course in Statistics (usually given in 26h, but now in 2h)  introduction of basic concepts of probability  concepts of parameter estimation.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
SPC for Real-World Processes William A. Levinson, P.E. Intersil Corporation Mountaintop, PA.
PARAMETRIC STATISTICAL INFERENCE
Essential Statistics Chapter 131 Introduction to Inference.
CHAPTER 14 Introduction to Inference BPS - 5TH ED.CHAPTER 14 1.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #23.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Statistics In HEP Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Confidence Intervals First ICFA Instrumentation School/Workshop At Morelia, Mexico, November 18-29, 2002 Harrison B. Prosper Florida State University.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #25.
Business Statistics for Managerial Decision Farideh Dehkordi-Vakil.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
DIJET STATUS Kazim Gumus 30 Aug Our signal is spread over many bins, and the background varies widely over the bins. If we were to simply sum up.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
2005 Unbinned Point Source Analysis Update Jim Braun IceCube Fall 2006 Collaboration Meeting.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Ivo van Vulpen Why does RooFit put asymmetric errors on data points ? 10 slides on a ‘too easy’ topic that I hope confuse, but do not irritate you
Probability and Statistics for Particle Physics Javier Magnin CBPF – Brazilian Center for Research in Physics Rio de Janeiro - Brazil.
Status of the Higgs to tau tau
arXiv:physics/ v3 [physics.data-an]
Chapter 5 STATISTICAL INFERENCE: ESTIMATION AND HYPOTHESES TESTING
Lecture 4 1 Probability (90 min.)
Modelling data and curve fitting
STA 291 Spring 2008 Lecture 5 Dustin Lueker.
STA 291 Spring 2008 Lecture 5 Dustin Lueker.
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Parametric Methods Berlin Chen, 2005 References:
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Statistics for HEP Roger Barlow Manchester University
Computing and Statistical Data Analysis / Stat 10
Lecture 8 Hypothesis Testing
Presentation transcript:

1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24

The Likelihood Function 2

Parameter Extraction The likelihood function discussed earlier is the basis for all further interpretations While the absolute value of L is not very telling on its own (need to calculate the p-value), relative changes in L are If you have a choice of parameters for the hypothesis, you can check how L is changing as a function of that parameter (fix “data” to what you saw in the data) Typical example: Likelihood function versus normalization of the signal S Some values lead to better agreement with the data than others 3

Bayesian Interpretation While the easiest way to pass all the information about your measurement is via the Likelihood function, one usually want a simple way to quantify the results Conventionally, people report 68% C.L. intervals for the parameters being Many ways to choose 68%, need a convention One example is highest posterior density intervals “Probability outside is less than on the inside” C.L. stands for: Confidence Level when using frequentist approach Credibility Level is Bayesian 4

Frequentist Limits Suppose that the true value of  is  0 :  0 lies between  1 (x) and  2 2(x) if and only if x lies between x 1 (  0 ) and x 2 (  0 ) The two events have the same probability, Since it is true for any value  0 This is Neyman procedure. Constructed to ensure the desired coverage Basically coverage is the probability that the true parameter is within the interval you report If you say that at 68% C.L. z=1+/-0.1, it better be that in 68% of experiments the true value of z is within the interval you report 5

Setting Frequentist Limits 6 Once you constructed the Likelihood, you can do pseudoexperiments Find what the limits should be to ensure desired coverage for any outcome x Report the limits for the x you saw in your data You still need how to define the borders of the confidence belt Also a convention, one option is require a/2 on each side Or can follow a different procedure (e.g. Feldman-Cousins)

Multi-Parameter Likelihoods Not too difficult to extend what we did onto a multi-dimensional space: Your reported intervals will become ellipses: The conventional way to report the numeric results “on paper” is to “overcover” by reporting rectangular box limits But you can also report the correlation matrix or make a 2D-plot like above if prudent, happens often 7

One Sided Limits It is typical in HEP to look for things that are at the edge of your sensitivity You frequently can’t “see” the signal, but you still want to say that the cross-section for that new process can’t be larger than X Also very useful information for theorists and model builders as your results can rule out a whole bunch of models or parameter space of the models Can do it for either Bayesian or Frequentist methods Most of the time fairly straightforward – either construct a one-sided intervals with known coverage or calculate the integral from 0 to x in Bayesian case 8

Bayesian vs Frequentist Practical differences usually only really relevant when the measured parameter is near a physical boundary Example: you looking for signal by measuring the number of some type of events during one year. You estimated that the background processes should yield B=100 events on average, so any “excess” in data (D=S+B) would indicate a potential signal If you observe 300 events, you clearly see signal and can easily do the interpretation If you see 100 events, you are still okay But what if you see 30 events or even zero events? What would be the upper 95% C.L. limit for signal in each case? Does that make sense? 9

Bayesian vs Frequentist 10

Practical Notes You usually will calculate –log(L) instead of L Assuming you are doing a measurement, the minimum of –log(L) is the maximum of L, so that gives the most likely parameter value Changing – log(L) by +/-1/2 (think of taking a log of a gaussian distribution – it will give you (x-x 0 ) 2 /2   so x shifted by sigma gives ½) gives you 1 sigma deviation in the parameter (68% C.L.) With the MLS method, you vary it by 1 instead: MINUIT is the most used minimization package in HEP (it is part of ROOT), easy to use in simple cases, some experience required for more complex ones 11

Some good to remember numbers Upper 90% and 95% limit on the rate of a Poisson process in the absence of background when you observe n events If you observe no events, the limit is 3 events In Bayesian case, this would also be true for any expected B rate 12

Un-Binned Likelihood Binned likelihood is easy to interpret and gives consistent and predictable outcomes But there is a choice of bin size and strictly speaking you are loosing information by clamping things together Can you avoid that? 13

Next Time Take as a use case Higgs results from CMS and walk through various plots Upper limits, various definitions of tests, central values, confidence intervals etc 14