Statistics for Particle Physics: Intervals Roger Barlow Karlsruhe: 12 October 2009.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Estimation, Variation and Uncertainty Simon French
Probability and Statistics Basic concepts II (from a physicist point of view) Benoit CLEMENT – Université J. Fourier / LPSC
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
Sampling: Final and Initial Sample Size Determination
1 LIMITS Why limits? Methods for upper limits Desirable properties Dealing with systematics Feldman-Cousins Recommendations.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
An Answer and a Question Limits: Combining 2 results Significance: Does  2 give  2 ? Roger Barlow BIRS meeting July 2006.
Experimental Uncertainties: A Practical Guide What you should already know well What you need to know, and use, in this lab More details available in handout.
Slide 1 Statistics for HEP Roger Barlow Manchester University Lecture 1: Probability.
Probability theory Much inspired by the presentation of Kren and Samuelsson.
Developments in Bayesian Priors Roger Barlow Manchester IoP meeting November 16 th 2005.
I The meaning of chance Axiomatization. E Plurbus Unum.
Introduction to Bayesian statistics Three approaches to Probability  Axiomatic Probability by definition and properties  Relative Frequency Repeated.
Thanks to Nir Friedman, HU
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Everything You Always Wanted To Know About Limits* Roger Barlow Manchester University YETI06 *But were afraid to ask.
Understanding sample survey data
Statistics for Marketing & Consumer Research Copyright © Mario Mazzocchi 1 Further advanced methods Chapter 17.
Lecture II-2: Probability Review
Standard error of estimate & Confidence interval.
Statistical Analysis of Systematic Errors and Small Signals Reinhard Schwienhorst University of Minnesota 10/26/99.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
1 Probability and Statistics  What is probability?  What is statistics?
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Confidence Interval Estimation.
Estimation in Sampling!? Chapter 7 – Statistical Problem Solving in Geography.
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
Dr. Gary Blau, Sean HanMonday, Aug 13, 2007 Statistical Design of Experiments SECTION I Probability Theory Review.
Theory of Probability Statistics for Business and Economics.
Bayesian Methods I: Parameter Estimation “A statistician is a person who draws a mathematically precise line from an unwarranted assumption to a foregone.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
Statistical Data Analysis and Simulation Jorge Andre Swieca School Campos do Jordão, January,2003 João R. T. de Mello Neto.
Practical Statistics for Particle Physicists Lecture 3 Harrison B. Prosper Florida State University European School of High-Energy Physics Anjou, France.
Likelihood function and Bayes Theorem In simplest case P(B|A) = P(A|B) P(B)/P(A) and we consider the likelihood function in which we view the conditional.
1 A Bayesian statistical method for particle identification in shower counters IX International Workshop on Advanced Computing and Analysis Techniques.
Sample variance and sample error We learned recently how to determine the sample variance using the sample mean. How do we translate this to an unbiased.
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Bayesian statistics Probabilities for everything.
Making sense of randomness
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Sampling and estimation Petter Mostad
BAYES and FREQUENTISM: The Return of an Old Controversy 1 Louis Lyons Imperial College and Oxford University CERN Summer Students July 2014.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
Analysis of Experimental Data; Introduction
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
Maximum likelihood estimators Example: Random data X i drawn from a Poisson distribution with unknown  We want to determine  For any assumed value of.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
Statistics and probability Dr. Khaled Ismael Almghari Phone No:
Review Day 2 May 4 th Probability Events are independent if the outcome of one event does not influence the outcome of any other event Events are.
Confidence Limits and Intervals 3: Various other topics Roger Barlow SLUO Lectures on Statistics August 2006.
Probability and Statistics
BAYES and FREQUENTISM: The Return of an Old Controversy
Bayes Net Learning: Bayesian Approaches
Bayes for Beginners Stephanie Azzopardi & Hrvoje Stojic
Modelling data and curve fitting
Bayes for Beginners Luca Chech and Jolanda Malamud
CS 594: Empirical Methods in HCC Introduction to Bayesian Analysis
CS639: Data Management for Data Science
Statistics for HEP Roger Barlow Manchester University
Applied Statistics and Probability for Engineers
Presentation transcript:

Statistics for Particle Physics: Intervals Roger Barlow Karlsruhe: 12 October 2009

Summary Techniques ΔΧ 2 =1, Δln L=-½ 1D and 2+D Integrating and/or profiling Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits2 Concepts Confidence and Probability Chi squared p-values Likelihood Bayesian Probability

Simple example Measurement: value and Gaussian Error ± 2.1 means: to 68% to 95% to 99.7% etc Thus provides a whole set of intervals and associated probability/confidence values Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits3

Aside (1): Why Gaussian? Central Limit Theorem: The cumulative effect of many different uncertainties gives a Gaussian distribution – whatever the form of the constituent distributions. Moral: don’t worry about nonGaussian distributions. They will probably be combined with others. Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits4

Aside(2) Probability and confidence “169.9 to 68%” What does this mean? Number either is in this interval or it isn’t. Probability is either 0 or 1. This is not like population statistics. Reminder: basic definition of probability as limit of frequency P(A)= Limit N(A)/N Interpretation. ‘The statement “Q lies in the range to 173.3” has a 68% probability of being true.’ Statement made with 68% confidence Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits5

Illustration Simple straight line fit y=a x Estimate a=Σ x i y i / Σ x i 2 Error on a given by σ/√(Σx i 2 ) (combination of errors) Also look at χ 2 =Σ (y i -ax i ) 2 /σ 2 Size contains information on quality of fit Parabolic function of a 2 nd derivative gives error on a Can be read off from points where χ 2 = increases by 1 Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits6 a χ2χ2 1

Illustration Simple straight line fit y=a x+b Estimate a, b Errors on a,b and correlation given by combination of errors Also look at χ 2 =Σ (y-ax-b) 2 /σ 2 Parabolic function of a and b χ 2 contours map out confidence regions Values 2.30 for 68%, 5.99 for 95%, for 99.73% Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits7 a b

χ2χ2 Χ 2 Distribution is convolution of N Gaussians Expected χ 2 ≈N If χ 2 >> N the model is implausible. Quantify this using standard function F(χ 2 ;N) Fitting a parameter just reduces N by 1 Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits8

Chi squared probability and p values p(χ 2 ;N)=Integral of F from χ 2 to ∞ An example of a p-value :the probability that the true model would give a result this bad, or worse. Correct p-values are distributed uniformly between 0 and 1 Notice the choice to be made as to what is ‘bad’ Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits9

Likelihood L(a;x)=ΠP(x i ;a) Ln L(a;x)=Σ ln P(x i ;a) Regarded as function of a for given data x. For set of Gaussian measurements, clearly ln L = -½ χ 2 So -2 ln L behaves like a χ 2 distribution Generalisation (Wilks’ Theorem) this is true in other cases Find 1-σ confidence interval by Δln L = -½ OK for parabolic likelihood function Extension to nonparabolic functions is not rigorous but everybody does it Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits10 a ln L

Extend to several variables Map out region in parameter space where likelihood is above appropriate value Appears in many presentations of results] Sometimes both/all parameters are important Sometimes not… “Nuisance Parameters”, or systematic errors Basic rule is to say what you’re doing. Can use profile likelihood technique to include effect. Or integrate. Dubious but probably OK. Bayesian Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits11 a b a b

Bayes theorem P(A|B) = P(B|A) P(A) P(B) Example: Particle ID Bayesian Probability P(Theory|Data) = P(Data|Theory) P(Theory) P(Data) Example: bets on tossing a coin P(Theory): Prior P(Theory|Data): Posterior Apparatus all very nice but prior is subjective. Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits12

Bayes and distributions Extend method. For parameter a have prior probability distribution P(a) and then posterior probability distribution P(a|x) Intervals can be read off directly. In simple cases, Bayesian and frequentist approach gives the same results and there is no real reason to use a Bayesian analysis. Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits13

Nuisance parameters L(a,b;x) and b is of no interest (e.g. experimental resolution). May have additional knowledge e.g. from another channel L’(a;x)=  L(a,b;x) P(b) db Seems natural – but be careful Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits14

Summary Techniques ΔΧ 2 =1, Δln L=-½ 1D and 2+D Integrating and/or profiling Karlsruhe: 12 October 2009Roger Barlow: Intervals and Limits15 Concepts Confidence and Probability Chi squared p-values Likelihood Bayesian Probability