Ivo van Vulpen Why does RooFit put asymmetric errors on data points ? 10 slides on a ‘too easy’ topic that I hope confuse, but do not irritate you

Slides:



Advertisements
Similar presentations
DO’S AND DONT’S WITH LIKELIHOODS Louis Lyons Oxford (CDF)
Advertisements

Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
Statistical Estimation and Sampling Distributions
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
1 LIMITS Why limits? Methods for upper limits Desirable properties Dealing with systematics Feldman-Cousins Recommendations.
I’ll not discuss about Bayesian or Frequentist foundation (discussed by previous speakers) I’ll try to discuss some facts and my point of view. I think.
Sampling Distributions (§ )
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Probability theory Much inspired by the presentation of Kren and Samuelsson.
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Standard error of estimate & Confidence interval.
Lecture 9: p-value functions and intro to Bayesian thinking Matthew Fox Advanced Epidemiology.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
Random variables Petter Mostad Repetition Sample space, set theory, events, probability Conditional probability, Bayes theorem, independence,
Probability theory 2 Tron Anders Moger September 13th 2006.
Statistical Decision Theory
The Triangle of Statistical Inference: Likelihoood
Overview course in Statistics (usually given in 26h, but now in 2h)  introduction of basic concepts of probability  concepts of parameter estimation.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
Why do Wouter (and ATLAS) put asymmetric errors on data points ? What is involved in the CLs exclusion method and what do the colours/lines mean ? ATLAS.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
MAS2317 Presentation Jake McGlen. Assuming vague prior knowledge, obtain the posterior distribution for µ and hence construct a 95% Bayesian confidence.
01/20151 EPI 5344: Survival Analysis in Epidemiology Maximum Likelihood Estimation: An Introduction March 10, 2015 Dr. N. Birkett, School of Epidemiology,
Learning Theory Reza Shadmehr Linear and quadratic decision boundaries Kernel estimates of density Missing data.
Likelihood Methods in Ecology November 16 th – 20 th, 2009 Millbrook, NY Instructors: Charles Canham and María Uriarte Teaching Assistant Liza Comita.
Likelihood function and Bayes Theorem In simplest case P(B|A) = P(A|B) P(B)/P(A) and we consider the likelihood function in which we view the conditional.
Bayesian statistics Probabilities for everything.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
Week 41 Estimation – Posterior mean An alternative estimate to the posterior mode is the posterior mean. It is given by E(θ | s), whenever it exists. This.
Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed.
Confidence Intervals First ICFA Instrumentation School/Workshop At Morelia, Mexico, November 18-29, 2002 Harrison B. Prosper Florida State University.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Confidence Interval & Unbiased Estimator Review and Foreword.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #24.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Estimating a Population Proportion ADM 2304 – Winter 2012 ©Tony Quon.
Bayesian Estimation and Confidence Intervals Lecture XXII.
arXiv:physics/ v3 [physics.data-an]
The asymmetric uncertainties on data points in Roofit
Bayesian Estimation and Confidence Intervals
Statistical methods in LHC data analysis introduction
Discrete Probability Distributions
Lecture 4 1 Probability (90 min.)
Maximum Likelihood Find the parameters of a model that best fit the data… Forms the foundation of Bayesian inference Slide 1.
More about Posterior Distributions
Bayesian Inference, Basics
Likelihood intervals vs. coverage intervals
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Sampling Distributions (§ )
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
CS639: Data Management for Data Science
Statistics for HEP Roger Barlow Manchester University
Why do Wouter (and ATLAS) put asymmetric errors on data points ?
Applied Statistics and Probability for Engineers
Presentation transcript:

Ivo van Vulpen Why does RooFit put asymmetric errors on data points ? 10 slides on a ‘too easy’ topic that I hope confuse, but do not irritate you ATLAS H  4 lepton peak Ivo van Vulpen (UvA/Nikhef)

Ivo van Vulpen ATLAS H  4 lepton peak - Summarize measurement - Make statement on underlying true value I’ll present 5 options. You tell me which one you prefer Why put an error on a data-point anyway ? Why put an error on a data-point anyway ?

Ivo van Vulpen Known λ (Poisson) Poisson distribution Probability to observe n events when λ are expected λ=4.00 Number of observed events #observed λ hypothesis fixed varying Binomial with n  ∞, p  0 en np=λ P(N obs |λ=4.0) ! !

Ivo van Vulpen Known λ (Poisson) Poisson distribution Probability to observe n events when λ are expected λ=4.90 Number of observed events #observed λ hypothesis fixed varying Binomial with n  ∞, p  0 en np=λ P(N obs |λ=4.9)

Ivo van Vulpen λ=1.00 λ=4.90 properties the famous √N (1) Mean: (2) Variance: (3) Most likely: first integer ≤ λ Option 1: Poisson spread for fixed λ

Ivo van Vulpen Treating it like a normal measurement What you have: 1) construct Likelihood λ as free parameter 3) Determine error interval: Δ(-2Log(lik)) = +1 2) Find value of λ that maximizes Likelihood

Ivo van Vulpen λ (hypothesis) P(N obs =4|λ) Note: normally you use -2log(Lik) Likelihood (ratio) Probability to observe n events when λ are expected Likelihood #observed λ hypothesis fixed varying Likelihood ! !

Ivo van Vulpen λ (hypothesis) P(N obs =4|λ) -2Log(P(N obs =4|λ)) Log(Likelihood) -2Log(Lik) ΔL=+1 Option 2: likelihood (ratio) Likelihood: P(4|λ)

Ivo van Vulpen Bayesian: statement on true value of λ Likelihood: Poisson distribution “what can I say about the measurement (number of observed events) given a theory expectation ?” Posterior pdf for λ: “what can I say about the underlying theory (true value of λ) given that I have observed of 4 events ?” What you have: What you want:

Ivo van Vulpen Choice of prior P(λ): Assume all values for λ are equally likely (“I know nothing”) λ (hypothesis) LikelihoodProbability density function Posterior PDF for λ  Integrate to get confidence interval Bayesian: statement on true value of λ

Ivo van Vulpen Option 3 and 4: Bayesian 16% 68% 8.3% 23.5% 68% Bayes central Bayes equal prob. option 1 option 2 Equal likelihood Ordering rule

Ivo van Vulpen Option 4: Frequentist largest λ fow which P(n≥n obs |λ)= smallest λ for which P(n≤n obs |λ)= λ = % Poisson λ = % Poisson If λ < 7.16 then probability to observe 4 events (or less) <16% Note: also using data that you did not observe

Ivo van Vulpen The options Poisson Likelihood Bayesian central Bayesian equal prob Frequentist Δλ:

Ivo van Vulpen Think about it and discuss with your colleagues tonight I’ll give the answer tomorrow

Ivo van Vulpen The options Poisson Likelihood Bayesian central Bayesian equal prob Frequentist Δλ:

Ivo van Vulpen ATLAS H  4 lepton peak Conclusion: - Now you know what RooFit uses - Hope you are a bit confused BobCousins_Poisson.pdf  Paper with details PoissonError.C  Implementation options shown here Ivo_Analytic_Poisson.pdf  Analytic properties (inverted) Poisson