Conditional Observables Joe Tuggle BaBar ROOT/RooFit Workshop December 2007.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Statistical Techniques I EXST7005 Start here Measures of Dispersion.
Part II – TIME SERIES ANALYSIS C3 Exponential Smoothing Methods © Angel A. Juan & Carles Serrat - UPC 2007/2008.
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 13, Slide 1 Chapter 13 From Randomness to Probability.
The Estimation Problem How would we select parameters in the limiting case where we had ALL the data? k → l  l’ k→ l’ Intuitively, the actual frequencies.
Simulation with ArenaAppendix C – A Refresher on Probability and StatisticsSlide 1 of 33 A Refresher on Probability and Statistics Appendix C.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #22.
Fermilab, June 26 th 2001 Thoughts on the fitting procedure for the  c + lifetime with the        channel Gianluigi Boca.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 6 The Standard Deviation as a Ruler and the Normal Model.
Theoretical Probability Distributions We have talked about the idea of frequency distributions as a way to see what is happening with our data. We have.
Reducing the Local Bias in Calibrating the General COCOMO.
MINOS Feb Antineutrino running Pedro Ochoa Caltech.
Probability theory Much inspired by the presentation of Kren and Samuelsson.
Cal State Northridge  320 Ainsworth Sampling Distributions and Hypothesis Testing.
Copyright © 2010 Pearson Education, Inc. Chapter 19 Confidence Intervals for Proportions.
Estimation and Hypothesis Testing. The Investment Decision What would you like to know? What will be the return on my investment? Not possible PDF for.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 20 Testing Hypotheses About Proportions.
Sampling Distributions and Hypothesis Testing. 2 Major Points An example An example Sampling distribution Sampling distribution Hypothesis testing Hypothesis.
1 Sampling Distributions Presentation 2 Sampling Distribution of sample proportions Sampling Distribution of sample means.
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
Statistics for Data Miners: Part I (continued) S.T. Balke.
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
Bias and Variability Lecture 28 Section 8.3 Tue, Oct 25, 2005.
Copyright © 2010 Pearson Education, Inc. Chapter 6 The Standard Deviation as a Ruler and the Normal Model.
Chapter 6 Continuous Distributions The Gaussian (Normal) Distribution.
Chapter 6 The Standard Deviation as a Ruler and the Normal Model.
Status of the Glasgow B→hh analysis CP Working group γ from loops 14 th October 2010 Paul Sail, Lars Eklund and Alison Bates.
Slide 6-1 Copyright © 2004 Pearson Education, Inc.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 6 The Standard Deviation as a Ruler and the Normal Model.
RooFit – tools for ML fits Authors: Wouter Verkerke and David Kirkby.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #23.
Fractions + & – Mixed Numbers Can you simplify the following with ease? 1 5 / / 4 or 2 7 / / 3 If not then run through this slide show,
Copyright © 2009 Pearson Education, Inc. Chapter 6 The Standard Deviation as a Ruler and the Normal Model.
Summary Five numbers summary, percentiles, mean Box plot, modified box plot Robust statistic – mean, median, trimmed mean outlier Measures of variability.
Statistics In HEP Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Multiple Random Variables Two Discrete Random Variables –Joint pmf –Marginal pmf Two Continuous Random Variables –Joint Distribution (PDF) –Joint Density.
Probabilistic Forecasting. pdfs and Histograms Probability density functions (pdfs) are unobservable. They can only be estimated. They tell us the density,
Issues in Estimation Data Generating Process:
Probability = Relative Frequency. Typical Distribution for a Discrete Variable.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Two Random Variables.
1 CS161 Introduction to Computer Science Topic #9.
Copyright © 2009 Pearson Education, Inc. Chapter 19 Confidence Intervals for Proportions.
Random Variable The outcome of an experiment need not be a number, for example, the outcome when a coin is tossed can be 'heads' or 'tails'. However, we.
Generalized Linear Models (GLMs) and Their Applications.
Sampling and estimation Petter Mostad
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Topic 5: Continuous Random Variables and Probability Distributions CEE 11 Spring 2002 Dr. Amelia Regan These notes draw liberally from the class text,
Using Probability Distributions for Project 2. Class Project How can probability distributions help us price a stock option? Last time we computed the.
Machine Learning 5. Parametric Methods.
INFERENTIAL STATISTICS DOING STATS WITH CONFIDENCE.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 6 The Standard Deviation as a Ruler and the Normal Model.
LECTURE 07: CLASSIFICATION PT. 3 February 15, 2016 SDS 293 Machine Learning.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 19 Confidence Intervals for Proportions.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Systematics in Hfitter. Reminder: profiling nuisance parameters Likelihood ratio is the most powerful discriminant between 2 hypotheses What if the hypotheses.
Guglielmo De Nardo for the BABAR collaboration Napoli University and INFN ICHEP 2010, Paris, 23 July 2010.
LNF 12/12/06 1 F.Ambrosino-T. Capussela-F.Perfetto Update on        Dalitz plot slope Where we started from A big surprise Systematic checks.
NASSP Masters 5003F - Computational Astronomy Lecture 4: mostly about model fitting. The model is our estimate of the parent function. Let’s express.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 6- 1.
Chapter 10 Confidence Intervals for Proportions © 2010 Pearson Education 1.
Prediction and Missing Data. Summarising Distributions ● Models are often large and complex ● Often only interested in some parameters – e.g. not so interested.
The asymmetric uncertainties on data points in Roofit
Top Changwatchai 18 October 2000
CMS RooStats Higgs Combination Package
Multi-dimensional likelihood
Weighted Least Squares Fit
Chapter 6 Confidence Intervals
Mathematical Foundations of BME Reza Shadmehr
Presentation transcript:

Conditional Observables Joe Tuggle BaBar ROOT/RooFit Workshop December 2007

2 Conditional PDFs A conditional probability density function answers the question, “Given y, what is the probability density of x?” y is the conditional observable. This is not a 2-dimensional probability density. That would be, “What is the combined probability density of x and y?” In effect, the observable y gets demoted from the argument of a function to a parameter of it.

3 Example Let m ES be defined in the range [5.2,5.3] GeV/c 2 and let  = –20. Given an endpoint of 5.29 GeV/c 2 what is the probability to find m ES in the range [5.275,5.285] GeV/c 2 ? If the endpoint is GeV/c 2, the probability becomes

4

5 Relation to 2D PDFs On the first slide I said that a conditional PDF p(x|y) is not a 2D PDF. We’ll denote the 2D PDF by p(x,y). However, these things are related: To get the full 2D PDF you need to know something about the distribution of y, whereas with the conditional version, you don’t care. p(x,y) =p(x|y)p(y)

6 Effect on Normalization When you want to get something useful out of your PDF, it needs to be normalized. To normalize a regular 2D PDF p(x,y) you divide by However, since a conditional PDF p(x|y) treats y as a parameter, you only divide a conditional PDF by

7 Fitting Your likelihood function is the product of normalized probability densities of all events in your data set: If your PDF is just p(x,y), then the constant denominator is irrelevant to maximizing the likelihood. However…

8 Fitting If you’re using p(x|y), then each event is weighted by and your likelihood function becomes

9 Fitting In RooFit, you can tell it to treat certain observables as conditional when you fit your PDF: RooRealVar x; RooRealVar y; RooDataSet ds; RooSomePdf myPdf(“myPdf”, “myPdf”, x, y); myPdf.fitTo(ds, ConditionalObservables(RooArgSet(y));

10 Plotting The different normalization required when using conditional observables also affects plotting. To project out an observable from a 2D PDF, you just integrate: However, if y is conditional, then it’s a parameter – not something you integrate over.

11 Plotting Since we can’t integrate over y, we average. Instead of We take an average over the data to get:

12 Plotting In RooFit, you can tell it to average over the data when plotting: Specifying the RooArgSet of conditional observables is crucial – otherwise you get an average over ALL of the observables in the dataset, which is probably not what you want. If this method takes too long to calculate, you can bin your conditional observable in a RooDataHist and pass that as the data set argument. RooRealVar x; RooRealVar y; RooDataSet ds; RooSomePdf myPdf(“myPdf”, “myPdf”, x, y); RooPlot* myPlot = x.frame(); myPdf.plotOn(myPlot, ProjWData(RooArgSet(y), ds));

13 Toy MC Since a conditional PDF knows nothing about the distribution of its conditional observable, that information has to come from somewhere else. Can’t just generate toy MC from your PDF. If you really don’t know anything about the observable, you can take it from your actual data set or the centrally-produced MC. Can run into statistics issues if you want to do a lot of toys. Otherwise, you can make up a reasonable PDF for it, and generate a prototype dataset from that PDF.

14 Toy MC To specify conditional observables and prototype data in RooFit, declare your RooMCStudy like this: RooAbsPdf theFullPdf; RooArgSet allObservables; RooDataSet* theYDistr = yPdf.Generate(nEvt); RooMCStudy myStudy(theFullPdf, allObservables, ConditionalObservables(RooArgSet(y)), ProtoData(theYDistr));

15 Pitfalls If your conditional observable has a different distribution for different species (like signal and background), then your fit can be biased. For a simple example, see arXiv:physics/ v1 [physics.data-an]. Not really a problem with the m ES endpoint example, but conceivably a problem in other applications. You might notice this in embedded toy studies. It can be explicitly tested in pure toys by generating with different distributions of your conditional observable for signal and background, then fitting with the conditional PDFs.

16 Summary Some observables have to be treated on an event- by-event basis. This affects the normalization of your PDFs, and therefore impacts the way you fit, plot, and validate your fitter. Watch out for conditional observables that look very different between signal and background events. If they are so different, maybe you should use this as a discriminating observable in your fit! RooFit has built-in methods to properly deal with conditional observables, but you have to tell it how to handle them.