Concluding Talk: Physics Gary Feldman Harvard University PHYSTAT 05 University of Oxford 15 September, 2005.

Slides:



Advertisements
Similar presentations
Group A: Limits and p-values with nuisance parameters Banff, July 2006.
Advertisements

Using the Profile Likelihood in Searches for New Physics / PHYSTAT 2011 G. Cowan 1 Using the Profile Likelihood in Searches for New Physics arXiv:
27 th March CERN Higgs searches: CL s W. J. Murray RAL.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
1 LIMITS Why limits? Methods for upper limits Desirable properties Dealing with systematics Feldman-Cousins Recommendations.
I’ll not discuss about Bayesian or Frequentist foundation (discussed by previous speakers) I’ll try to discuss some facts and my point of view. I think.
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #22.
An Answer and a Question Limits: Combining 2 results Significance: Does  2 give  2 ? Roger Barlow BIRS meeting July 2006.
Statistics In HEP 2 Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Maximum likelihood (ML) and likelihood ratio (LR) test
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Maximum likelihood (ML) and likelihood ratio (LR) test
1 Nuisance parameters and systematic uncertainties Glen Cowan Royal Holloway, University of London IoP Half.
Optimization of Signal Significance by Bagging Decision Trees Ilya Narsky, Caltech presented by Harrison Prosper.
Evaluating Hypotheses
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Basics of discriminant analysis
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 41 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability.
8-1 Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall Chapter 8 Confidence Interval Estimation Statistics for Managers using Microsoft.
Copyright ©2011 Pearson Education 8-1 Chapter 8 Confidence Interval Estimation Statistics for Managers using Microsoft Excel 6 th Global Edition.
G. Cowan Lectures on Statistical Data Analysis Lecture 13 page 1 Statistical Data Analysis: Lecture 13 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
1 Statistical Inference Problems in High Energy Physics and Astronomy Louis Lyons Particle Physics, Oxford BIRS Workshop Banff.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Maximum likelihood (ML)
Statistical Analysis of Systematic Errors and Small Signals Reinhard Schwienhorst University of Minnesota 10/26/99.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Confidence Interval Estimation Basic Business Statistics 11 th Edition.
Confidence Interval Estimation
Probability theory 2 Tron Anders Moger September 13th 2006.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
Random Sampling, Point Estimation and Maximum Likelihood.
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
SFT group meeting 1 Phystat05 : Trip Report A. Kreshuk, L. Moneta.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Practical Statistics for Particle Physicists Lecture 3 Harrison B. Prosper Florida State University European School of High-Energy Physics Anjou, France.
Statistics In HEP Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Section 10.1 Confidence Intervals
Statistical Methods for Data Analysis Introduction to the course Luca Lista INFN Napoli.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
Chap 8-1 Chapter 8 Confidence Interval Estimation Statistics for Managers Using Microsoft Excel 7 th Edition, Global Edition Copyright ©2014 Pearson Education.
Sampling and estimation Petter Mostad
Experience from Searches at the Tevatron Harrison B. Prosper Florida State University 18 January, 2011 PHYSTAT 2011 CERN.
Practical Statistics for Particle Physicists Lecture 2 Harrison B. Prosper Florida State University European School of High-Energy Physics Parádfürdő,
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Lectures on Statistical Data Analysis Lecture 4 page 1 Lecture 4 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Lecture 8: Measurement Errors 1. Objectives List some sources of measurement errors. Classify measurement errors into systematic and random errors. Study.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 12 page 1 Statistical Data Analysis: Lecture 12 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
G. Cowan RHUL Physics Statistical Issues for Higgs Search page 1 Statistical Issues for Higgs Search ATLAS Statistics Forum CERN, 16 April, 2007 Glen Cowan.
1 Nuisance parameters and systematic uncertainties Glen Cowan Royal Holloway, University of London IoP Half.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Practical Statistics for LHC Physicists Frequentist Inference Harrison B. Prosper Florida State University CERN Academic Training Lectures 8 April, 2015.
arXiv:physics/ v3 [physics.data-an]
Two Interpretations of What it Means to Normalize the Low Energy Monte Carlo Events to the Low Energy Data Atms MC Atms MC Data Data Signal Signal Apply.
Lecture 4 1 Probability (90 min.)
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Introduction to Statistics − Day 4
Lecture 4 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Statistics for HEP Roger Barlow Manchester University
Presentation transcript:

Concluding Talk: Physics Gary Feldman Harvard University PHYSTAT 05 University of Oxford 15 September, 2005

Gary Feldman PHYSTAT September Topics l I will restrict my comments to two topics, both of which I am interested in and both of which received some attention at this meeting: l Event classification l Nuisance parameters

Gary Feldman PHYSTAT September Event Classification l The problem: Given a measurement of an event X = (x 1,x 2,…x n ), find the function F(X) which returns 1 if the event is signal (s) and 0 if the event is background (b) to optimize a figure of merit, say signal.

Gary Feldman PHYSTAT September Theoretical Solution l In principle the solution is straightforward: Use a Monte Carlo simulation to calculate the likelihood ratio L s (X)/L b (X) and derive F(X) from it. By the Neyman-Pearson Theorem, this is the optimum solution. l Unfortunately, this does not work due to the “curse of dimensionality.” In a high-dimension space, even the largest data set is sparse with the distance between neighboring events comparable to the radius of the space.

Gary Feldman PHYSTAT September Practical Solutions l Thus, we are forced to substitute cleverness for brute force. l In recent years, physicists have come to learn that computers may be cleverer than they are. l They have turned to machine learning: One gives the computer samples of signal and background events and lets the computer figure out what F(X) is.

Gary Feldman PHYSTAT September Artificial Neural Networks l Originally most of this effort was in artificial neural networks (ANN). Although used successfully in many experiments, ANNs tend to be finicky and often require real cleverness from their creators. l At this conference, there was an advance in ANNs reported by Harrison Prosper. The technique is to average over a collection of networks. Each network is constructed by sampling the weight probability density constructed from the training sample.

Gary Feldman PHYSTAT September Trees and Rules l In the past couple of years, interest has started to shift to other techniques, such as decision trees, at least partially sparked by Jerry Friedman’s talk at PHYSTAT 03. l A single decision tree has limited power, but its power can be increased by techniques that effectively sum many trees. A cartoon from Roe’s talk

Gary Feldman PHYSTAT September Rules and Bagging Trees l Jerry Friedman gave a talk on rules, which effectively combines a series of trees. l Harrison Prosper gave a talk (for Ilya Narsky) on bagging (Bootstrap AGGregatING) trees. In this technique, one builds a collection of trees by selecting a sample of the training data and, optionally, a subset of the variables. Results on significance of B   e at BaBar Single decision tree 2.16  Boosted decision trees 2.62  (not optimized) Bagging decision trees 2.99 

Gary Feldman PHYSTAT September Boosted Decision Trees l Byron Roe gave a talk on the use of boosted trees in MiniBooNE. Misclassified events in one tree are given a higher weight and a new tree is generated. Repeat to generate 1000 trees. The final classifier is a weighted sum of all of the trees. l Comparison to ANN: Also more robust. % of signal retained 52 variables 21 variables

Gary Feldman PHYSTAT September Other Talks l There were a couple of other talks on this subject by Puneet Sarda and Alex Gray, which I could not attend.

Gary Feldman PHYSTAT September Nuisance Parameters l Nuisance parameters are parameters with unknown true values for which coverage is required in a frequentist analysis. l They may be statistical, such as number of background events in a sideband used for estimating the background under a peak. l They may be systematic, such as the shape of the background under the peak, or the error caused by the uncertainty of the hadronic fragmentation model in the Monte Carlo. l Most experiments have a large number of systematic uncertainties.

Gary Feldman PHYSTAT September New Concerns for the LHC l Although the statistical treatment of these uncertainties is probably the analysis question that I have been asked the most, Kyle Cranmer has pointed out that these issues will be even more important at the LHC. l If the statistical error is O(1) and the systematic error is O(0.1), the the systematic error will contribute as its square or O(0.01) and it does not much matter how you treat it. l However, at the LHC, we may have process with 100 background events and 10% systematic errors. Even more critical, we want 5  for a discovery level.

Gary Feldman PHYSTAT September Why 5  ? l LHC searches: 500 searches each of which has 100 resolution elements (mass, angle bins, etc.) x 5 x 10 4 chances to find something. One experiment: False positive rate at 5   (5 x 10 4 ) (3 x ) = OK. l Two experiments: l Allowable false positive rate: (5 x 10 4 ) (1 x ) = 10  3.7  required. Required other experiment verification: (1 x )(10) = 0.01  3.1  required. l Caveats: Is the significance real? Are there common systematic errors?

Gary Feldman PHYSTAT September A Cornucopia of Techniques l At this meeting we have seen a wide series of techniques discussed for constructing confidence intervals in the presence of nuisance parameters. l Everyone has expressed a concern that their methods cover, at least approximately. This appears to be important for LHC physics in light of Cranmer’s concerns.

Gary Feldman PHYSTAT September Bayesian with Coverage l Joel Heinrich presented a decision by CDF to do Bayesian analyses with priors that cover. Advantage is Bayesian conditioning with frequentist coverage. Possibly the maximum amount of work for the experimenter. l Example of coverage with a single Poisson with normalization and background nuisance parameters: Flat priors

Gary Feldman PHYSTAT September Bayesian with Coverage Example of coverage with flat and 1/  and 1/b priors for a 4-channel Poisson with normalization and background nuisance parameters Flat priors 1/  and 1/b priors

Gary Feldman PHYSTAT September Frequentist/Bayesian Hybrid l Fredrik Tegenfeldt presented a likelihood-ratio ordered (LR) Neyman construction after integrating out the nuisance parameters with a flat priors. In a single channel test, there was no undercoverage. l What happens for a multi-channel case? My guess is that the confidence belt will be distorted by the use of flat priors, but that the method will still cover due to the construction. l Cranmer considers a similar technique, as was used for LEP Higgs searches. l Both are call “Cousins-Highland,” although probably neither actually is.

Gary Feldman PHYSTAT September Profile Likelihood l 44 years ago, Kendall and Stuart told us how to eliminate nuisance parameters and do a LR construction:

Gary Feldman PHYSTAT September One (Minor) Problem l The Kendall-Stuart prescription leads to the problem that for Poisson problems as the nuisance parameter is better and better known, the confidence intervals do not converge to the limit of being perfectly known. The reason is that the introduction of a nuisance par- ameter breaks the discreteness of the Poisson distribution. From Punzi’s talk

Gary Feldman PHYSTAT September One More Try l Since this was referred to in a parallel session as “the Feldman problem” and since two plenary speakers made fun of my Fermilab Workshop plots, I will try to explain them again.

Gary Feldman PHYSTAT September The Cousins-Highland Problem l This correction also solves what Bob and I refer to as the Cousins-Highland problem (as opposed to method). l Cousins and Highland turned to a Bayesian approach to calculate the effect of a normalization error because the frequentist approach gave an answer with the wrong sign. l We now understand this was due to simply breaking the discreteness of the Poisson distribution. l In one test case, using this correction reproduced the Cousins-Highland result x/  2.

Gary Feldman PHYSTAT September Use of Profile Likelihood Wolfgang Rolke presented a talk on eliminating the nuisance parameters via profile likelihood, but with the Neyman construction replaced by the -  lnL hill-climbing approximation. This is also what MINUIT does. The coverage is good with some minor undercoverage. Cranmer also considers this method.

Gary Feldman PHYSTAT September Full Neyman Constructions l Both Giovanni Punzi and Kyle Cranmer attempted full Neyman constructions for both signal and nuisance parameters. l I don’t recommend you try this at home for the following reasons: l The ordering principle is not unique. Both Punzi and Cranmer ran into some problems. l The technique is not feasible for more than a few nuisance parameters. l It is unnecessary since removing the nuisance parameters through profile likelihood works quite well.

Gary Feldman PHYSTAT September Cranmer’s (Revised) Conclusions l In Cranmer’s talk, he had an unexpected result for the coverage of Rolke’s method (“profile”). He did in fact have an error and it is corrected here:

Gary Feldman PHYSTAT September Final Comments on Nuisance Parameters l My preference is to eliminate at least the major nuisance parameters through profile likelihood and then do a LR Neyman construction. It is straightforward and has excellent coverage properties. l However, whatever method you choose, you should check the coverage of the method. l Cranmer makes the point that if you can check the coverage, you can also do a Neyman construction. I don’t completely agree, but it is worth considering.