iSTEP 2016 Tsinghua University, Beijing July 10-20, 2016

Slides:



Advertisements
Similar presentations
G. Cowan TAE 2013 / Statistics Problems1 TAE Statistics Problems Glen Cowan Physics Department Royal Holloway, University of London
Advertisements

Using the Profile Likelihood in Searches for New Physics / PHYSTAT 2011 G. Cowan 1 Using the Profile Likelihood in Searches for New Physics arXiv:
G. Cowan RHUL Physics Profile likelihood for systematic uncertainties page 1 Use of profile likelihood to determine systematic uncertainties ATLAS Top.
G. Cowan RHUL Physics Comment on use of LR for limits page 1 Comment on definition of likelihood ratio for limits ATLAS Statistics Forum CERN, 2 September,
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 41 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability.
G. Cowan RHUL Physics Higgs combination note status page 1 Status of Higgs Combination Note ATLAS Statistics/Higgs Meeting Phone, 7 April, 2008 Glen Cowan.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination using shapes ATLAS Statistics Meeting CERN, 19 December, 2007 Glen Cowan.
G. Cowan Lectures on Statistical Data Analysis Lecture 7 page 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan 2009 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 4 Lecture 1 Probability Random variables, probability densities,
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
G. Cowan Statistical Methods in Particle Physics1 Statistical Methods in Particle Physics Day 3: Multivariate Methods (II) 清华大学高能物理研究中心 2010 年 4 月 12—16.
Comparison of Bayesian Neural Networks with TMVA classifiers Richa Sharma, Vipin Bhatnagar Panjab University, Chandigarh India-CMS March, 2009 Meeting,
G. Cowan CLASHEP 2011 / Topics in Statistical Data Analysis / Lecture 21 Topics in Statistical Data Analysis for HEP Lecture 2: Statistical Tests CERN.
G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Lecture 2 1 Probability Definition, Bayes’ theorem, probability densities and their properties,
G. Cowan, RHUL Physics Discussion on significance page 1 Discussion on significance ATLAS Statistics Forum CERN/Phone, 2 December, 2009 Glen Cowan Physics.
פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
G. Cowan IDPASC School of Flavour Physics, Valencia, 2-7 May 2013 / Statistical Analysis Tools 1 Statistical Analysis Tools for Particle Physics IDPASC.
2005 Unbinned Point Source Analysis Update Jim Braun IceCube Fall 2006 Collaboration Meeting.
1 Introduction to Statistics − Day 2 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
G. Cowan Lectures on Statistical Data Analysis Lecture 6 page 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Aachen 2014 / Statistics for Particle Physics, Lecture 21 Statistical Methods for Particle Physics Lecture 2: statistical tests, multivariate.
G. Cowan CERN Academic Training 2010 / Statistics for the LHC / Lecture 21 Statistics for the LHC Lecture 2: Discovery Academic Training Lectures CERN,
G. Cowan SLAC Statistics Meeting / 4-6 June 2012 / Two Developments 1 Two developments in discovery tests: use of weighted Monte Carlo events and an improved.
G. Cowan CERN Academic Training 2012 / Statistics for HEP / Lecture 21 Statistics for HEP Lecture 2: Discovery and Limits Academic Training Lectures CERN,
Discussion on significance
Chapter 7. Classification and Prediction
Computing and Statistical Data Analysis / Stat 11
Background on Classification
Tutorial on Statistics TRISEP School 27, 28 June 2016 Glen Cowan
Statistics for the LHC Lecture 3: Setting limits
Some Statistical Tools for Particle Physics
LAL Orsay, 2016 / Lectures on Statistics
Comment on Event Quality Variables for Multivariate Analyses
Advanced Analytics Using Enterprise Miner
Data Analysis in Particle Physics
Tutorial on Multivariate Methods (TMVA)
ISTEP 2016 Final Project— Project on
TAE 2017 Centro de ciencias Pedro Pascual Benasque, Spain
School on Data Science in (Astro)particle Physics
Graduierten-Kolleg RWTH Aachen February 2014 Glen Cowan
Discrete Event Simulation - 4
Computing and Statistical Data Analysis / Stat 8
Computing and Statistical Data Analysis Stat 3: The Monte Carlo Method
Lecture 3 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
Project on H →ττ and multivariate methods
Using Single Photons for WIMP Searches at the ILC
TAE 2018 / Statistics Lecture 3
Computing and Statistical Data Analysis Stat 5: Multivariate Methods
Statistical Methods for the LHC
TAE 2018 Benasque, Spain 3-15 Sept 2018 Glen Cowan Physics Department
Computing and Statistical Data Analysis / Stat 6
Computing and Statistical Data Analysis / Stat 7
TRISEP 2016 / Statistics Lecture 2
Statistical Methods for HEP Lecture 3: Discovery and Limits
Statistical Analysis Tools for Particle Physics
Decomposition of Stat/Sys Errors
TAE 2018 Centro de ciencias Pedro Pascual Benasque, Spain
Introduction to Statistics − Day 4
Project on H→ττ and multivariate methods
Statistical Methods for Data Analysis Multivariate discriminators with TMVA Luca Lista INFN Napoli.
Why do Wouter (and ATLAS) put asymmetric errors on data points ?
Presentation transcript:

Statistical Methods for Particle Physics Tutorial on multivariate methods http://indico.ihep.ac.cn/event/5966/ iSTEP 2016 Tsinghua University, Beijing July 10-20, 2016 Glen Cowan (谷林·科恩) Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAA

Test example with TMVA Suppose signal (blue) and background (red) events are each characterized by 3 variables, x, y, z: G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Test example (x, y, z) y y no cut on z z < 0.75 x x y y z < 0.25 G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Goal of test example: discovery of signal We want to define a region of (x,y,z) space using a test statistic t, e.g., t > tcut, and to search for evidence of the signal process by counting events in this region. The number of events n that we observe will follow a Poisson distribution with mean s + b, where s and b are the expected numbers of signal and background events with t > tcut. Goal is to maximize the expected significance for test of s = 0 hypothesis if signal is present. The expected discovery significance can be estimated using (For s << b this is approximately s/√b.) G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Code for tutorial The code for the tutorial is here: http://www.pp.rhul.ac.uk/~cowan/stat/istep2016/tmva Copy the file tmvaExamples.tar to your own working directory and unpack using tar –xvf tmvaExamples.tar This will create some subdirectories, including generate train test analyze inc G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Generate the data cd into the subdirectory generate, build and run the program generateData by typing make ./generateData This creates two data files: trainingData.root and testData.root an each of which contains a TTree (n-tuple) for signal and background events. G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Train the classifier cd into the subdirectory train, build and run the program generateData by typing make ./tmvaTrain This uses the data in trainingData.root to train a Fisher discriminant and writes the coefficients to a file tmvaTest_Fisher.weights.xml. It also creates a file TMVA.root. Copy this file into the subdirectory test, and then you can look at various diagnostic histograms using the macros there. Try e.g. .x plotall.C G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Distribution of classifier output (Fisher) G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Analyze the data cd into the subdirectory analyze, build and run the program analyzeData by typing make ./analyzeData This reads in the data from testData.root and selects events with values of the Fisher statistic t greater than a given threshold tcut (set initially to zero). The program counts the number of events passing the threshold and from these estimates the efficiencies G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Discovery significance Suppose the data sample we have corresponds to an integrated luminosity of L = 20 fb-1, and that the cross sections of the signal and background processes are σs = 0.2 fb and σb = 10 fb. The program then calculates the expected number of signal and background events after the cut using From these numbers we can compute the expected discovery significance using the “Asimov” formula: G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Extending the example We would like to optimize the value of tcut to obtain the maximum expected discovery significance. To do this, modify the code in analyzeData.cc to loop over a range of values of tcut. You should find a plot something like G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Hints for making plots To make a simple plot, you can use the root class TGraph. See https://root.cern.ch/root/html/TGraph.html Suppose you have values x1,..., xn and y1,..., yn in arrays x and y of type double and (int) length n. Then plot with TGraph* g = new TGraph(n, x, y); g->Draw(); This can be done either in your C++ code or in a root macro. To make a plot with error bars use TGraphErrors. G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Using different classifiers The Fisher discriminant is a linear classifier and is not expected to give the optimal performance for this problem. You can try adding classifiers to tmvaTrain.cc by adding the lines (adds Boosted Decision Tree with 200 boosting iterations) (adds a Multilayer Perceptron with a single hidden layer with 3 nodes) You need to make corresponding changes in analyzeData.cc Try using these classifiers to find the maximum expected discovery significance. You can also try modifying the architecture of the classifiers or use different classifiers (see TMVA manual at tmva.sourceforge.net). G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Extension of TMVA Project select For the TMVA Project, you defined a test statistic t to separate between signal and background events. t tcut You selected events with t > tcut, calculated s and b, and estimated the expected discovery significance. This is OK for a start, but does not use all of the available information from each event’s value of the statistic t. G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Binned analysis Choose some number of bins (~20) for the histogram of the test statistic. In bin i, find the expected numbers of signal/background: Likelihood function for strength parameter μ with data n1,..., nN Statistic for test of μ = 0: (Asimov Paper: CCGV EPJC 71 (2011) 1554; arXiv:1007.1727) G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Discovery sensitivity First one should (if there is time) write a toy Monte Carlo program and generate data sets (n1,..., nN) following the μ = 0 hypothesis, i.e., ni ~ Poisson(bi) for the i = 1,..., N bins of the histogram. This can be done using the random number generator TRandom3 (see generateData.cc for an example and use ran->Poisson(bi).) From each data set (n1,..., nN), evaluate q0 and enter into a histogram. Repeat for at least 107 simulated experiments. You should see that the distribution of q0 follows the asymptotic “half-chi-square” form. G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Hints for computing q0 You should first show that ln L(μ) can be written where C represents terms that do not depend on μ. Therefore, to find the estimator μ, you need to solve ⌃ To do this numerically, you can use the routine fitPar.cc (header file fitPar.h). Put fitPar.h in the subdirectory inc and fitPar.cc in analyze. Modify GNUmakefile to have SOURCES = analyzeData.cc fitPar.cc INCLFILES = Event.h fitPar.h G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

To plot the histogram and superimpose To “book” (initialize) a histogram: TH1D* h = new TH1D(“h”, “my histogram”, numBins, xMin, xMax); To fill a value x into the histogram: h->Fill(x); To display the histogram: h->Draw(); To superimpose ½ times the chi-square distribution curve on the histogram use: TF1* func = new TF1("func", ScaledChi2, 0., 50., 2); func->SetParameter(0,1.0); // degrees of freedom func->SetParameter(1, 0.5); // scale factor 0.5 func->Draw("same"); You can get the function ScaledChi2.C. from subdirectory tools. G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Background-only distribution of q0 For background-only (μ = 0) toy MC, generate ni ~ Poisson(bi). Large-sample asymptotic formula is “half-chi-square”. G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

Discovery sensitivity Provided that the asymptotic approximation is valid, we can estimate the discovery significance (significance of test of μ= 0)) from the formula Median significance of test of background-only hypothesis under assumption of signal+background from “Asimov data set”: You can use the Asimov data set to evaluate q0 and use this with the formula Z = √q0 to estimate the median discovery significance. This should give a higher significance that what was obtained from the analysis based on a single cut. G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial

The Higgs Machine Learning Challenge An open competition (similar to our previous example) took pace from May to September 2014 using simulated data from the ATLAS experiment. Information can be found opendata.cern.ch/collection/ATLAS-Higgs-Challenge-2014 800k simulated ATLAS events for signal (H → ττ) and background (ttbar and Z → ττ) now publicly available. Each event characterized by 30 kinematic variables and a weight. Weights defined so that their sum gives expected number of events for 20 fb-1. Some code using TMVA is here (download and unpack as usual with tar –xvf): www.pp.rhul.ac.uk/~cowan/higgsml/tmvaHiggsML.tar G. Cowan iSTEP 2016, Beijing / Statistics for Particle Physics / Tutorial