1 Xavier Prudent - LAPP BaBar Collaboration Meeting June 2006 - Montreal Andreas Höcker (ATLAS), Kai Voss (ATLAS), Helge Voss (LHCb), Jörg Stelzer (BaBar),

Slides:



Advertisements
Similar presentations
HEP Data Mining with TMVA
Advertisements

Statistical Methods for Data Analysis Multivariate discriminators with TMVA Luca Lista INFN Napoli.
Statistical Issues at LHCb Yuehong Xie University of Edinburgh (on behalf of the LHCb Collaboration) PHYSTAT-LHC Workshop CERN, Geneva June 27-29, 2007.
S.Towers TerraFerMA TerraFerMA A Suite of Multivariate Analysis tools Sherry Towers SUNY-SB Version 1.0 has been released! useable by anyone with access.
Matthew Schwartz Harvard University March 24, Boost 2011.
Kalanand Mishra BaBar Coll. Meeting Sept 26, /12 Development of New SPR-based Kaon, Pion, Proton, and Electron Selectors Kalanand Mishra University.
RooUnfold unfolding framework and algorithms
TMVA TMVA Workshop 2011, CERN, Jan 21 Cuts and Likelihood Classifiers in TMVA Jörg Stelzer – Michigan State University TMVA Workshop 2011, CERN, Geneva,
TMVA – Toolkit for Multivariate Analysis
TMVA A Toolkit for (Parallel) MultiVariate Data Analysis
Searching for Single Top Using Decision Trees G. Watts (UW) For the DØ Collaboration 5/13/2005 – APSNW Particles I.
Handwritten Character Recognition Using Artificial Neural Networks Shimie Atkins & Daniel Marco Supervisor: Johanan Erez Technion - Israel Institute of.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Optimization of Signal Significance by Bagging Decision Trees Ilya Narsky, Caltech presented by Harrison Prosper.
G. Cowan 2007 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 3 Lecture 1 Probability Random variables, probability densities,
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
July 11, 2001Daniel Whiteson Support Vector Machines: Get more Higgs out of your data Daniel Whiteson UC Berkeley.
Vector Boson Scattering At High Mass
Artificial Neural Networks
TMVA Andreas Höcker (CERN) CERN meeting, Oct 6, 2006 Toolkit for Multivariate Data Analysis.
G. Cowan Lectures on Statistical Data Analysis Lecture 7 page 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem 2Random variables and.
1 Helge Voss Nikhef 23 rd - 27 th April 2007TMVA Toolkit for Multivariate Data Analysis: ACAT 2007 TMVA Toolkit for Multivariate Data Analysis with ROOT.
Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
G. Cowan Statistical Methods in Particle Physics1 Statistical Methods in Particle Physics Day 3: Multivariate Methods (II) 清华大学高能物理研究中心 2010 年 4 月 12—16.
TMVA Jörg Stelzer: Machine Learning withCHEP 2007, Victoria, Sep 5 Machine Learning Techniques for HEP Data Analysis with TMVA Toolkit for Multivariate.
Comparison of Bayesian Neural Networks with TMVA classifiers Richa Sharma, Vipin Bhatnagar Panjab University, Chandigarh India-CMS March, 2009 Meeting,
Matthew Schwartz Harvard University with J. Gallicchio, PRL, 105:022001,2010 (superstructure) with K. Black, J. Gallicchio, J. Huth, M. Kagan and B. Tweedie.
Use of Multivariate Analysis (MVA) Technique in Data Analysis Rakshya Khatiwada 08/08/2007.
1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker.
Training of Boosted DecisionTrees Helge Voss (MPI–K, Heidelberg) MVA Workshop, CERN, July 10, 2009.
Statistical Methods for Data Analysis Introduction to the course Luca Lista INFN Napoli.
1 PAT meeting, CERN, April 5, 2006 Kai Voss – TMVA Toolkit for MultiVariate Analysis TMVA – toolkit for parallel multivariate data analysis – Andreas Höcker.
Parallelization of likelihood functions for data analysis Alfio Lazzaro CERN openlab Forum on Concurrent Programming Models and Frameworks.
G. Cowan 2011 CERN Summer Student Lectures on Statistics / Lecture 31 Introduction to Statistics − Day 3 Lecture 1 Probability Random variables, probability.
1 Krakow Seminar, Feb 27, 2007A. Hoecker: Data Mining with TMVA HEP Data Mining with TMVA  ToolKit for Multivariate Analysis with ROOT  Andreas Hoecker.
Physics of the B-factories, MVA chapter 12nd workshop, KEK, May 2010 José Ocariz LPNHE – IN2P3 et Université Paris Diderot On behalf of the MVA chapter.
פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב
1 Teilchenseminar, April 26, 2007K. Voss: TMVA toolkit TMVA  ToolKit for Multivariate Analysis with ROOT  Kai Voss ( * ) Teilchenseminar, Bonn, April.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Kalanand Mishra BaBar Coll. Meeting February, /8 Development of New Kaon Selectors Kalanand Mishra University of Cincinnati.
1 Introduction to Statistics − Day 2 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
1 CATPhys meeting, CERN, Mar 13, 2006 Andreas Höcker – TMVA – toolkit for parallel multivariate data analysis – Andreas Höcker (ATLAS), Helge Voss (LHCb),
Multivariate Classifiers or “Machine Learning” in TMVA
G. Cowan Lectures on Statistical Data Analysis Lecture 6 page 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem 2Random variables and.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Neural Network Analysis of Dimuon Data within CMS Shannon Massey University of Notre Dame Shannon Massey1.
One framework for most common MVA-techniques, available in ROOT Have a common platform/interface for all MVA classification and regression-methods: Have.
LCFI physics studies meeting, 7 th December 04Sonja Hillertp. 1 Charge efficiency and leakage rates  purity vs efficiency plots give only part of the.
Multivariate Data Analysis with TMVA4 Jan Therhaag ( * ) (University of Bonn) ICCMSE09 Rhodes, 29. September 2009 ( * ) On behalf of the present core developer.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Multivariate Methods of
Machine Learning Developments in ROOT Sergei Gleyzer, Lorenzo Moneta
Multivariate Methods of
Cluster Classification Studies with MVA techniques
iSTEP 2016 Tsinghua University, Beijing July 10-20, 2016
Search for b → u transitions in B+ → {Kpp0}DK+
Theoretical Motivation Analysis Procedure Systematics Results
Tutorial on Statistics TRISEP School 27, 28 June 2016 Glen Cowan
Multivariate Data Analysis
Multi-dimensional likelihood
Advanced Analytics Using Enterprise Miner
10th China HEPS Particle Physics Meeting
Machine Learning Today: Reading: Maria Florina Balcan
Computing and Statistical Data Analysis Stat 5: Multivariate Methods
Toolkit for Multivariate Data Analysis Helge Voss, MPI-K Heidelberg
Statistical Methods for Data Analysis Multivariate discriminators with TMVA Luca Lista INFN Napoli.
Measurement of the Single Top Production Cross Section at CDF
Presentation transcript:

1 Xavier Prudent - LAPP BaBar Collaboration Meeting June Montreal Andreas Höcker (ATLAS), Kai Voss (ATLAS), Helge Voss (LHCb), Jörg Stelzer (BaBar), Peter Speckmayer (CERN)

2 Need of a global tool that would : provide with the most common MV methods do both the training and evaluation of these methods enable easy computation of systematics Multi-variable analysis widely used in HEP (LEP, BaBar, Belle, D0, MiniBooNE, …) Common Reproaches to Multi-Variable Methods : “Black box methods” In case of correlations cuts are not transparent anymore Training sample may not describe correctly the data Creates no bias, only bad performance Systematics ? Need a control sample Independent & uneasy implementations …

3 “TMVA” means Toolkit for Multivariate Analysis MV Methods available so far : Rectangular cut optimization Correlated Likelihood estimator (PDE) Multi-dimensional likelihood estimator (PDE) Fisher & Mahalanobis discriminant H-Matrix ( χ 2 estimator) Neural Network (2 different implementations) Boosted decision tree Root package written by Andreas Höcker, Kai Voss, Helge Voss, Jörg Stelzer, Peter Speckmayer for the evaluation of MV methods in parallel of an analysis TMVA provide training, testing & evaluation of these methods A dedicated class enables to plug the training results in your favorite analysis

4 Cut Optimization  Scan in signal efficiency for highest background rejection Correlated & de-correlated likelihood  PDE approach, generalization to multi dimension likelihood, Output transformed by an inverse Fermi function (  less peaked) De-correlation possible with the square root of the covariance matrix Fisher discriminant and H-matrix  Classical definitions Neural Network  2 NNs, both multi-layer perceptrons with stochastic learning Clermont Ferrand ANN (used for ALEPH Higgs analysis) TMultiLayerPerceptron (ANN from ROOT) (Boosted) Decision trees  Inspired from MiniBooNE Sequential applications of cuts

5 Each event has a weight W i (= 1 to start) Training Split until minimal #event reached or limit of purity reached Final node = “leaf” : if P>P min “signal leaf” If P<P min “background leaf” Boosting : if a signal event is on a bkg leaf or if a bkg event is on a signal leaf  weight modified Then training re-performed with the new weights (x1000 trees) S/B 52/48 Var1 > x1 Var1 ≤ x1 Var2 > x2Var2 ≤ x2 S/B 4/37 S/B 48/11 S/B 2/10 S/B 46/1 Testing Start from root node, events go through the 1000 boosted trees Each time an event ends on leaf signal or background signal its weight is modified (↔ Neural Net output) (  smoother output than classical discrete output) Cut on the variable that optimizes the separation, based on the purity P (=“node”) Optimization by scanning, Genetic Algorithm soon What is a Boosted Decision Tree ?

6 How to get and use TMVA ?

7 How to download TMVA ? Get a tgz file : from TMVA website then click onhttp://tmva.sourceforge.net Via cvs : cvs –z3 co –P TMVA  Automatic creation of 6 directories src/ Source for TMVA library example/ example of how to use TMVA lib/ TMVA library once compiled reader/ all functionalities to apply MV weights macros/ root macro to display results development/ working & testing directories For your own analysis : > cp example myTMVA  Modify “makefile” for compilation in /myTMVA Download

8 /home cd TMVA /home/TMVA source setup.csh /home/TMVA cd src/ /home/TMVA/src make Include TMVA/lib in PATH Compile the librairy  libTMVA.so How to compile TMVA ? Detailed Steps for the Example

9 How to choose the MV method I want ? … Bool_t Use_Cuts = 1; Bool_t Use_Likelihood = 0; Bool_t Use_LikelihoodD = 0; Bool_t Use_PDERS = 0; Bool_t Use_HMatrix = 0; Bool_t Use_Fisher = 1; Bool_t Use_CFMlpANN = 1; Bool_t Use_TMlpANN = 0; Bool_t Use_BDT_GiniIndex = 0; Bool_t Use_BDT_CrossEntro = 0; Bool_t Use_BDT_SdivStSpB = 0; Bool_t Use_BDT_MisClass = 0; … You just have to “switch” on or off the Booleans ! Here for instance I will compare Cuts, Fisher, Neural Net CFM /home/TMVA/src cd../examples Go to exemple/ directory and open You will find a list of available methods (Booleans) Switch to 1/0 the method you want/don’t want TMVAnalysis.cpp

10 How to point to TMVA the training samples & variables ? In /examples/data : toy_sig.dat bkg_toy.dat How to point the variables (example with 4 variables) : How to point the input ascii files : In TMVAnalysis.cpp Both ascii or root files can be used as input Creation of the factory object

11 How to change training options ? In TMVAnalysis.cpp Training cycles, #hidden layer, #neurons per layer  Description of every options in class BookMethod factory->PreparTrainigAndTestTree( mycut, 2000, 4000 ); #events used :trainingtesting

12 How to I run TMVA ? /home/TMVA/src cd../examples /home/TMVA/examples make /home/TMVA/examples TMVAnalysis “myOutput.root” Name of output root file What does it create ? Some weights files for each trained MV methods in weight/ A root file in main directory with the MV outputs and efficiencies How to look at the results ? Use the nice ROOT macros in directory macros/ /home/TMVA/examples root –l /home/TMVA/examples.L../macros/efficiencies.C /home/TMVA/examples efficiencies("MyOutput.root") Plots created in directory plots/

13 Which ROOT macros are available ? (1) variables.C  Distributions of input variables

14 Which ROOT macros are available ? (2) correlations.C  Colored correlation matrix of input variables Numeric values displayed during TMVA running

15 Which ROOT macros are available ? (3) mvas.C  Outputs of MV methods

16 Which ROOT macros are available ? (4) efficiencies.C  Background rejection vs. Signal efficiency Direct comparison of all MV methods !

17 I have trained the MV method I want … I have the weight files … How to use this MV method in my analysis ? Detailed example is Dedicated class The next slide shows what must be included in your analysis program … Work in progress (being implemented in ROOT) thus possible differences with later version … TMVA/reader/TMVApplication.cpp reader/TMVA_reader.hh

18 #include “TMVA_reader.h” using TMVApp::TMVA_Reader; Void MyAnalysis() { vector inputVars; inputVars.push_back( "var1" ); inputVars.push_back( "var2" ); inputVars.push_back( "var3" ); inputVars.push_back( "var4" ); TMVA_Reader *tmva = new TMVA_Reader( inputVars ); tmva->BookMVA(TMVA_Reader::Fisher, “TMVAnalysis_Fisher.weights"); vector varValues; varValues.push_back( var1 ); varValues.push_back( var2 ); varValues.push_back( var3 ); varValues.push_back( var4 ); double mvaFi = tmva->EvaluateMVA( varValues, TMVA_Reader::Fisher ); delete tmva;} [1] Include the reader class [2] Create an array of the input variables names (here 4var) [3] Create the reader class [4] Read the weights and build the MV tool [5] Create an array with the input variables values [6] Compute the value of the MV, it is the value you will cut on

19 TMVA is already used by several AWG in BaBar Group Dalitz Charmless UK : TMVA Fisher for continuum rejection in the Dalitz-plot analyzes of K s    - and K +    + ( BADs: 1376 and 1512 ). Use of 11 input variables, pictures taken from BAD 1376

20 Group D0h0 : TMVA ClmF NN for continuum rejection in the measurement of the BFs of the color suppressed modes B 0  D 0 h 0 (h 0 = ω, η, η’, ρ, π 0 ) and in the measurement of CKM β angle Use of 4 input variables

21 Measurement of sin(2α) with B  ρπ  Uses Clermont-Ferrand NN to get rid of combinatory background Measurement of CKM angle γ with GLW method (Emmanuel Latour – LLR)  Uses Fisher to get rid of combinatory background Signal = MC signal B  D*K, D*  D0p0, D0  Kp Background = MCs udsc

22 A similar tool has been developed by Ilya Narsky ( StatPatternRecognition ) What to keep in mind about TMVA ? A powerful Multivariate toolkit with 12 different methods (more are coming) User friendly package from training to plots ! Already used in BaBar Comparison possible & easy between the different MV methods C++ & Root functionalities, announced in ROOT version V Have a look at !! Talk by Kai Voss at CERN TMVA Tutorial Physics analysis HN advertisement

23 Back Up Slides

24 Available Options For Every Methods in TMVAnalysis.cpp Rectangular cut optimization Correlated Likelihood estimator (PDE) Multi-dimensional likelihood estimator (PDE) Fisher & Mahalanobis discriminant H-Matrix ( χ 2 estimator) Neural Network (2 different implementations) Boosted decision tree

25 Rectangular cuts Option for each variables - "FMax" : ForceMax (the max cut is fixed to maximum of variable i) - "FMin" : ForceMin (the min cut is fixed to minimum of variable i) - "FSmart": ForceSmart (the min or max cut is fixed to min/max, based on mean value) - Adding "All" to "option_vari", eg, "AllFSmart" will use this option for all variables - if "option_vari" is empty (== ""), no assumptions on cut min/max are made TMVA method Method of cut - "MC" : Monte Carlo optimization (recommended) - "FitSel": Minuit Fit: "Fit_Migrad" or "Fit_Simplex" - "FitPDF": PDF-based: only useful for uncorrelated input variables factory->BookMethod( "MethodCuts", “Method : nBin : OptionVar1: … :OptionVarn" ); # bins in the hist of efficiency S/B

26 Likelihood factory->BookMethod( "MethodLikelihood", “TypeOfSpline : NbSmooth : NbBin : Decorr"); TMVA method Option for decorrelation or not - “NoDecorr” – do not use square-root-matrix to decorrelate variable space - “Decorr” – decorrelate variable space Which spline is used for smoothing the pdfs “Splinei” [i=1,2,3,5] How often the input histos are smoothed Average num of events per PDF bin to trigger warning

27 Fisher Discriminant and H matrix factory->BookMethod( "MethodFisher", "Fisher" ); TMVA method Which method -“Fisher” - “Mahalanobis” (another definition of distance) factory->BookMethod( "MethodHMatrix" ); TMVA method

28 Artificial Neural Network factory->BookMethod( “WhichANN”, “NbCycles:NeuronsL1:NeuronsL2:…:NeuronsLn" ); Which type of NN - “MethodCFMlpANN” Clermond Ferrand NN, used for Higgs search in ALEPH - “MethodTMlpANN” ROOT NN Number of training cycles Number of neurons in each layer The 1st layer has necessarily as many neurons as input variables

29 Boosted Decision Trees factory->BookMethod( "MethodBDT", “nTree : BoostType : SeparationType : nEvtMin : MaxNodePurity : nCuts”); TMVA method Method for evaluating the misclassification - GiniIndex - CrossEntropy - SdivSqrtSplusB - MisClassificationError Number or trees Method of boosting - AdaBoost - EpsilonBoost Minimum Number of events in a node (leaf criteria) Higher bound for leave or intermediate node Number of steps in the optimization of the cut for a node