Multivariate Data Analysis with TMVA4 Jan Therhaag ( * ) (University of Bonn) ICCMSE09 Rhodes, 29. September 2009 ( * ) On behalf of the present core developer.

Slides:



Advertisements
Similar presentations
HEP Data Mining with TMVA
Advertisements

Statistical Methods for Data Analysis Multivariate discriminators with TMVA Luca Lista INFN Napoli.
S.Towers TerraFerMA TerraFerMA A Suite of Multivariate Analysis tools Sherry Towers SUNY-SB Version 1.0 has been released! useable by anyone with access.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
An Overview of Machine Learning
TMVA ― Status and Developments
TMVA – Toolkit for Multivariate Analysis
Lecture 14 – Neural Networks
Introduction to Statistics and Machine Learning 1 How do we: understandunderstand interpretinterpret our measurements How do we get the data for our measurements.
Sparse vs. Ensemble Approaches to Supervised Learning
DESY Computing Seminar Hamburg
Optimization of Signal Significance by Bagging Decision Trees Ilya Narsky, Caltech presented by Harrison Prosper.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
MACHINE LEARNING 12. Multilayer Perceptrons. Neural Networks Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Sparse vs. Ensemble Approaches to Supervised Learning
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Jens Zimmermann, MPI für Physik München, ACAT 2005 Zeuthen1 Backups Jens Zimmermann Max-Planck-Institut für Physik, München Forschungszentrum.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Walter Hop Web-shop Order Prediction Using Machine Learning Master’s Thesis Computational Economics.
July 11, 2001Daniel Whiteson Support Vector Machines: Get more Higgs out of your data Daniel Whiteson UC Berkeley.
Comparing the Parallel Automatic Composition of Inductive Applications with Stacking Methods Hidenao Abe & Takahira Yamaguchi Shizuoka University, JAPAN.
TMVA Andreas Höcker (CERN) CERN meeting, Oct 6, 2006 Toolkit for Multivariate Data Analysis.
G. Cowan Lectures on Statistical Data Analysis Lecture 7 page 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem 2Random variables and.
Machine Learning with TMVA A ROOT based Tool for Multivariate Data Analysis PANDA Computing Workshop Groningen The TMVA developer team: The TMVA.
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
1 Helge Voss Nikhef 23 rd - 27 th April 2007TMVA Toolkit for Multivariate Data Analysis: ACAT 2007 TMVA Toolkit for Multivariate Data Analysis with ROOT.
Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
TMVA Jörg Stelzer: Machine Learning withCHEP 2007, Victoria, Sep 5 Machine Learning Techniques for HEP Data Analysis with TMVA Toolkit for Multivariate.
Comparison of Bayesian Neural Networks with TMVA classifiers Richa Sharma, Vipin Bhatnagar Panjab University, Chandigarh India-CMS March, 2009 Meeting,
1 / 41 LAL Seminar, June 21, 2007A. Hoecker: Machine Learning with TMVA Machine Learning Techniques for HEP Data Analysis with TMVA Andreas Hoecker ( *
1 Helge Voss Genvea 28 th March 2007ROOT 2007: TMVA Toolkit for MultiVariate Analysis TMVA A Toolkit for MultiVariate Data Analysis with ROOT Andreas Höcker.
1 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA Machine Learning Techniques for HEP Data Analysis with TMVA Andreas Hoecker.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
1 Krakow Seminar, Feb 27, 2007A. Hoecker: Data Mining with TMVA HEP Data Mining with TMVA  ToolKit for Multivariate Analysis with ROOT  Andreas Hoecker.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב
1 Teilchenseminar, April 26, 2007K. Voss: TMVA toolkit TMVA  ToolKit for Multivariate Analysis with ROOT  Kai Voss ( * ) Teilchenseminar, Bonn, April.
Chapter 6. Classification and Prediction Classification by decision tree induction Bayesian classification Rule-based classification Classification by.
Perceptrons Michael J. Watts
1 / 44 LPNHE Seminar, June 20, 2007A. Hoecker: Machine Learning with TMVA Machine Learning Techniques for HEP Data Analysis with TMVA Andreas Hoecker (
1 CATPhys meeting, CERN, Mar 13, 2006 Andreas Höcker – TMVA – toolkit for parallel multivariate data analysis – Andreas Höcker (ATLAS), Helge Voss (LHCb),
Multivariate Classifiers or “Machine Learning” in TMVA
C. Kiesling, MPI for Physics, Munich - ACAT03 Workshop, KEK, Japan, Dec Jens Zimmermann, Christian Kiesling Max-Planck-Institut für Physik, München.
1/21 ACAT 2011Eckhard von Toerne 06.Sept 2011 Status of TMVA the Toolkit for MultiVariate Analysis Eckhard von Toerne (University of Bonn) For the TMVA.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Neural Network Analysis of Dimuon Data within CMS Shannon Massey University of Notre Dame Shannon Massey1.
Multivariate Data Analysis with TMVA Andreas Hoecker ( * ) (CERN) SOS, Strasbourg, France, July 4, 2008 ( * ) On behalf of the present core team: A. Hoecker,
One framework for most common MVA-techniques, available in ROOT Have a common platform/interface for all MVA classification and regression-methods: Have.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Helge VossAdvanced Scientific Computing Workshop ETH Multivariate Methods of data analysis Helge Voss Advanced Scientific Computing Workshop ETH.
Helge VossINFN School of Statistics Multivariate Discriminators Helge Voss INFN School of Statistics 2015 Ischia (Napoli, Italy) MAX-PLANCK-INSTITUT.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Multivariate Methods of
Multivariate Data Analysis with TMVA
COMP61011 : Machine Learning Ensemble Models
Christoph Rosemann and Helge Voss DESY FLC
Multivariate Data Analysis
Machine Learning Basics
Machine Learning Ensemble Learning: Voting, Boosting(Adaboost)
Computing and Statistical Data Analysis Stat 5: Multivariate Methods
Toolkit for Multivariate Data Analysis Helge Voss, MPI-K Heidelberg
Statistical Methods for Data Analysis Multivariate discriminators with TMVA Luca Lista INFN Napoli.
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
What is Artificial Intelligence?
Presentation transcript:

Multivariate Data Analysis with TMVA4 Jan Therhaag ( * ) (University of Bonn) ICCMSE09 Rhodes, 29. September 2009 ( * ) On behalf of the present core developer team: A. Hoecker, P. Speckmayer, J. Stelzer, J. Therhaag, E. von Toerne, H. Voss On the web:

22 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 2 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Classification and Regression Tasks in HEP Can a machine learn that for you? Yes, with TMVA! Classification of signal/background: How to find the best decision boundary? Regression: How to determine the correct model and its parameters?

33 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 3 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, TMVA provides methods for multivariate analysis… …based on supervised learning… …and so much more: –a common interface for all MVA techniques –a common interface for classification and regression –easy training and testing of all methods on the same datasets consistent evaluation and comparison common data preprocessing –a complete user analysis framework and examples –embedding in ROOT –creation of standalone C++ classes (ROOT independent) –an understandable Users Guide What is TMVA? Multivariate classifiers/regressors condense (correlated) multi-variable input information into a single scalar output variable Supervised learning means learning by example: the program extracts patterns from training data

44 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 4 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Overview TMVA Classifiers / Regressors The TMVA Workflow

55 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 5 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, The TMVA Classifiers Cuts, Fisher, Likelihood, Functional Discriminant, kNN, Neural Networks, Support Vector Machine, Boosted Decision Trees, Rule Fit…

66 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 6 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Linear Discriminant (Fisher) Fisher‘s discriminant is a linear model which projects the data on the (hyper)plane of best separation Advantages: –easy to understand –robust, fast Disadvantages: –not very flexible –poor performace in complex settings PerformanceSpeedRobustnessCurse of Dim. TransparencyRegression No/linear correlations Nonlinear correlations TrainingResponseOvertrainingWeak input vars 1Dmulti D  

77 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 7 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, PerformanceSpeedRobustnessCurse of Dim. TransparencyRegression No/linear correlations Nonlinear correlations TrainingResponseOvertrainingWeak input vars 1Dmulti D  K-Nearest-Neighbour / Multidimensional PDE Idea: Estimate the multidimensional probability densities by counting the events of each class in a predefined or adaptive volume Advantages: –reconstruction of PDFs possible Disadvantages: –„curse of dimensionality“

88 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 8 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, PerformanceSpeedRobustnessCurse of Dim. TransparencyRegression No/linear correlations Nonlinear correlations TrainingResponseOvertrainingWeak input vars 1Dmulti D   Artificial Neural Networks 1 i N 1 input layerk hidden layers1 ouput layer 1 j M1M MkMk 2 output classes (signal and background) N var discriminating input variables (“Activation” function) with: Feed-forward Multilayer Perceptron Modelling of arbitrary nonlinear functions as a nonlinear combination of simple „neuron activation functions“ Advantages: –very flexible, no assumption about the function necessary Disadvantages: –„black box“ –needs tuning –seed dependent

99 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 9 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Boosted Decision Trees (AdaBoost, Bagging, GradBoost) PerformanceSpeedRobustnessCurse of Dim. TransparencyRegression No/linear correlations Nonlinear correlations TrainingResponseOvertrainingWeak input vars 1Dmulti D   A decision tree is a set of binary splits of the input space Grow a forest of decision trees and determine the event class/target by majority vote Advantages: –ignores weak variables –works out of the box Disadvantages: –vulnerable to overtraining

10 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 10 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Metaclassifiers – Category Classifier and Boosting The category classifier is custom-made for HEP –Use different classifiers for different phase space regions and combine them into a single output TMVA supports boosting for all classifiers –Use a collection of “weak learners“ to improve their performace (boosted Fisher, boosted neural nets with few neurons each…)

11 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 11 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Using TMVA

12 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 12 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, The TMVA Workflow – Training, Testing, Application Training: –Classification: Learn the features of the different event classes from a sample with known signal/background composition –Regression: Learn the functional dependence between input variables and targets Testing: –Evaluate the performance of the trained classifier/regressor on an independent test sample –Compare different methods Application: –Apply the classifier/regressor to real data

13 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 13 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Selection of variables TMVA supports… …ROOT TTree or ASCII files as input …any combination or function of available input variables (including TMath functions) …independent signal/bg preselection cuts ….any input variable as individual event weight …various methods to split data into training/test trees

14 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 14 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Note that in cases with non-Gaussian distributions and/or nonlinear correlations decorrelation may do more harm than any good Data Preprocessing Decorrelation may improve the performance of simple classifiers… … but complex correlations call for more sophisticated classifiers

15 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 15 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, void TMVAnalysis( ) { TFile* outputFile = TFile::Open( "TMVA.root", "RECREATE" ); TMVA::Factory *factory = new TMVA::Factory( "MVAnalysis", outputFile,"!V"); TFile *input = TFile::Open("tmva_example.root"); factory->AddVariable("var1+var2", 'F'); factory->AddVariable("var1+var2", 'F'); //factory->AddTarget("tarval", 'F'); factory->AddSignalTree ( (TTree*)input->Get("TreeS"), 1.0 ); factory->AddBackgroundTree ( (TTree*)input->Get("TreeB"), 1.0 ); //factory->AddRegressionTree ( (TTree*)input->Get("regTree"), 1.0 ); factory->BookMethod( TMVA::Types::kLikelihood, "Likelihood", "!V:!TransformOutput:Spline=2:NSmooth=5:NAvEvtPerBin=50" ); factory->BookMethod( TMVA::Types::kMLP, "MLP", "!V:NCycles=200:HiddenLayers=N+1,N:TestRate=5" ); factory->TrainAllMethods(); // factory->TrainAllMethodsForRegression(); factory->TestAllMethods(); factory->EvaluateAllMethods(); outputFile->Close(); delete factory; } A complete TMVA training/testing session Create Factory, the only object the user needs! Add variables/ targets Initialize Trees Book MVA methods and transformations Train, test and evaluate

16 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 16 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Evaluation of classification Check for overtraining, compare classifier performance and select the optimal cut value for every classifier

17 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 17 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Evaluation of regression Compare deviation from target for different methods

18 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 18 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Functionality in preparation... Multiclass classification Automatic classifier tuning via cross validation Advanced method combination Bayesian classifiers …. Train Test Classifier1Classifier2Classifier3 Combined Classifier

19 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 19 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Conclusion Acknowledgments: The fast development of TMVA would not have been possible without the contribution and feedback from many developers and users to whom we are indebted. A full list of contributors is available in the TMVA User’s Guide ( ). TMVA is a powerful toolkit for both classification and regression in HEP All MVA methods can be accessed through a common interface, data preprocessing can be applied TMVA supports the user in the MVA selection process via a collection of evaluation scripts TMVA is open source software and can be downloaded from

20 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 20 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Backup

21 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 21 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Projective likelihood (naive bayesian classifier) PerformanceSpeedRobustnessCurse of Dim. TransparencyRegression No/linear correlations Nonlinear correlations TrainingResponseOvertrainingWeak input vars 1Dmulti D   discriminating variables Species: signal, background types Likelihood ratio for event i event PDFs Idea: –create PDFs from variable distributions, then calculate likelihood ratio Advantages: –optimal when no correlations Disadvantages: –correlations not taken into account

22 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 22 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Support Vector Machines PerformanceSpeedRobustnessCurse of Dim. TransparencyRegression No/linear correlations Nonlinear correlations TrainingResponseOvertrainingWeak input vars 1Dmulti D  support vectors x1x1 x2x2 margin Separable data optimal hyperplane x1x1 x3x3 x1x1 x2x2  (x 1,x 2 ) Non-separable data Linearly separable data: –find separating hyperplane (Fisher) Non-separable data: –fransform variables into a high dimensional space where linear separation is possible Advantages: –flexibility combined with robustness Disadvantages: –slow training –complex algorithm

23 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 23 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, TMVA supports individual transformations for each method  transformations can be chained  transformations are calculated on the fly when the event is read  TMVA supports decorrelation, Gaussianization and normalization Data Preprocessing original SQRT derorr. PCA derorr.

24 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 24 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Decorrelation original SQRT derorr. PCA derorr. Decorrelations can drastically improve the performance of the simple classifiers But complex correlations call for more sophisticated classifiers…

25 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 25 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Gaussianisation Decorrelation may improve by pre-Gaussianisation of the variables –First step: obtain uniform distribution through Rarity transformation –Second step: make Gaussian via inverse error function Rarity transform of variable kMeasured valuePDF of variable k

26 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 26 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Evaluation – The TMVA GUI TMVA is not only a collection of classifiers, but also a complete MVA framework! –a collection of evaluation macros can be accessed via the convenient GUI Plot all signal (S) and background (B) input variables with and without pre-processing Correlation scatters and linear coefficients for S & B Classifier outputs (S & B) for test and training samples (spot overtraining) Classifier significance with optimal cuts B rejection versus S efficiency (ROC) Classifier-specific plots: Likelihood reference distributions Classifier PDFs (for probability output and Rarity) Network architecture, weights and convergence Visualization of decision trees

27 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 27 J. Therhaag ― Multivariate Data Analysis with TMVA4Rhodes, Classifier specific evaluation average no. of nodes before/after pruning: 4193 / 968