1 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction TMVA Workshop – Introduction Andreas Hoecker (CERN) TMVA Workshop, 21 January 2011, CERN, Switzerland.

Slides:



Advertisements
Similar presentations
Statistical Issues at LHCb Yuehong Xie University of Edinburgh (on behalf of the LHCb Collaboration) PHYSTAT-LHC Workshop CERN, Geneva June 27-29, 2007.
Advertisements

S.Towers TerraFerMA TerraFerMA A Suite of Multivariate Analysis tools Sherry Towers SUNY-SB Version 1.0 has been released! useable by anyone with access.
Machine learning continued Image source:
Matthew Schwartz Harvard University March 24, Boost 2011.
Lecture 3 Nonparametric density estimation and classification
TMVA – Toolkit for Multivariate Analysis
Chapter 4: Linear Models for Classification
14 Sept 2004 D.Dedovich Tau041 Measurement of Tau hadronic branching ratios in DELPHI experiment at LEP Dima Dedovich (Dubna) DELPHI Collaboration E.Phys.J.
What is Statistical Modeling
Summary of Results and Projected Sensitivity The Lonesome Top Quark Aran Garcia-Bellido, University of Washington Single Top Quark Production By observing.
Introduction to Statistics and Machine Learning
Bayesian Decision Theory Chapter 2 (Duda et al.) – Sections
x – independent variable (input)
8. Statistical tests 8.1 Hypotheses K. Desch – Statistical methods of data analysis SS10 Frequent problem: Decision making based on statistical information.
8. Hypotheses 8.4 Two more things K. Desch – Statistical methods of data analysis SS10 Inclusion of systematic errors LHR methods needs a prediction (from.
Particle Identification in the NA48 Experiment Using Neural Networks L. Litov University of Sofia.
Statistical Decision Theory, Bayes Classifier
Optimization of Signal Significance by Bagging Decision Trees Ilya Narsky, Caltech presented by Harrison Prosper.
G. Cowan 2007 CERN Summer Student Lectures on Statistics1 Introduction to Statistics − Day 3 Lecture 1 Probability Random variables, probability densities,
Machine Learning CMPT 726 Simon Fraser University
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Ensemble Learning (2), Tree and Forest
METU Informatics Institute Min 720 Pattern Classification with Bio-Medical Applications PART 2: Statistical Pattern Classification: Optimal Classification.
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Inference for the mean vector. Univariate Inference Let x 1, x 2, …, x n denote a sample of n from the normal distribution with mean  and variance 
July 11, 2001Daniel Whiteson Support Vector Machines: Get more Higgs out of your data Daniel Whiteson UC Berkeley.
Principles of Pattern Recognition
G. Cowan Lectures on Statistical Data Analysis Lecture 7 page 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem 2Random variables and.
11 Top Workshop, LPSC, Oct 18–20, 2007A. Hoecker: Multivariate Analysis with TMVA 1 CERN, April 14., 2010P. Speckmayer ― Multivariate Data Analysis with.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
G. Cowan Statistical Methods in Particle Physics1 Statistical Methods in Particle Physics Day 3: Multivariate Methods (II) 清华大学高能物理研究中心 2010 年 4 月 12—16.
TMVA Jörg Stelzer: Machine Learning withCHEP 2007, Victoria, Sep 5 Machine Learning Techniques for HEP Data Analysis with TMVA Toolkit for Multivariate.
Comparison of Bayesian Neural Networks with TMVA classifiers Richa Sharma, Vipin Bhatnagar Panjab University, Chandigarh India-CMS March, 2009 Meeting,
Matthew Schwartz Harvard University with J. Gallicchio, PRL, 105:022001,2010 (superstructure) with K. Black, J. Gallicchio, J. Huth, M. Kagan and B. Tweedie.
Multivariate Data Analysis with TMVA Andreas Hoecker ( * ) (CERN) Seminar at Edinburgh, Scotland, Dec 5, 2008 ( * ) On behalf of the present core developer.
Statistics In HEP Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Training of Boosted DecisionTrees Helge Voss (MPI–K, Heidelberg) MVA Workshop, CERN, July 10, 2009.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
1 Krakow Seminar, Feb 27, 2007A. Hoecker: Data Mining with TMVA HEP Data Mining with TMVA  ToolKit for Multivariate Analysis with ROOT  Andreas Hoecker.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Particle Identification. Particle identification: an important task for nuclear and particle physics Usually it requires the combination of informations.
Kalanand Mishra BaBar Coll. Meeting February, /8 Development of New Kaon Selectors Kalanand Mishra University of Cincinnati.
G. Cowan IDPASC School of Flavour Physics, Valencia, 2-7 May 2013 / Statistical Analysis Tools 1 Statistical Analysis Tools for Particle Physics IDPASC.
1 Introduction to Statistics − Day 2 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
Multivariate Classifiers or “Machine Learning” in TMVA
1/21 ACAT 2011Eckhard von Toerne 06.Sept 2011 Status of TMVA the Toolkit for MultiVariate Analysis Eckhard von Toerne (University of Bonn) For the TMVA.
G. Cowan Lectures on Statistical Data Analysis Lecture 6 page 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Statistical Data Analysis: Lecture 5 1Probability, Bayes’ theorem 2Random variables and.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Neural Network Analysis of Dimuon Data within CMS Shannon Massey University of Notre Dame Shannon Massey1.
Overfitting, Bias/Variance tradeoff. 2 Content of the presentation Bias and variance definitions Parameters that influence bias and variance Bias and.
One framework for most common MVA-techniques, available in ROOT Have a common platform/interface for all MVA classification and regression-methods: Have.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Multivariate Data Analysis with TMVA4 Jan Therhaag ( * ) (University of Bonn) ICCMSE09 Rhodes, 29. September 2009 ( * ) On behalf of the present core developer.
Helge VossAdvanced Scientific Computing Workshop ETH Multivariate Methods of data analysis Helge Voss Advanced Scientific Computing Workshop ETH.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Multivariate Methods of
Multivariate Methods of
Comment on Event Quality Variables for Multivariate Analyses
Multi-dimensional likelihood
ISTEP 2016 Final Project— Project on
Computing and Statistical Data Analysis Stat 5: Multivariate Methods
Prospects for quarkonium studies at LHCb
TRISEP 2016 / Statistics Lecture 2
Learning From Observed Data
Presentation transcript:

1 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction TMVA Workshop – Introduction Andreas Hoecker (CERN) TMVA Workshop, 21 January 2011, CERN, Switzerland

2 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Goals of the Workshop Introduction to multivariate classification and regression with TMVA Pedagogical talks on various multivariate methods (morning) Talks from users (14:00) Tutorial (15:30)

3 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction TMVA ROOT: is the analysis framework used by most (HEP)-physicists Idea: rather than just implementing new MVA techniques and making them available in ROOT: ﹣ Have one common platform / interface for all MVA methods ﹣ Have common data pre-processing capabilities ﹣ Train and test all classifiers on same data sample and evaluate consistently ﹣ Provide common analysis (ROOT scripts) and application framework ﹣ Provide access with and without ROOT, through macros, C++ executables or python

4 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction TMVA TMVA started in 2006 on the Sourceforge development platformSourceforge 6 core developers, 21 contributors so far TMVA is written in C++, and relies on ROOT functionality Since ROOT 5.15 / TMVA v3.7.2 TMVA is part of ROOT, and developed directly in ROOT SVNROOT SVN ﹣ Continue to maintain primary tmva-users mailing list on Sourceforgetmva-users ﹣ New TMVA versions also published as downloadable tgz files on Sourceforge ﹣ For bug reports, use ROOT SavannahROOT Savannah

5 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Simulated Higgs Event in CMS Higgs event in an LHC proton–proton collision at high luminosity (together with ~24 other inelastic events) 7 TeV LHC 2010 Such events occur only in a tiny fraction of the proton-proton collisions O(10 −10 )

6 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Event Classification in HEP Most HEP analyses require discrimination of signal from background: ﹣ Event level (Higgs searches, …) ﹣ Cone level (Tau-vs-jet reconstruction, …) ﹣ Track level (particle identification, …) ﹣ Object level (flavour tagging, …) ﹣ Parameter estimation (significance, mass, CP violation in B system, …) The multivariate input information used for this has various sources ﹣ Kinematic variables (masses, momenta, decay angles, …) ﹣ Event properties (jet/lepton multiplicity, sum of charges, …) ﹣ Event shape (sphericity, Fox-Wolfram moments, …) ﹣ Detector response (silicon hits, dE/dx, Cherenkov angle, shower profiles, muon hits, …) Traditionally few powerful input variables were combined. New methods allow to use up to 100 and more variables w/o loss of classification power e.g. MiniBooNE: NIMA 543 (2005), or D0 single top: Phys.Rev. D78, (2008)

7 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Event Classification Suppose data sample with two types of events: H 0, H 1 ﹣ We have found discriminating input variables x 1, x 2, … ﹣ What decision boundary should we use to select events of type H 1 ? Linear boundary? A nonlinear one? Rectangular cuts? Low variance (stable), high bias methodsHigh variance, small bias methods

8 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Event Classification Suppose data sample with two types of events: H 0, H 1 ﹣ We have found discriminating input variables x 1, x 2, … ﹣ What decision boundary should we use to select events of type H 1 ? Linear boundary? A nonlinear one? Rectangular cuts? How can we decide this in an optimal way ?  Let the machine learn it !

9 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Parameter Regression How to estimate a “functional behaviour” from a set of measurements? ﹣ Energy deposit in a the calorimeter, distance between overlapping photons, … ﹣ Entry location of a particle in the calorimeter or on a silicon pad, … x f(x) x x Linear function ? A non-linear one ? Constant ? Looks trivial? What if we have many input variables?

10 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Multivariate Event Classification

11 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Each event, Signal or Background, has “D” measured variables.  D “feature space” Find a mapping from D-dimensional input-observable = ”feature” space to one dimensional output  class labels

12 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Each event, Signal or Background, has “D” measured variables.  D “feature space” y(x)y(x)  Most general form y = y(x); x in  D x = {x 1,….,x D }: input variables y(x): R n  R: Plotting the resulting y(x) values: Find a mapping from D-dimensional input-observable = ”feature” space to one dimensional output  class labels

13 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Each event, Signal or Background, has “D” measured variables.  D “feature space”  y(x): R n  R: y(x)y(x) y(x): “test statistic” in D-dimensional space of input variables Distributions of y(x): PDF S (y) and PDF B (y) Overlap of PDF S (y) and PDF B (y) affects separation power, purity > cut: signal = cut: decision boundary < cut: background y(x): Used to set the selection cut ! y(x) = const: surface defining the decision boundary  Efficiency and purity y(B)  0, y(S)  1

14 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Multi-Class Classification Binary classification: two classes, “signal” and “background” SignalBackground

15 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Class 1 Class 2 Class 3 Class 5 Class 6 Class 4 Multi-Class Classification Multi-class classification – natural extension for many classifiers

16 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction P(Class=C|x) (or simply P(C|x)) : probability that the event class is of type C, given the measured observables x = {x 1,….,x D }  y(x) Prior probability to observe an event of “class C”, i.e., the relative abundance of “signal” versus “background” Overall probability density to observe the actual measurement y(x), i.e., Probability density distribution according to the measurements x and the given mapping function Posterior probability Event Classification

17 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction AND Minimum error in misclassification if C chosen such that it has maximum P(C|y)  To select S(ignal) over B(ackground), place decision on: Likelihood ratio as discriminating function y(x) Prior odds ratio of choosing a signal event (relative probability of signal vs. bkg) “c” determines efficiency and purity x = {x 1,….,x D }: measured observables y = y(x) [ Or any monotonic function of P(S|y) / P(B|y) ] Posterior odds ratio Bayes Optimal Classification

18 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Trying to select signal events: (i.e. try to disprove the null-hypothesis stating it were “only” a background event) Type-2 error: Fail to identify an event from Class C as such (reject a hypothesis although it would have been true) (fail to reject the null-hypothesis/accept null hypothesis although it is false)  loss of efficiency (in selecting signal events) Decide to treat an event as “Signal” or “Background” Signal Back- ground Signal Type-2 error Back- ground Type-1 error Accept as: Truly is: Type-1 error: Classify event as Class C even though it is not (accept a hypothesis although it is not true) (reject the null-hypothesis although it would have been the correct one)  loss of purity (in the selection of signal events) Any Decision Involves a Risk Significance α:Type-1 error rate: (=p-value): α = background selection “efficiency” Size β: Type-2 error rate: Power: 1  β = signal selection efficiency should be small ! “A”: region where event is called signal

19 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Neyman-Peason: The Likelihood ratio used as “selection criterion” y(x) gives for each selection efficiency the best possible background rejection. i.e. it maximises the area under the “Receiver Operation Characteristics” (ROC) curve Neyman-Peason: The Likelihood ratio used as “selection criterion” y(x) gives for each selection efficiency the best possible background rejection. i.e. it maximises the area under the “Receiver Operation Characteristics” (ROC) curve  Varying y(x) > “cut” moves the working point (efficiency and purity) along the ROC curve How to choose “cut”?  need to know prior probabilities (S, B abundances) ﹣ Measurement of signal cross section: maximum of S/√(S+B) or equiv. √(  ·p) ﹣ Discovery of a signal :maximum of S/√(B) ﹣ Precision measurement:high purity (p) ﹣ Trigger selection:high efficiency (  (sometimes high background rejection)  backgr.  signal random guessing good classification better classification “limit” in ROC curve given by likelihood ratio Type-1 error small Type-2 error large Type-1 error large Type-2 error small Neyman-Pearson Lemma

20 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Unfortunately, the true probability densities functions are typically unknown:  Neyman-Pearson’s lemma doesn’t really help us…  Supervised (machine) learning * Hyperplane in the strict sense goes through the origin. Here is meant an “affine set” to be precise. Use MC simulation, or more generally: set of known (already classified) “events” Use these “training” events to: Try to estimate the functional form of P(x|C) from which the likelihood ratio can be obtained e.g. D-dimensional histogram, Kernel densitiy estimators, MC-based matrix-element methods, … Find a “discrimination function” y(x) and corresponding decision boundary (i.e. hyperplane* in the “feature space”: y(x) = const) that optimally separates signal from background e.g. Linear Discriminator, Neural Networks, Boosted Decision, … Realistic Event Classification

21 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Unfortunately, the true probability densities functions are typically unknown:  Neyman-Pearson’s lemma doesn’t really help us…  Supervised (machine) learning * Hyperplane in the strict sense goes through the origin. Here is meant an “affine set” to be precise. Use MC simulation, or more generally: set of known (already classified) “events” Use these “training” events to: Try to estimate the functional form of P(x|C) from which the likelihood ratio can be obtained e.g. D-dimensional histogram, Kernel densitiy estimators, MC-based matrix-element methods, … Find a “discrimination function” y(x) and corresponding decision boundary (i.e. hyperplane* in the “feature space”: y(x) = const) that optimally separates signal from background e.g. Linear Discriminator, Neural Networks, … Realistic Event Classification

22 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Multivariate Analysis Methods in TMVA Examples for classifiers and regression methods – Rectangular cut optimisation – Projective and multidimensional likelihood estimator – k-Nearest Neighbor algorithm – Fisher and H-Matrix discriminants – Function discriminants – Artificial neural networks – Boosted decision trees – RuleFit – Support Vector Machine Preprocessing methods: – Decorrelation, Principal Value Decomposition, Gaussianisation Examples for synthesis methods: – Boosting, Categorisation (valid for all methods, and their combinations) Joerg Jan Helge Eckhard Peter

23 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction We have a Users Guide ! Available on TMVA Users Guide 142pp, incl. code examples arXiv physics/

24 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction  T MVA tutorial U s i n g T M V A A typical TMVA analysis consists of two main steps: 1.Training phase: training, testing and evaluation of classifiers using data samples with known signal and background composition 2.Application phase: using selected trained classifiers to classify unknown data samples

25 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction  T MVA tutorial Code Flow for Training and Application

26 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Can be ROOT scripts, C++ executables or python scripts (via PyROOT), or any other high-level language that interfaces with ROOT  T MVA tutorial Code Flow for Training and Application

27 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction  T MVA tutorial Strong Methods need Strong Evaluation A lot of evaluation information is already provided in the logging output of the training Simple GUIs provide access to evaluation plots and tools for single and multi-class classification and regression

28 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction  T MVA tutorial Involved Methods need Thorough Evaluation

29 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction  T MVA tutorial Involved Methods need Thorough Evaluation

30 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction  T MVA tutorial Involved Methods need Thorough Evaluation average no. of nodes before/after pruning: 4193 / 968

31 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Join the tutorial this afternoon !!  T MVA tutorial

32 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Two Wishes to Users :  T MVA tutorial

33 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction ﹣ A detector element may only exist in the barrel, but not in the endcaps ﹣ A variable may have different distributions in barrel, overlap, endcap regions Ignoring this dependence may reduce performance and creates correlations between variables, which must be learned by the classifier ﹣ Classifiers such as the projective likelihood, which do not account for correlations, significantly loose performance if the sub-populations are not separated Categorisation means splitting the data sample into categories defining disjoint data samples with the following (idealised) properties: ﹣ Events belonging to the same category are statistically indistinguishable ﹣ Events belonging to different categories have different properties Multivariate training samples often have distinct sub-populations of data

34 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction MethodCategory is Your Friend ! It provides fully transparent support for categorisation of your input data, applicable to any TMVA method  See Peter’s talk later today

35 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction The HEP community has already a lot of experience with MVA classification In particular for rare events searches, O(all) mature experiments use it They increase the experiment’s sensitivity, and may reduce systematic errors due to a smaller background component MVAs are not black boxes, but (possibly involved) R n  R mapping functions Should acquire more experience in HEP with multivariate regression Our calibration schemes are often still quite simple: linear or simple functions, look-up-table based, mostly depending on few variables (e.g., η, p T ) Non-linear multivariate regression may significantly boost calibration and corrections applied, in particular if it is possible to train from data Available since TMVA 4 for: LD, FDA, k-NN, PDERS, PDEFoam, MLP, BDT

36 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction S t a t u s & O u t l o o k 2-class classification supported by all methods Multi-class classification supported by MLP (NN), BDTG, FDA Single-target regression: PDE-RS, PDE-Foam, k-NN, LD, FDA, MLP, BDT Multi-target regression: PDE-Foam, k-NN, MLP All methods support categorised classification and generalised boosting Priority on to-do list for future releases ﹣ Automatic self-optimisation of parameter settings for all methods ﹣ For this: full support of cross validation ﹣ Increase support of multi-dimensional classification and regression ﹣ Individual improvements of methods (see, eg, MLP talk by Jan) ﹣ Introduction of unsupervised learning

37 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Several similar data mining efforts with rising importance in most fields of science and industry Important for HEP: ﹣ Parallelised MVA training and evaluation pioneered by Cornelius package (BABAR) ﹣ Also frequently used: StatPatternRecognition package by I. Narsky (Cal Tech) ﹣ Many implementations of individual classifiers exist TMVA is open source software Use & redistribution of source permitted according to terms in BSD licenseBSD license Contributed to TMVA have: Andreas Hoecker (CERN, Switzerland), Jörg Stelzer (CERN, Switzerland), Peter Speckmayer (CERN, Switzerland), Jan Therhaag (Universität Bonn, Germany), Eckhard von Toerne (Universität Bonn, Germany), Helge Voss (MPI für Kernphysik Heidelberg, Germany), Moritz Backes (Geneva University, Switzerland), Tancredi Carli (CERN, Switzerland), Asen Christov (Universität Freiburg, Germany), Or Cohen (CERN, Switzerland and Weizmann, Israel), Krzysztof Danielowski (IFJ and AGH/UJ, Krakow, Poland), Dominik Dannheim (CERN, Switzerland), Sophie Henrot-Versille (LAL Orsay, France), Matthew Jachowski (Stanford University, USA), Kamil Kraszewski (IFJ and AGH/UJ, Krakow, Poland), Attila Krasznahorkay Jr. (CERN, Switzerland, and Manchester U., UK), Maciej Kruk (IFJ and AGH/UJ, Krakow, Poland), Yair Mahalalel (Tel Aviv University, Israel), Rustem Ospanov (University of Texas, USA), Xavier Prudent (LAPP Annecy, France), Arnaud Robert (LPNHE Paris, France), Doug Schouten (S. Fraser U., Canada), Fredrik Tegenfeldt (Iowa University, USA, until Aug 2007), Alexander Voigt (CERN, Switzerland), Kai Voss (University of Victoria, Canada), Marcin Wolter (IFJ PAN Krakow, Poland), Andrzej Zemla (IFJ PAN Krakow, Poland). Copyrights & Credits

38 TMVA Workshop, 21 Jan 2011 Andreas Hoecker – Introduction Software packages for Multivariate Data Analysis/Classification: Individual classifier software: e.g. “JETNET” C.Peterson, T. Rognvaldsson, L.Loennblad, many, many other packages! “All inclusive” packages StatPatternRecognition: I.Narsky, arXiv: physics/ TMVA: Hoecker, Speckmayer, Stelzer, Therhaag, von Toerne, Voss, arXiv: physics/ or every ROOT distribution WEKA: Huge data analysis library available in “R”: Literature: T. Hastie, R. Tibshirani, J. Friedman, “The Elements of Statistical Learning”, Springer 2001 C.M. Bishop, “Pattern Recognition and Machine Learning”, Springer 2006 Conferences: PHYSTAT, ACAT,… A few References