GLAST/LAT analysis group 31Jan05 - T. Burnett 1 Creating Decision Trees for GLAST analysis: A new C++-based procedure Toby Burnett Frank Golf.

Slides:



Advertisements
Similar presentations
DECISION TREES. Decision trees  One possible representation for hypotheses.
Advertisements

Michael Alves, Patrick Dugan, Robert Daniels, Carlos Vicuna
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
Decision Tree Approach in Data Mining
Introduction Training Complexity, Pruning CART vs. ID3 vs. C4.5
Chapter 7 – Classification and Regression Trees
Chapter 7 – Classification and Regression Trees
Decision Tree under MapReduce Week 14 Part II. Decision Tree.
Decision Trees Instructor: Qiang Yang Hong Kong University of Science and Technology Thanks: Eibe Frank and Jiawei Han.
Searching for Single Top Using Decision Trees G. Watts (UW) For the DØ Collaboration 5/13/2005 – APSNW Particles I.
Analysis Meeting 24Oct 05T. Burnett1 UW classification: new background rejection trees.
Analysis Meeting 26 Sept 05T. Burnett1 UW classification: background rejection (A summary – details at a future meeting)
Bill Atwood, July, 2003 GLAST 1 A GLAST Analysis Agenda Overarching Approach & Strategy Flattening Analysis Variables Classification Tree Primer Sorting.
T. Burnett: IRF status 30-Jan-061 DC2 IRF Status Meeting agenda, references at:
Optimization of Signal Significance by Bagging Decision Trees Ilya Narsky, Caltech presented by Harrison Prosper.
Analysis Meeting 31Oct 05T. Burnett1 Classification, etc. status at UW Implement “nested trees” Generate “boosted” trees for all of the Atwood categories.
ICS 273A Intro Machine Learning
Bill Atwood, SCIPP/UCSC, Dec., 2003 GLAST 1 DC1 Instrument Response Agenda Where are we? V3R3P7 Classification Trees Covariance Scaled PSF Pair Energies.
Analysis Meeting 31Oct 05T. Burnett1 Classification, etc. status Implement pre-filters Energy estimation.
Bill Atwood, SCIPP/UCSC, August, 2005 GLAST 1 Energy Method Selections Data Set: All Gamma (GR-HEAD1.615) Parametric Tracker Last Layer Profile 4 Methods.
C&A 8 May 06 1 Point Source Localization: Optimizing the CTBCORE cut Toby Burnett University of Washington.
Bill Atwood, Rome, Sept, 2003 GLAST 1 Instrument Response Studies Agenda Overarching Approach & Strategy Classification Trees Sorting out Energies PSF.
Analysis Meeting 17 Oct 05T. Burnett1 UW classification: background rejection Update with new all gamma and background samples.
ICS 273A Intro Machine Learning
Bill Atwood, Nov. 2002GLAST 1 Classification PSF Analysis A New Analysis Tool: Insightful Miner Classification Trees From Cuts Classification Trees: Recasting.
Analysis Meeting 4/09/05 - T. Burnett 1 Classification tree Update Toby Burnett Frank Golf.
1 Using Insightful Miner Trees for Glast Analysis Toby Burnett Analysis Meeting 2 June 2003.
Bill Atwood, SCIPP/UCSC, DC2 Closeout, May 31, 2006 GLAST 1 Bill Atwood & Friends.
T. Burnett: IRF status 5-feb-061 DC2 IRF Status Meeting agenda, references at:
Analysis Meeting 14Nov 05T. Burnett1 Classification, etc. status at UW New application to apply trees to existing tuple Background rejection for the v7r2.
Classification and Regression Trees for Glast Analysis: to IM or not to IM? Toby Burnett Data Challenge Meeting 15 June 2003.
Analysis Meeting 8 Aug 05T. Burnett1 Status of UW classification and PSF fits Data selection Classification PSF.
Bill Atwood, SCIPP/UCSC, Jan., 2006 GLAST 1 The 3 rd Pass Back Rejection Analysis using V7R3P4 (repo) 1)Training sample: Bkg: Runs (SLAC) v7r3p4.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Ensemble Learning (2), Tree and Forest
Decision Tree Models in Data Mining
Microsoft Enterprise Consortium Data Mining Concepts Introduction to Directed Data Mining: Decision Trees Prepared by David Douglas, University of ArkansasHosted.
Introduction to Directed Data Mining: Decision Trees
1 Chapter 1 Analysis Basics. 2 Chapter Outline What is analysis? What to count and consider Mathematical background Rates of growth Tournament method.
Chapter 9 – Classification and Regression Trees
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
Learning from Observations Chapter 18 Through
DC2 at GSFC - 28 Jun 05 - T. Burnett 1 DC2 C++ decision trees Toby Burnett Frank Golf Quick review of classification (or decision) trees Training and testing.
Decision Tree Learning Debapriyo Majumdar Data Mining – Fall 2014 Indian Statistical Institute Kolkata August 25, 2014.
B-tagging Performance based on Boosted Decision Trees Hai-Jun Yang University of Michigan (with X. Li and B. Zhou) ATLAS B-tagging Meeting February 9,
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Pass 6: GR V13R9 Muons, All Gammas, & Backgrounds Muons, All Gammas, & Backgrounds Data Sets: Muons: allMuon-GR-v13r9 (14- Jan-2008) AG: allGamma-GR-v13r9.
CS690L Data Mining: Classification
Training of Boosted DecisionTrees Helge Voss (MPI–K, Heidelberg) MVA Workshop, CERN, July 10, 2009.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
Feb. 7, 2007First GLAST symposium1 Measuring the PSF and the energy resolution with the GLAST-LAT Calibration Unit Ph. Bruel on behalf of the beam test.
ECE 471/571 – Lecture 20 Decision Tree 11/19/15. 2 Nominal Data Descriptions that are discrete and without any natural notion of similarity or even ordering.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Machine Learning Recitation 8 Oct 21, 2009 Oznur Tastan.
RECITATION 4 MAY 23 DPMM Splines with multiple predictors Classification and regression trees.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
BY International School of Engineering {We Are Applied Engineering} Disclaimer: Some of the Images and content have been taken from multiple online sources.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
Background Rejection Activities in Italy
DECISION TREES An internal node represents a test on an attribute.
Decision Trees an introduction.
Background Rejection Prep Work
Artificial Intelligence
Data Science Algorithms: The Basic Methods
Roberto Battiti, Mauro Brunato
LAT performance for DC2 and the “4-panel plot”
Decision Trees By Cole Daily CSCI 446.
CR Photon Working Group
STT : Intro. to Statistical Learning
Presentation transcript:

GLAST/LAT analysis group 31Jan05 - T. Burnett 1 Creating Decision Trees for GLAST analysis: A new C++-based procedure Toby Burnett Frank Golf

GLAST/LAT analysis group 31Jan05 - T. Burnett 2 Outline Quick review of classification (or decision) trees Training and testing How Bill does it with Insightful Miner Application to the “good-gamma” trees: how does it compare?

GLAST/LAT analysis group 31Jan05 - T. Burnett 3 Quick Review of Decision Trees Introduced to GLAST by Bill Atwood, using InsightfulMiner Each branch node is a predicate, or cut on a variable, like CalCsIRLn > If true, this defines the right branch, otherwise the left branch. If there is no branch, the node is a leaf; a leaf contains the purity of the sample that reaches that point Thus the tree defines a function of the event variables used, returning a value for the purity from the training sample predicate purity true: right false: left

GLAST/LAT analysis group 31Jan05 - T. Burnett 4 Training and testing procedure Analyze a training sample containing a mixture of “good” and “bad” events: I use the even events Choose set of variables and find the optimal cut for such that the left and right subsets are purer than the orignal. Two standard criteria for this: “Gini” and entropy. I currently use the former. W S : sum of signal weights W B : sum of background weights Gini = 2 W S W B /(W S +W B ) Thus Gini wants to be small. Actually maximize the improvement: Gini(parent)-Gini(left child)-Gini(right child) Apply this recursively until too few events. (100 for now) Finally test with the odd events: measure purity for each node

GLAST/LAT analysis group 31Jan05 - T. Burnett 5 Evaluate by Comparing with Bill CAL-Low CT Probabilities CAL-High CT Probabilities All Good Bad Good All Bad Good From Bill’s Rome ’03 talk: The “good cal” analysis

GLAST/LAT analysis group 31Jan05 - T. Burnett 6 Compare with Bill, cont Since Rome: Three energy ranges, three trees. CalEnergySum: 5-350; ; >3500 Resulting classification trees implemented in Gleam by a “classification” package: results saved to the IMgoodCal tuple variable. Data set for training, comparison: the all_gamma from v4r2 760 K events E from 16 MeV to 160 GeV (uniform in log(E), and 0<  < 90 (uniform in cos(  ) ). Contains IMgoodCal for comparison

GLAST/LAT analysis group 31Jan05 - T. Burnett 7 Bill-type plots for all_gamma

GLAST/LAT analysis group 31Jan05 - T. Burnett 8 Another way of looking at performance Define efficiency as fraction of good events after given cut on node purity; determine bad fraction for that cut

GLAST/LAT analysis group 31Jan05 - T. Burnett 9 The new classifier package Properties Implements Gini separation (entropy on todo list) Reads ROOT files Flexible specification of variables to use Simple and compact ascii representation of decision trees Not (yet) implemented: multiple trees, pruning Currently in users/burnett, I plan, after suitable notification, to copy it to the cvs root.

GLAST/LAT analysis group 31Jan05 - T. Burnett 10 Growing trees For simplicity, run a single tree: initially try all of Bill’s classification tree variables: list at right. The run over the full 750 K events, trying each of the 70 variables takes only a few minutes! "AcdActiveDist", "AcdTotalEnergy", "CalBkHalfRatio", "CalCsIRLn", "CalDeadTotRat", "CalDeltaT", "CalEnergySum", "CalLATEdge", "CalLRmsRatio", "CalLyr0Ratio", "CalLyr7Ratio", "CalMIPDiff", "CalTotRLn", "CalTotSumCorr", "CalTrackDoca", "CalTrackSep", "CalTwrEdge", "CalTwrGap", "CalXtalRatio", "CalXtalsTrunc", "EvtCalETLRatio", "EvtCalETrackDoca", "EvtCalETrackSep", "EvtCalEXtalRatio", "EvtCalEXtalTrunc", "EvtCalEdgeAngle", "EvtEnergySumOpt", "EvtLogESum", "EvtTkr1EChisq", "EvtTkr1EFirstChisq", "EvtTkr1EFrac", "EvtTkr1EQual", "EvtTkr1PSFMdRat", "EvtTkr2EChisq", "EvtTkr2EFirstChisq", "EvtTkrComptonRatio", "EvtTkrEComptonRatio", "EvtTkrEdgeAngle", "EvtVtxEAngle", "EvtVtxEDoca", "EvtVtxEEAngle", "EvtVtxEHeadSep", "Tkr1Chisq", "Tkr1DieEdge", "Tkr1FirstChisq", "Tkr1FirstLayer", "Tkr1Hits", "Tkr1KalEne", "Tkr1PrjTwrEdge", "Tkr1Qual", "Tkr1TwrEdge", "Tkr1ZDir", "Tkr2Chisq", "Tkr2KalEne", "TkrBlankHits", "TkrHDCount", "TkrNumTracks", "TkrRadLength", "TkrSumKalEne", "TkrThickHits", "TkrThinHits", "TkrTotalHits", "TkrTrackLength", "TkrTwrEdge", "VtxAngle", "VtxDOCA", "VtxHeadSep", "VtxS1", "VtxTotalWgt", "VtxZDir"

GLAST/LAT analysis group 31Jan05 - T. Burnett 11 Performance of the truncated list Note: this is cheating: evaluation using the same events as for training 1691 leaf nodes 3381 total Nameimprovement CalCsIRLn21000 CalTrackDoca3000 AcdTotalEnergy1800 CalTotSumCorr800 EvtCalEXtalTrunc600 EvtTkr1EFrac560 CalLyr7Ratio470 CalTotRLn400 CalEnergySum370 EvtLogESum370 EvtEnergySumOpt310 EvtTkrEComptonRatio290 EvtCalETrackDoca220 EvtVtxEEAngle140

GLAST/LAT analysis group 31Jan05 - T. Burnett 12 Plans Implement multiple trees: “boosting” Averaging Try truncation to reduce size of tree without losing performance This is only one of Bill’s variables: also need trees for: PSF tail reduction But we do not implement “regression trees” vertex choice: one track or vertex good-gamma prediction