Searching for Single Top Using Decision Trees G. Watts (UW) For the DØ Collaboration 5/13/2005 – APSNW Particles I.

Slides:



Advertisements
Similar presentations
DECISION TREES. Decision trees  One possible representation for hypotheses.
Advertisements

Introduction Training Complexity, Pruning CART vs. ID3 vs. C4.5
Matthew Schwartz Harvard University March 24, Boost 2011.
Recent Results on the Possibility of Observing a Standard Model Higgs Boson Decaying to WW (*) Majid Hashemi University of Antwerp, Belgium.
Chapter 7 – Classification and Regression Trees
Chapter 7 – Classification and Regression Trees
Summary of Results and Projected Sensitivity The Lonesome Top Quark Aran Garcia-Bellido, University of Washington Single Top Quark Production By observing.
Single Top Production at the Tevatron Top Quark Symposium University of Michigan April 7, 2005 Gordon Watts for DØ and CDF Searches.
Data Mining Techniques Outline
Top Physics at the Tevatron Mike Arov (Louisiana Tech University) for D0 and CDF Collaborations 1.
Optimization of Signal Significance by Bagging Decision Trees Ilya Narsky, Caltech presented by Harrison Prosper.
Introduction to Single-Top Single-Top Cross Section Measurements at ATLAS Patrick Ryan (Michigan State University) The measurement.
Bayesian Neural Networks Pushpa Bhat Fermilab Harrison Prosper Florida State University.
Three kinds of learning
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Single-Top Cross Section Measurements at ATLAS Patrick Ryan (Michigan State University) Introduction to Single-Top The measurement.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
T-CHANNEL MODELING UNCERTAINTIES AND FURTHER QUESTIONS TO TH AND NEW FIDUCIAL MEASUREMENTS Julien Donini, Jose E. Garcia, Dominic Hirschbuehl, Luca Lista,
Optimizing Higgs Analysis at DØ John Sandy – SULI Intern (Texas Tech University) Mentors: Michael P Cooke and Ryuji Yamada (Fermilab National Accelerator.
Vector Boson Scattering At High Mass
G. Cowan Lectures on Statistical Data Analysis Lecture 7 page 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem 2Random variables and.
Chapter 9 – Classification and Regression Trees
Michigan REU Final Presentations, August 10, 2006Matt Jachowski 1 Multivariate Analysis, TMVA, and Artificial Neural Networks Matt Jachowski
G. Cowan Statistical Methods in Particle Physics1 Statistical Methods in Particle Physics Day 3: Multivariate Methods (II) 清华大学高能物理研究中心 2010 年 4 月 12—16.
Matthew Schwartz Harvard University with J. Gallicchio, PRL, 105:022001,2010 (superstructure) with K. Black, J. Gallicchio, J. Huth, M. Kagan and B. Tweedie.
Use of Multivariate Analysis (MVA) Technique in Data Analysis Rakshya Khatiwada 08/08/2007.
B-tagging Performance based on Boosted Decision Trees Hai-Jun Yang University of Michigan (with X. Li and B. Zhou) ATLAS B-tagging Meeting February 9,
Today Ensemble Methods. Recap of the course. Classifier Fusion
KIRTI RANJANDIS, Madison, Wisconsin, April 28, Top Quark Production Cross- Section at the Tevatron Collider On behalf of DØ & CDF Collaboration KIRTI.
Evidence for Single Top Quark Production at DØ Ann Heinson University of California, Riverside for the DØ Collaboration CERN Particle Physics Seminar Tuesday.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
B-Tagging Algorithms for CMS Physics
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
By Henry Brown Henry Brown, LHCb, IOP 10/04/13 1.
Analysis of H  WW  l l Based on Boosted Decision Trees Hai-Jun Yang University of Michigan (with T.S. Dai, X.F. Li, B. Zhou) ATLAS Higgs Meeting September.
B-tagging based on Boosted Decision Trees
1 Measurement of the Mass of the Top Quark in Dilepton Channels at DØ Jeff Temple University of Arizona for the DØ collaboration DPF 2006.
Single top quark physics Peter Dong, UCLA on behalf of the CDF and D0 collaborations Les Rencontres de Physique de la Vallee d’Aoste Wednesday, February.
G. Cowan Lectures on Statistical Data Analysis Lecture 6 page 1 Statistical Data Analysis: Lecture 6 1Probability, Bayes’ theorem 2Random variables and.
Top Quark Properties and Search for Single Top Quark at the Tevatron Meenakshi Narain Boston University Presented at EPS 2005.
Search for the Standard Model Higgs in  and  lepton final states P. Grannis, ICHEP 2012 for the DØ Collaboration Tevatron, pp √s = 1.96 TeV -
Jessica Levêque Rencontres de Moriond QCD 2006 Page 1 Measurement of Top Quark Properties at the TeVatron Jessica Levêque University of Arizona on behalf.
EPS Manchester Daniela Bortoletto Associated Production for the Standard Model Higgs at CDF D. Bortoletto Purdue University Outline: Higgs at the.
1 Reinhard Schwienhorst, MSU Top Group Meeting W' Search in the single top quark channel Reinhard Schwienhorst Michigan State University Top Group Meeting,
1 Reinhard Schwienhorst, Michigan State University Search for Single Top Quark Production at DØ in Run II Reinhard Schwienhorst for the DØ Collaboration.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Single Top Quark Production at D0, L. Li (UC Riverside) EPS 2007, July Liang Li University of California, Riverside On Behalf of the DØ Collaboration.
I'm concerned that the OS requirement for the signal is inefficient as the charge of the TeV scale leptons can be easily mis-assigned. As a result we do.
Studies for the Single Top Quark t-channel measurement with the ATLAS experiment 1 Ph. Sturm D. Hirschbühl, W. Wagner Bergische Universität Wuppertal BND.
Investigation on CDF Top Physics Group Ye Li Graduate Student UW - Madison.
Eric COGNERAS LPC Clermont-Ferrand Prospects for Top pair resonance searches in ATLAS Workshop on Top Physics october 2007, Grenoble.
Low Mass Standard Model Higgs Boson Searches at the Tevatron Andrew Mehta Physics at LHC, Split, Croatia, September 29th 2008 On behalf of the CDF and.
1 Donatella Lucchesi July 22, 2010 Standard Model High Mass Higgs Searches at CDF Donatella Lucchesi For the CDF Collaboration University and INFN of Padova.
The Hunt For Single Top Production
First Evidence for Electroweak Single Top Quark Production
Ch9: Decision Trees 9.1 Introduction A decision tree:
Optimizing Foam, A Monte Carlo Event Generator
Search for WHlnbb at the Tevatron DPF 2009
Multi-dimensional likelihood
10th China HEPS Particle Physics Meeting
W boson helicity measurement
Computing and Statistical Data Analysis Stat 5: Multivariate Methods
Greg Heath University of Bristol
Prospects on Lonely Top Quarks searches in ATLAS
Tim Scanlon Imperial College, London on behalf of the DØ Collaboration
The Hunt For Single Top Production
Measurement of the Single Top Production Cross Section at CDF
Northern Illinois University / NICADD
Presentation transcript:

Searching for Single Top Using Decision Trees G. Watts (UW) For the DØ Collaboration 5/13/2005 – APSNW Particles I

Gordon Watts (UW) APSNW Meeting May 13, SingleTop Challenges Overwhelming Background! Straight Cuts Difficulty taking advantage of correlations (and counting experiments) Multivariate Cuts (and shape fitting) Designed to take advantage of correlations and irreducible backgrounds

Gordon Watts (UW) APSNW Meeting May 13, Asymmetries in t-Channel Production b Pair Production Lots of variables give small separation (Use ME, phase space, etc.)

Gordon Watts (UW) APSNW Meeting May 13, Combine Variables! Multivariate Likelihood Fit 7 variables means 7 dimensions… Neural Network Many inputs and a single output Trained on signal and background sample Well understood and mostly accepted in HEP Decision Tree Many inputs and a single output Trained on signal and background sample Used mostly in life sciences & business (MiniBOONE - physics/ ).

Gordon Watts (UW) APSNW Meeting May 13, Decision Tree Trained Decision Tree (Binned Likelihood Fit) (Limit)

Gordon Watts (UW) APSNW Meeting May 13, Internals of a Trained Tree Every Event belongs to a single leaf node! “Rooted Binary Tree” “You can see a decision tree”

Gordon Watts (UW) APSNW Meeting May 13, Training Determine a branch point Calculate Gini Improvement As a function of a interesting variable (H T in this case) Choose the largest improvement as the cut point Repeat for all interesting variables HT, Jet pT, Angular Variables, etc. Best improvement is this node’s decision.

Gordon Watts (UW) APSNW Meeting May 13, Gini Process Requires a Variable to optimize separation. W s – Weight of Signal Events W b – Weight of Background Events Purity Gini G is zero for pure background or signal!

Gordon Watts (UW) APSNW Meeting May 13, Gini Improvement Data (S) S1S1 S2S2 For each node GI = G(S) – G(S 1 ) – G(S 2 ) Repeat the process for each subdivision of data

Gordon Watts (UW) APSNW Meeting May 13, And Cut… Determine the Purity of each leaf Stop process and generate a leaf. We used statistical sample error (# of events) Use Tree as Estimator of Purity Each event belongs to a unique leaf The leaf’s purity is the estimator of the event

Gordon Watts (UW) APSNW Meeting May 13, DT in the Single Top Search DT Wbb DT tt l+jets Two DTs 2d Histogram used in binned likelihood fit Trained on signal and Wbb as background Trained on signal and tt  lepton + jets as background DØ This part is identical to a NN based analysis Separate DT for muon & electron Backgrounds: W+Jets, QCD, top Pair Production Fake Leptons

Gordon Watts (UW) APSNW Meeting May 13, Results Expected Limits s-channel: 4.5 pb (NN: 4.5) t-channel: 6.4 pb (NN: 5.8) Actual Limits s-channel: 8.3 pb (NN: 6.4) t-channel: 8.1 pb (NN: 5.0) Expected Results Close to NN

Gordon Watts (UW) APSNW Meeting May 13, Future of the Analysis Use a Single Decision Tree Train it against all backgrounds Pruning Train until each leaf has only a single event Recombine leaves (pruning) using statistical estimator Boosting Combine multiple trees, each weighted Train trees on event samples that have mis- classified event weights enhanced

Gordon Watts (UW) APSNW Meeting May 13, References & Introduction MiniBooNE Paper: hep-ex/ Recent Advances in Predictive (Machine) Learning Jerome H. Friedman, Conf. Proceedings I have then linked and other on my web page conferences

Gordon Watts (UW) APSNW Meeting May 13, Conclusions Decision Trees are good… –Model is obvious in form of 2d binary tree. –Not as sensitive to outliers in input data as other methods –Easily accommodate integer inputs (N Jets ) or missing variable inputs. –Easy to implement (several months to go from scratch to working code) Decision Trees aren’t so good… –Well understood input variables are a must Similar for Neural Networks, of course. –Minor changes in the input events can make for major changes in tree layout and results. –Estimator is not a continuous function Don’t have to deal with hidden nodes –Separate training of background or other issues