Vladimir Litvin, Harvey Newman, Sergey Schevchenko Caltech CMS Scott Koranda, Bruce Loftis, John Towns NCSA Miron Livny, Peter Couvares, Todd Tannenbaum,

Slides:



Advertisements
Similar presentations
Peter Couvares and Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison
Advertisements

1 Concepts of Condor and Condor-G Guy Warner. 2 Harvesting CPU time Teaching labs. + Researchers Often-idle processors!! Analyses constrained by CPU time!
INTRODUCTION TO e/ ɣ IN ATLAS In order to acquire the full physics potential of the LHC, the ATLAS electromagnetic calorimeter must be able to identify.
1 Analysis of Prompt Diphoton Production at the Large Hadron Collider. Andy Yen Mentor: Harvey Newman Co-Mentors: Marat Gataullin, Vladimir Litvine California.
Condor-G: A Computation Management Agent for Multi-Institutional Grids James Frey, Todd Tannenbaum, Miron Livny, Ian Foster, Steven Tuecke Reporter: Fu-Jiun.
A Computation Management Agent for Multi-Institutional Grids
Condor and GridShell How to Execute 1 Million Jobs on the Teragrid Jeffrey P. Gardner - PSC Edward Walker - TACC Miron Livney - U. Wisconsin Todd Tannenbaum.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1  trigger optimization in CMS Tracker Giuseppe Bagliesi On behalf of  tracking group Workshop on B/tau Physics at LHC Helsinki, May 30 - June 1, 2002.
GRID workload management system and CMS fall production Massimo Sgaravatto INFN Padova.
1 Rutherford Appleton Laboratory The 13th Annual International Conference on Supersymmetry and Unification of the Fundamental Interactions Durham, 2005.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
1 N. Davidson E/p single hadron energy scale check with minimum bias events Jet Note 8 Meeting 15 th May 2007.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
GRID Workload Management System Massimo Sgaravatto INFN Padova.
Workload Management Massimo Sgaravatto INFN Padova.
The Difficulties of Distributed Data Douglas Thain Condor Project University of Wisconsin
US CMS Testbed A Grid Computing Case Study Alan De Smet Condor Project University of Wisconsin at Madison
Zach Miller Computer Sciences Department University of Wisconsin-Madison What’s New in Condor.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Vladimir Litvin, Harvey Newman Caltech CMS Scott Koranda, Bruce Loftis, John Towns NCSA Miron Livny, Peter Couvares, Todd Tannenbaum, Jamie Frey Wisconsin.
Workflow Management in Condor Gökay Gökçay. DAGMan Meta-Scheduler The Directed Acyclic Graph Manager (DAGMan) is a meta-scheduler for Condor jobs. DAGMan.
L3 Filtering: status and plans D  Computing Review Meeting: 9 th May 2002 Terry Wyatt, on behalf of the L3 Algorithms group. For more details of current.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
Tau Jet Identification in Charged Higgs Search Monoranjan Guchait TIFR, Mumbai India-CMS collaboration meeting th March,2009 University of Delhi.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Job Submission Condor, Globus, Java CoG Kit Young Suk Moon.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
28 April 0 Yaxian Mao, Daicui Zhou, Yves Schutz In ALICE Physics Workgroup: High p T and photons ( for ALICE collaboration -- Wuhan)
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Development Timelines Ken Kennedy Andrew Chien Keith Cooper Ian Foster John Mellor-Curmmey Dan Reed.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Welcome and Condor Project Overview.
Grid Workload Management Massimo Sgaravatto INFN Padova.
Grid Compute Resources and Job Management. 2 Local Resource Managers (LRM)‏ Compute resources have a local resource manager (LRM) that controls:  Who.
22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Report from USA Massimo Sgaravatto INFN Padova. Introduction Workload management system for productions Monte Carlo productions, data reconstructions.
W + /W - and l + /l - A Means to investigate PDFs T. Schörner-Sadenius, G. Steinbrück Hamburg University HERA-LHC Workshop, CERN, October 2004.
July 11-15, 2005Lecture3: Grid Job Management1 Grid Compute Resources and Job Management.
Review of Condor,SGE,LSF,PBS
Kati Lassila-Perini/HIP HIP CMS Software and Physics project evaluation1/ Electron/ physics in CMS Kati Lassila-Perini HIP Activities in the.
US CMS Centers & Grids – Taiwan GDB Meeting1 Introduction l US CMS is positioning itself to be able to learn, prototype and develop while providing.
FIMCMS, 26 May, 2008 S. Lehti HIP Charged Higgs Project Preparative Analysis for Background Measurements with Data R.Kinnunen, M. Kortelainen, S. Lehti,
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
Higgs Reach Through VBF with ATLAS Bruce Mellado University of Wisconsin-Madison Recontres de Moriond 2004 QCD and High Energy Hadronic Interactions.
GRID activities in Wuppertal D0RACE Workshop Fermilab 02/14/2002 Christian Schmitt Wuppertal University Taking advantage of GRID software now.
Peter F. Couvares Computer Sciences Department University of Wisconsin-Madison Condor DAGMan: Managing Job.
A Fully Automated Fault- tolerant System for Distributed Video Processing and Off­site Replication George Kola, Tevfik Kosar and Miron Livny University.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Study of pair-produced doubly charged Higgs bosons with a four muon final state at the CMS detector (CMS NOTE 2006/081, Authors : T.Rommerskirchen and.
Γ +Jet Analysis for the CMS Pooja Gupta, Brajesh Choudhary, Sudeep Chatterji, Satyaki Bhattacharya & R.K. Shivpuri University of Delhi, India.
Peter Couvares Computer Sciences Department University of Wisconsin-Madison Condor DAGMan: Introduction &
Grid Compute Resources and Job Management. 2 Grid middleware - “glues” all pieces together Offers services that couple users with remote resources through.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Condor and (the) Grid (one of.
Régis Lefèvre (LPC Clermont-Ferrand - France)ATLAS Physics Workshop - Lund - September 2001 In situ jet energy calibration General considerations The different.
DZero Monte Carlo Production Ideas for CMS Greg Graham Fermilab CD/CMS 1/16/01 CMS Production Meeting.
University of Iowa Study of qq->qqH  qq ZZ Alexi Mestvirishvili November 2004, CMS PRS Workshop at FNAL.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
V. Pozdnyakov Direct photon and photon-jet measurement capability of the ATLAS experiment at the LHC Valery Pozdnyakov (JINR, Dubna) on behalf of the HI.
Workload Management Workpackage
Condor DAGMan: Managing Job Dependencies with Condor
U.S. ATLAS Grid Production Experience
Harnessing the Power of Condor for Human Genetics
US CMS Testbed.
Grid Canada Testbed using HEP applications
Condor Tutorial Asia Pacific Grid Workshop Tokyo, Japan October 2001
Condor-G Making Condor Grid Enabled
Northern Illinois University / NICADD
Presentation transcript:

Vladimir Litvin, Harvey Newman, Sergey Schevchenko Caltech CMS Scott Koranda, Bruce Loftis, John Towns NCSA Miron Livny, Peter Couvares, Todd Tannenbaum, Jamie Frey Wisconsin Condor Using of Grid Prototype Infrastructure for QCD Background Study to the H   Process on Alliance Resources

CMS Physics The CMS detector at the LHC will probe fundamental forces in our Universe and search for the yet undetected Higgs Boson Detector expected to come online 2007

CMS Physics

Leveraging Alliance Grid Resources The Caltech CMS group is using Alliance Grid resources today for detector simulation and data processing prototyping Even during this simulation and prototyping phase the computational and data challenges are substantial

Goal to simulate QCD background The QCD jet-jet background cross section is huge (~ pb). Previous studies of QCD jet-jet background have got the estimation of the rate, R jet, when jet might be misidentified as photon and, due to the limited CPU power, for QCD jet-jet background rates where simply squared (R jet 2 ). Hence, the correlations within event have not been taken into account in previous studies Previous simulations have been done with simplified geometry and non-gaussian tails in the resolution have not been adequately simulated Our goal is to make full simulation of relatively large QCD sample, measure the rate of diphoton misidentification and compare it with other types of background

Generation of QCD background QCD jet cross section strongly depends on the p T of the parton in hard interaction QCD jet cross section is huge. We need reasonable preselection at the generator level before pass events through full detector simulation The optimal choice of p T is needed Our choice is p T = 35GeV p T = 35 GeV is safe cut: we do not lose significant fraction of events, which could fake the Higgs signal, at the preselection level

Generator level cuts QCD background Standard CMS cuts: E t1 >40 GeV, E t2 >25 GeV, |  1,2 |<2.5 at least one pair of any two neutral particles (  0, , e, ,  ’, ,  ) with –E t1 > 37.5 GeV –E t2 > 22.5 GeV –|  1,2 | < 2.5 –m inv in GeV Rejection factor at generator level ~3000 Photon bremsstrahlung background Standard CMS cuts: E t1 >40 GeV, E t2 >25 GeV, |  1,2 |<2.5 at least one neutral particle (  0, , e, ,  ’, ,  ) with –E t > 37.5 GeV –|  | < 2.5 Rejection factor at generator level ~6

Challenges of a CMS Run CMS run naturally divided into two phases –Monte Carlo detector response simulation –100’s of jobs per run –each generating ~ 1 GB –all data passed to next phase and archived –reconstruct physics from simulated data –100’s of jobs per run –jobs coupled via Objectivity database access –~100 GB data archived Specific challenges –each run generates ~100 GB of data to be moved and archived –many, many runs necessary –simulation & reconstruction jobs at different sites –large human effort starting & monitoring jobs, moving data

Tools Generation level - PYTHIA (CTEQ 4L structure functions) Full Detector Simulation - CMSIM 121 (includes full silicon version of the tracker) Reconstruction - ORCA with pileup at L = 2 *10 33 cm -2 /s (~30 pileup events per signal event) -

Analysis Chain Full analysis chain

Meeting Challenge With Globus and Condor Globus middleware deployed across entire Alliance Grid remote access to computational resources dependable, robust, automated data transfer Condor strong fault tolerance including checkpointing and migration job scheduling across multiple resources layered over Globus as “personal batch system” for the Grid

CMS Run on the Alliance Grid Caltech CMS staff prepares input files on local workstation Pushes “one button” to launch master Condor job Input files transferred by master Condor job to Wisconsin Condor pool (~700 CPUs) using Globus GASS file transfer Master Condor job running at Caltech Caltech workstation Input files via Globus GASS WI Condor pool

CMS Run on the Alliance Grid Master Condor job at Caltech launches secondary Condor job on Wisconsin pool Secondary Condor job launches 100 Monte Carlo jobs on Wisconsin pool –each runs 12~24 hours –each generates ~1GB data –Condor handles checkpointing & migration –no staff intervention Master Condor job running at Caltech Secondary Condor job on WI pool 100 Monte Carlo jobs on Wisconsin Condor pool

CMS Run on the Alliance Grid When each Monte Carlo job completes data automatically transferred to UniTree at NCSA –each file ~ 1 GB –transferred using Globus-enabled FTP client “gsiftp” –NCSA UniTree runs Globus-enabled FTP server –authentication to FTP server on user’s behalf using digital certificate 100 Monte Carlo jobs on Wisconsin Condor pool NCSA UniTree with Globus-enabled FTP server 100 data files transferred via gsiftp, ~ 1 GB each

CMS Run on the Alliance Grid When all Monte Carlo jobs complete Secondary Condor reports to Master Condor at Caltech Master Condor at Caltech launches job to stage data from NCSA UniTree to NCSA Linux cluster –job launched via Globus jobmanager on cluster –data transferred using Globus-enabled FTP –authentication on user’s behalf using digital certificate Master starts job via Globus jobmanager on cluster to stage data Secondary Condor job on WI pool NCSA Linux cluster Secondary reports complete to master Master Condor job running at Caltech gsiftp fetches data from UniTree

CMS Run on the Alliance Grid Master Condor at Caltech launches physics reconstruction jobs on NCSA Linux cluster –job launched via Globus jobmanager on cluster –Master Condor continually monitors job and logs progress locally at Caltech –no user intervention required –authentication on user’s behalf using digital certificate Master Condor job running at Caltech Master starts reconstruction jobs via Globus jobmanager on cluster NCSA Linux cluster

CMS Run on the Alliance Grid When reconstruction jobs complete data automatically archived to NCSA UniTree –data transferred using Globus-enabled FTP After data transferred run is complete and Master Condor at Caltech s notification to staff NCSA Linux cluster data files transferred via gsiftp to UniTree for archiving

Condor Details for Experts Use CondorG –Condor + Globus –allows Condor to submit jobs to remote host via a Globus jobmanager –any Globus-enabled host reachable (with authorization) –Condor jobs run in the “Globus” universe –use familiar Condor classads for submitting jobs universe = globus globusscheduler = beak.cs.wisc.edu/jobmanager- condor-INTEL-LINUX environment = CONDOR_UNIVERSE=scheduler executable = CMS/condor_dagman_run arguments = -f -t -l. -Lockfile cms.lock -Condorlog cms.log -Dag cms.dag -Rescue cms.rescue input = CMS/hg_90.tar.gz remote_initialdir = Prod2001 output = CMS/hg_90.out error = CMS/hg_90.err log = CMS/condor.log notification = always queue

Condor Details for Experts Exploit Condor DAGman –DAG=directed acyclic graph –submission of Condor jobs based on dependencies –job B runs only after job A completes, job D runs only after job C completes, job E only after A,B,C & D complete… –includes both pre- and post-job script execution for data-staging, cleanup, or the like Job jobA_632 Prod2000/hg_90_gen_632.cdr Job jobB_632 Prod2000/hg_90_sim_632.cdr Script pre jobA_632 Prod2000/pre_632.csh Script post jobB_632 Prod2000/post_632.csh PARENT jobA_632 CHILD jobB_632 Job jobA_633 Prod2000/hg_90_gen_633.cdr Job jobB_633 Prod2000/hg_90_sim_633.cdr Script pre jobA_633 Prod2000/pre_633.csh Script post jobB_633 Prod2000/post_633.csh PARENT jobA_633 CHILD jobB_633

Monte Carlo Samples Simulated and Reconstructed

CPU timing

All cuts except isolation are applied Distributions are normalized to L int = 40 pb -1

Isolation Tracker isolation Isolation cut: Number of tracks with p T > 1.5 GeV in R = 0.30 cone around photon candidate is zero Still optimizing p T threshold and cone sizes Ecal isolation Sum of E t energy in the cone around photon candidate, using E t energies of ECAL clusters Isolation cut: Sum of E t energy in R = 0.30 cone around photon candidate is less than 0.8 GeV

Background Cross Section

Conclusions The goal of this study is to increase efficiency of computer resources and to reduce and minimize human intervention during simulation and reconstruction –“proof of concept” - it is possible to create the distributed system based on GLOBUS and Condor (MOP is operational now) –A lot of work ahead in order to make this system as automatic as possible Important results are obtained for the Higgs boson search in two photon decay mode –the main background is the background with one prompt photon plus bremsstrahlung photon or isolated  0, which is ~50% of the total background. QCD background is reduced down to the 15% of the total background –More precise studies need much more CPU time