7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland1 Object Reconstruction for CMS Analysis The Design Implementation and Deployment of a Functional.

Slides:



Advertisements
Similar presentations
CMS HLT production using Grid tools Flavia Donno (INFN Pisa) Claudio Grandi (INFN Bologna) Ivano Lippi (INFN Padova) Francesco Prelz (INFN Milano) Andrea.
Advertisements

1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Upgrading the CMS simulation and reconstruction David J Lange LLNL April CHEP 2015D. Lange.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
L.M.Barone – INFN Rome October 2000 ACAT FNAL Management of Large Scale Data Productions for the CMS Experiment Presented by L.M.Barone Università.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Framework for Automated Builds Natalia Ratnikova CHEP’03.
Introduction to RUP Spring Sharif Univ. of Tech.2 Outlines What is RUP? RUP Phases –Inception –Elaboration –Construction –Transition.
CMS Software and Computing FNAL Internal Review of USCMS Software and Computing David Stickland Princeton University CMS Software and Computing Deputy.
Dave Newbold, University of Bristol24/6/2003 CMS MC production tools A lot of work in this area recently! Context: PCP03 (100TB+) just started Short-term.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
Introduzione al Software di CMS N. Amapane. Nicola AmapaneTorino, Aprile Outline CMS Software projects The framework: overview Finding more.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
3rd Nov 2000HEPiX/HEPNT CDF-UK MINI-GRID Ian McArthur Oxford University, Physics Department
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
Computing in CMS May 24, 2002 NorduGrid Helsinki workshop Veikko Karimäki/HIP.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
11 December 2000 Paolo Capiluppi - DataGrid Testbed Workshop CMS Applications Requirements DataGrid Testbed Workshop Milano, 11 December 2000 Paolo Capiluppi,
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
David N. Brown Lawrence Berkeley National Lab Representing the BaBar Collaboration The BaBar Mini  BaBar  BaBar’s Data Formats  Design of the Mini 
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Summary of 1 TB Milestone RD Schaffer Outline : u Goals and assumptions of the exercise u The hardware constraints u The basic model u What have we understood.
5 May 98 1 Jürgen Knobloch Computing Planning for ATLAS ATLAS Software Week 5 May 1998 Jürgen Knobloch Slides also on:
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
19 November 98 1 Jürgen Knobloch ATLAS Computing ATLAS Computing - issues for 1999 Jürgen Knobloch Slides also on:
David Adams ATLAS DIAL status David Adams BNL November 21, 2002 ATLAS software meeting GRID session.
Fifth Lecture Hour 9:30 – 10:20 am, September 9, 2001 Framework for a Software Management Process – Life Cycle Phases (Part II, Chapter 5 of Royce’ book)
Andreas Morsch, CERN EP/AIP CHEP 2003 Simulation in ALICE Andreas Morsch For the ALICE Offline Project 2003 Conference for Computing in High Energy and.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
CMS Computing and Core-Software USCMS CB Riverside, May 19, 2001 David Stickland, Princeton University CMS Computing and Core-Software Deputy PM.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
Chapter 8 Workflows of the Process Taken from Walker Royce’s textbook – Software Project Management plus a number of Personal Comments.
Introduction to CMSSW Framework Concepts Simulation & Reconstruction Liz Sexton-Kennedy January 10, 2008.
GDB Meeting - 10 June 2003 ATLAS Offline Software David R. Quarrie Lawrence Berkeley National Laboratory
The CMS Simulation Software Julia Yarba, Fermilab on behalf of CMS Collaboration 22 m long, 15 m in diameter Over a million geometrical volumes Many complex.
June 29, 2000DOE/NSF USCMS Computing and Software Report. HLT Studies D. Acosta1 High-Level Trigger Studies Darin Acosta University of Florida DOE/NSF.
0 Fermilab SW&C Internal Review Oct 24, 2000 David Stickland, Princeton University CMS Software and Computing Status The Functional Prototypes.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
Computing R&D and Milestones LHCb Plenary June 18th, 1998 These slides are on WWW at:
23/2/2000Status of GAUDI 1 P. Mato / CERN Computing meeting, LHCb Week 23 February 2000.
Claudio Grandi INFN-Bologna CHEP 2000Abstract B 029 Object Oriented simulation of the Level 1 Trigger system of a CMS muon chamber Claudio Grandi INFN-Bologna.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
David Stickland CMS Core Software and Computing
Vincenzo Innocente, CERN/EPUser Collections1 Grid Scenarios in CMS Vincenzo Innocente CERN/EP Simulation, Reconstruction and Analysis scenarios.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
January 18, 2000DOE/NSF USCMS Computing and Software Review. HLT Studies D. Acosta1 High-Level Trigger Studies Darin Acosta University of Florida DoE/NSF.
Detector SimOOlation activities in ATLAS A.Dell’Acqua CERN-EP/ATC May 19th, 1999.
USCMS May 2002Jim Branson 1 Physics in US CMS US CMS Annual Collaboration Meeting May 2002 FSU Jin Branson.
L. Perini DATAGRID WP8 Use-cases 19 Dec ATLAS short term grid use-cases The “production” activities foreseen till mid-2001 and the tools to be used.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
1 CMS Virtual Data Overview Koen Holtman Caltech/CMS GriPhyN all-hands meeting, Marina del Rey April 9, 2001.
VI/ CERN Dec 4 CMS Software Architecture vs Hybrid Store Vincenzo Innocente CMS Week CERN, Dec
ATLAS Physics Analysis Framework James R. Catmore Lancaster University.
Marco Cattaneo, 3-June Event Reconstruction for LHCb  What is the scope of the project?  What are the goals (short+medium term)?  How do we organise.
CDF SAM Deployment Status Doug Benjamin Duke University (for the CDF Data Handling Group)
CMS High Level Trigger Configuration Management
The COMPASS event store in 2002
An ODBMS approach to persistency in CMS
US ATLAS Physics & Computing
Vincenzo Innocente CERN/EP/CMC
Ch 4. The Evolution of Analytic Scalability
ATLAS DC2 & Continuous production
Use of GEANT4 in CMS The OSCAR Project
CMS Software Architecture
Presentation transcript:

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland1 Object Reconstruction for CMS Analysis The Design Implementation and Deployment of a Functional Prototype OO Reconstruction Software for CMS CHEP February 2000 David Stickland,Princeton University

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland2 The ORCA Project Project Started in Fall 1998 l Primary Goals è Supply the collaboration with high-quality tools for detector optimization, trigger and physics studies è In-depth prototyping of SW architectures and tools l Secondary Goals è Accomplish a transition to Object Oriented style (people and code) è Increase base of skilled developers è Allow development of a user base è Test-bed for deployment of other OO projects

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland3 Transition Strategy l Top Down design of entire project impractical è Gain experience by deploying the project in phases l ORCA 1 (Dec 1998) è Basic reconstruction in all sub detectors, tolerate wrapped Fortran l ORCA 2 (Spring 1999) è OO Implementations of all major components l ORCA 3 (Oct 1999) è Persistency implemented è Used for Trigger Studies è In “Production” l ORCA 4 (Feb 2000) è Satisfy “Functional Prototype” Milestone. è A “complete” reconstruction package More than a set of reconstruction algorithms! We are Here

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland4 Training l Developers: è Established a finite set of understandable rules è Summer 99: many developers followed CMS organized OO Analysis/Design and UML courses. To date more than 50 physicists and SW engineers have code in ORCA l Users: è Many users need to be able to use ORCA without wanting to know much about OO design è Very successful web-based and interactive tutorial given an CERN in December ‘ New users followed this introductory course

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland5 Worldwide Collaboration l Worldwide collaboration and accessibility is at the crux both now and when LHC runs. SW is written by a few people at each of many institutes SW is used by a few (often different) users at each of many other institutes è Project Management weekly meetings held as VRVS conference (late PM CERN time, early AM Ca. time) p Typical ~5 remote sites connect p US users also organized their own SW/ORCA meetings è SCRAM developed in CMS. Very effective tool for code distribution and build management. p ORCA installed at steadily growing number of remote sites: Caltech, Fermilab, Moscow, Ecole Polytechnique, INFN (Bari)...

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland6 Persistency l Currently we use GEANT3 simulation via CMSIM package. Migration to GEANT4 during 2000 l Digitization is very expensive in CPU (200 min-bias events for each signal event at high-luminosity) l Objectivity Database è Use of Database is transparent to user program. No user code needs to be changed. è Objectivity handles etc, do not appear in user code è All Objectivity access is via CARF Framework l ORCA 3: Calorimeter and Muon digits persistent

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland7 ORCA and HLT l Trigger (HLT)/Physics studies have been first “user” for ORCA. l In October 99 we ran ORCA3 as a production job reading CMSIM signal and minimum bias files and filling two large databases with persistent ORCA results. è Used by Physicists studying HLT p 19 CPU’s running production (Linux and Sun) p Two production federations p Multiple AMS to reduce interference / timeouts p Automatic tape backup to HPSS as jobs complete p Automatic copying to two user federations p 30k events/day processed (>700GB in Obj/Db) è Ran essentially faultlessly

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland8 First HLT studies with ORCA Pythia Zebra files with HITS ORCA Digitization Objectivity Database HEPEVT ntuples CMSIM ORCA user Analysis ntuples PAW User Analysis ORCA ntuple production MC Prod. DB pop. User Analysis

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland9 Turn-over curves of DT Trigger Very preliminary!

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland10 March 2000 HLT Production Plans l 2M events ORCA reconstructed with high-luminosity pile-up l 2-4 Tera-Bytes in Objectivity/Db l 400 CPU-weeks l ~6 Production-Units è import existing catalogs, è schedule and monitor jobs è ooTidy and copy to hpss è export new catalog and system files l ~1-2 Production Units off CERN site l Copy of all data at CERN in hpss, use of IT/ASD AMS- backend to stage data to ~1TB of disk pools l Mirroring of Data to a few off-site centers, including trans- Atlantic Users want (need!) now what they were promised for

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland11 ORCA Production 2000 Signal Zebra files with HITS ORCA Digitization (merge signal and MB) Objectivity Database HEPEVT ntuples CMSIM HLT Algorithms New Reconstructed Objects MC Prod. ORCA Prod. HLT Grp Databases ORCA ooHit Formatter Objectivity Database MB Objectivity Database Catalog import Objectivity Database Objectivity Database ytivitcejbOesabataD Mirrored Db’s (US, Russia, Italy..)

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland12 CARF Experiences l CARF Architecture: On-demand reconstruction (see V.Innocente talk -this session) è Multi-layered applications p read CMSIM files, fill ooHit Db’s p merge min-bias and signal ooHit’s, create “crossing”, digitize p read ooDigi’s, create new RecObj p filter on meta-data, create new event-collections è All possible within one program but separate applications simplify production è Runtime loading p simplifies users responsibilities, application is built as actions are demanded p package initializers know the dependencies p accessing a new Db object triggers library loading

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland13 On-Demand & Production l On demand processing makes user code very simple è “give me an iterator over jets of type 1 constructed from Calorimeter recHits”: p jet algo 1 requests iterator over Calo recHits p Calo hit reconstructor requests iterator over CaloDigi p Calo Digitizer requests iterator over caloHits p Calo Hit Maker processes CMSIM files to extract GEANT3 hits è process branches to Db at any level that has been previously completed and made persistent è user needs to know nothing about deeper levels l Production process also profits from multi-layered approach.

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland14 Stability and Robustness l We have made very rapid progress due to (meeting) an aggressive set of milestones. è Cost in stability and robustness l We feel compelled to embark on a phase of consolidation, design and process improvement è Stability è Quality testing and quality improvement è Coherence p to ease use rather than to enforce monolithic style è Formal requirements, design and implementation documentation p to disseminate understanding rather than to accumulate paper Professional Engineering Crucial

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland15 After ORCA 4 l Close the macro- life-cycle: analysis  design  implementation  testing  analysis… è Re-analyze requirements and assess current design and implementations è More emphasis on data handling and its impact on design l Prepare for Pre-production Software.

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland16 Software & Architecture l When project started, most people were worried about ways to bring on the physicists, develop the sub-detector software etc. è Important, major emphasis of the last year, but actually less critical in the long term l Engineering of the architecture, and crucially the data- handling issues, are really the critical items è Tracking algorithms can, and will, be rewritten many times. But having an architecture that allows and keeps track of plug-and-play is vital. è Even now we face very large datasets (multi TB). Production, automation, mirroring, evolution are (some of) the hard issues. Reconstruction is much more than the reconstruction code

7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland17 Conclusions l In one year ORCA has gone from plan to reality è We have demonstrated it is possible to achieve an OO transition in a reasonably ordered way è We are supplying useful code for CMS design/optimization è The ORCA team now numbers about 50 developers è Persistency is the bedrock, not an addition è CARF Architecture proving suitable and flexible è We have made the transition from development code to production environment Now entering a critical phase of supporting deployed code and simultaneously undergoing full reanalysis and design...