1 The MEG Software Project PSI 9/2/2005Corrado Gatto Offline Architecture Computing Model Status of the Software Organization.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Kondo GNANVO Florida Institute of Technology, Melbourne FL.
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
MEG-Review Feb MEG Software Group MEG Software Status Framework for MC, Schedule, DC reconstruction update and Discussion on a ROOT-based offline.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
O. Stézowski IPN Lyon AGATA Week September 2003 Legnaro Data Analysis – Team #3 ROOT as a framework for AGATA.
Report from the Lecce’s Offline Group Rome 6-Feb-2006 The Achievements The People Future Plans.
August 98 1 Jürgen Knobloch ATLAS Software Workshop Ann Arbor ATLAS Computing Planning ATLAS Software Workshop August 1998 Jürgen Knobloch Slides also.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
19 July 2005Fabrizio Cei1 Status and Perspectives of MEG software Fabrizio Cei INFN & University of Pisa On behalf of the MC/Offline group MEG Review Meeting.
Shuei MEG review meeting, 2 July MEG Software Status MEG Software Group Framework Large Prototype software updates Database ROME Monte Carlo.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
The GlueX Collaboration Meeting October 4-6, 2012 Jefferson Lab Curtis Meyer.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Offline Coordinators  CMSSW_7_1_0 release: 17 June 2014  Usage:  Generation and Simulation samples for run 2 startup  Limited digitization and reconstruction.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
As of 28 Juni 2005Getting Starged with GEM - Shuei Yamada 1 Getting Started with GEM Shuei YAMADA ICEPP, University of Tokyo What is GEM? Before you start.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
David N. Brown Lawrence Berkeley National Lab Representing the BaBar Collaboration The BaBar Mini  BaBar  BaBar’s Data Formats  Design of the Mini 
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
14 February 2007Fabrizio Cei1 INFN and University of Pisa PSI Review Meeting PSI, 14 February 2007 Status of MEG Software.
BESIII MC Release notes & planned development Dengzy, Hem, Liuhm, Youzy, Yuany Nov. 23, 2005.
5 May 98 1 Jürgen Knobloch Computing Planning for ATLAS ATLAS Software Week 5 May 1998 Jürgen Knobloch Slides also on:
Prediction W. Buchmueller (DESY) arXiv:hep-ph/ (1999)
CBM Software Workshop for Future Challenges in Tracking and Trigger Concepts, GSI, 9 June 2010 Volker Friese.
19 November 98 1 Jürgen Knobloch ATLAS Computing ATLAS Computing - issues for 1999 Jürgen Knobloch Slides also on:
October 20th, 2006 C. Gatto 1 4 th Concept Software Status Report.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Framework for MDO Studies Amitay Isaacs Center for Aerospace System Design and Engineering IIT Bombay.
CMS pixel data quality monitoring Petra Merkel, Purdue University For the CMS Pixel DQM Group Vertex 2008, Sweden.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
06 February 2007Fabrizio Cei1 INFN and University of Pisa INFN Scientific Commission I Rome, 06 February 2007 Status of MEG Software.
CMS Computing and Core-Software USCMS CB Riverside, May 19, 2001 David Stickland, Princeton University CMS Computing and Core-Software Deputy PM.
Introduction CMS database workshop 23 rd to 25 th of February 2004 Frank Glege.
Databases in CMS Conditions DB workshop 8 th /9 th December 2003 Frank Glege.
Hall-D/GlueX Software Status 12 GeV Software Review III February 11[?], 2015 Mark Ito.
STAR Event data storage and management in STAR V. Perevoztchikov Brookhaven National Laboratory,USA.
5/2/  Online  Offline 5/2/20072  Online  Raw data : within the DAQ monitoring framework  Reconstructed data : with the HLT monitoring framework.
20 September 2005Fabrizio Cei1 Status and Perspectives of MEG Software Fabrizio Cei INFN & University of Pisa INFN Scientific Committee I Napoli, 20 September.
News on GEM Readout with the SRS, DATE & AMORE
Online Reconstruction 1M.Ellis - CM th October 2008.
Paul Scherrer Institut 5232 Villigen PSI CHEP 2006 in Mumbay / / Matthias Schneebeli ROME CHEP 2006 Presented by Matthias Schneebeli a universally.
Linda R. Coney – 5 November 2009 Online Reconstruction Linda R. Coney 5 November 2009.
University user perspectives of the ideal computing environment and SLAC’s role Bill Lockman Outline: View of the ideal computing environment ATLAS Computing.
Computing R&D and Milestones LHCb Plenary June 18th, 1998 These slides are on WWW at:
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
The JANA Reconstruction Framework David Lawrence - JLab May 25, /25/101JANA - Lawrence - CLAS12 Software Workshop.
Online Monitoring System at KLOE Alessandra Doria INFN - Napoli for the KLOE collaboration CHEP 2000 Padova, 7-11 February 2000 NAPOLI.
General requirements for BES III offline & EF selection software Weidong Li.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
The MEG Offline Project General Architecture Offline Organization Responsibilities Milestones PSI 2/7/2004Corrado Gatto INFN.
Victoria, Sept WLCG Collaboration Workshop1 ATLAS Dress Rehersals Kors Bos NIKHEF, Amsterdam.
Proposal for the MEG Offline System Assisi 9/21/2004Corrado Gatto General Architecture Computing Model Organization & Responsibilities Milestones.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
A proposal for the KM3NeT Computing Model Pasquale Migliozzi INFN - Napoli 1.
MAUS Status A. Dobbs CM43 29 th October Contents MAUS Overview Infrastructure Geometry and CDB Detector Updates CKOV EMR KL TOF Tracker Global Tracking.
1 GlueX Software Oct. 21, 2004 D. Lawrence, JLab.
9 Feb, MEG Software Status -for the coming review committee- Question from A.Blondel Our answer Schedule and Man power MEG Software Group.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
Monthly video-conference, 18/12/2003 P.Hristov1 Preparation for physics data challenge'04 P.Hristov Alice monthly off-line video-conference December 18,
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Database Replication and Monitoring
CMS High Level Trigger Configuration Management
Migration of reconstruction and analysis software to C++
LHC experiments Requirements and Concepts ALICE
ALICE analysis preservation
Commissioning of the ALICE HLT, TPC and PHOS systems
ProtoDUNE SP DAQ assumptions, interfaces & constraints
ATLAS DC2 & Continuous production
Development of LHCb Computing Model F Harris
Presentation transcript:

1 The MEG Software Project PSI 9/2/2005Corrado Gatto Offline Architecture Computing Model Status of the Software Organization

2 The Offline Architecture

3 Dataflow and Reconstruction Requirements 100 Hz L3 trigger evt size : 1.2 MB Raw Data throughput: (10+10)Hz  1.2Mb/Phys evt  Hz  0.01Mb/bkg evt = 3.5Mb/s : 35 kB Total raw data storage: 3.5Mb/s  10 7 s = 35 TB/yr

4 Requirements for Software Architecture Geant3 compatible (at least at the beginning) Easy interface with existing packages: – Geant3, Geant4, external (fortran) event generators Scalability Simple structure to be used by non-computing experts Written and maintained by few people Portability Use a world-wide accepted framework Use ROOT + An existing Offline Package as starting point

5 Expected Raw Performance Pure Linux setup 20 data sources FastEthernet local connection

6 MC & Data Processing Scheme Simu sigSimu bkg 1 hits files sighits files bkg 1 SDigitizer SDigits files sig SDigitizer SDigits files bkg Digitizer Merged Digits filesRaw Data Reconstruct ESD Reconstruct ESD C++ Fortran & G3 Or C++ & G3/G4/Fluka Simu bkg 2 hits files bkg 2 SDigitizer SDigits files bkg Fortran & G3 Or C++

7 General Architecture: Guidelines Ensure high level of modularity (for easy of maintenance) Absence of code dependencies between different detector modules (to C++ header problems) The structure of every detector package is designed so that static parameters (like geometry and detector response parameters) are stored in distinct objects The data structure is build up as ROOT TTree-objects Access is possible to either the full set of correlated data (i.e., the event) or only one or more sub-sample, stored in different branches of the data structure (TBranch class) and corresponding to one or more detector.

8 Computing Model

9 Elements of a Computing Model Components – Data Model Event data sizes, formats, streaming Data “Tiers” (DST/ESD/AOD etc) Roles, accessibility, distribution,… Calibration/Conditions data Flow, latencies, update freq Simulation, Sizes, distribution File size – Analysis Model Canonical group needs in terms of data, streams, re-processing, calibrations Data Movement, Job Movement, Priority management Interactive analysis Implementation – Computing Strategy and Deployment Roles of Computing Tiers Data Distribution between Tiers Data Management Architecture Databases Masters, Updates, Hierarchy Active/Passive Experiment Policy – Computing Specifications Profiles (Tier N & Time) –Processors, –Storage, –Network (Wide/Local), –DataBase services, –Specialized servers Middleware requirements

10 CPU Needed Calibration: CPU’s MC Production: CPU’s MC Reconstruction: CPU/repr. Raw data reconstruction: 3 – 14 CPU/repr. Alignment: to be estimated Assume: – Trigger rate: 20 Hz – Beam Duty Cycle: 50% – MC = data – Calibration: 1Hz Estimates by scaling existing code and a C++ framework Optimal solution for reconstruction:take Beam Duty Cycle: 100% and use no-beam time for reprocessing. -> Double the CPU for reconstruction.

11 Storage Needed Assume: – Raw data (compressed) event size: 120 kB – Hits (from MC): 9 kB/evt – Sdigit+Digit size (from MC): 30kB + 120kB – Track reference (from MC): 15 kByte – Kinematics (from MC): 4 kByte – ESD size (data or MC): 30 kByte Storage required yearly ( L3 trigger: 20Hz and DC: 50%): – Raw data: 12 Tbyte (+ calib) – Hits (from MC): 0.9 TByte – Digit size (from MC): 15 TByte – Track reference (from MC): 1.5 TByte – Kinematics (from MC): 0.4 TByte – ESD size (data or MC): 3/Tbyte /reproc. From Alice data model

12 A Computing Model for MEG The crucial decisions are being taken at the collaboration level Very dependent on CPU and storage requirement Two degrees of complexity are being considered INFN has requested a CM design by Nov 2005.

13 Data Access: ROOT + RDBMS Model histograms Calibrations Geometries Run/File Catalog Trees Event Store ROOT files Oracle MySQL

14 Computing Deployment: Implementation #1 Concentrated computing – Will be considered only in the case the CPU’s needed would exceed PSI’s capabilities – Easiest implementation – It requires the least manpower for maintenance – PROOF is essential for Analysis and DQM

15 Computing Deployment: Implementation #2 Distributed computing – Posting of the data/Catalog synchronization – Data distribution dependent upon connection speed – PSI -> INFN OK (3 MB/sec) – PSI -> Japan needs GRID Department    Desktop PSI – Tier 0 (PSI - Tier 1) Tier 1 X Y Z 622 Mbps 2.5 Gbps 622 Mbps 155 mbps Tier2 Lab a Uni b Lab c Uni n  "Transparent" user access to applications and all data

16 Processing Flow Ana DPD Raw Calib & Reco 1 ESD AOD Tag Reco Reproc MC ESD AOD Tag

17 Computing Model Configuration

18 How We Will Proceed 1. Decide for single-Tier or multi-Tier CM (depends heavily on CPU+Storage needed) 2. PSI is OK for Tier-0 (existing infrastructure + Horizon Cluster) 3. Find the candidate sites for Tier-1 Requirement for a Tier FTE for job running, control FTE for data export, catalog maintenance and DQM 3. If GRID is needed, FTE for GRID maintenance 4. Software installation would be responsibility of the Offline group (after startup)

19 Computing Model Status Started interaction with PSI’s Computing Center (J. Baschnagel and coll.) PSI CC survey was very encouraging: Physical space and electrical power not a limitation However, CC is running near 80% of Cooling and Clean Power capabilities Relevant services (backup, connection to exp. Area) are OK Alternative options are being considered (setup the farm in the experimental area and use the CC only for backup) Horizon Cluster PSI stand well as Tier-0 candidate (in a MultiTier CM. It might have enough resources to fulfill all MEG’s computing requirement

20 Status of the Software

21 Software Organization Development of Montecarlo (G3+Fortran) and Calibration/Reconstruction (VMC+ROOT) will initially proceed in parallel Migration will eventually occur later (end 2005) Montecarlo group is in charge of the simulation (Geometry+Event Generator) Offline group in charge of the rest (architecture, framework, reconstruction, computing, etc.) Detector specific code developed along with Detector Experts Present analyses for R&D within the ROME environment (same as Online)

22 Status of the Montecarlo: DCH (P. Cattaneo) Progress – Implementation of cable duct and various mechanical supports – Reorganization of the GEANT volume inside the chambers – First integration of the time to distance profile from GARFIELD Next – Completing integration of GARFIELD profiles – Calculation of time profiles for signals – Effect on signals of electrode patterns – Electronic simulation – Signal digitization – Improving details cards, cables

23 Status of the Montecarlo: EMCAL Progress – Pisa model: implement new geometry – Pisa model: new tracking model (not based on GTNEXT) – Tokio model: GEANT based (GNEXT) photon tracking including Refraction PMT quartz window Refraction PMT holder Tokio model: absorption Tokio model: scattering – Tokio model: ZEBRA output Number of p.e. in each PMT PMT position (to be replaced by MySQL/ XML) – Tokio model: ZEBRA2NTuple converter Next – Update geometry and complete integration between Pisa and Tokio – Signal digitization – ZEBRA output for hit timing

24 Status of the Montecarlo: TOF Progress – Implementation of the actual geometry: tapered slanted bar, square fibers – Addition of phototubes, photodiodes and related mechanics Next – Generation of photons in the bars/fiber – Propagation of photons to the PMT & photodiodes – Electronics and PMT/ photodiodes simulation – Improving details of material distribution, e.g. better PMT, – cables – Add mechanical support – Signal digitization

25 Status of the Montecarlo: more Graphics display of geometry only (no event) – Some zooming capability – Possibility of displaying a single volume – Addition of phototubes, photodiodes and related mechanics Beam, Target and Trigger – Preliminary approach outside GEM

26 Softwares used in LP beam test (R. Sawada) Midas Data taking Slow control Data logging Run control ROME + ROOT Online monitoring Offline analysis MySQL Channel information Geometry information Calibration constants

Proceedure of data take BoR EoR Event run number time trigger mode Configuration ID trigger mode channel info. geometry info. calibration data Histogram Tree (Ntuple) time # of events calibration

Proceedure of Offline Analysis BoR EoR Event trigger mode channel info. geometry info. calibration data processed data calibration Bofore a Run calibration run number

29 Status of the Offline Project Preliminary Offline architecture approved by MEG (June 2004) INFN – CSN1 also approved the Project (Sept. 2004) Funds have been provided for installing a minifarm in Lecce (22 kEUR) Real work started November 2004

30 Immediate Tasks and Objectives Setup a system to test the Offline code DAQ Prompt Calib Reco farm Online Disk Server Staging to tape Prototype the Main Program Modules Classes Steering program FastMC Test functionalities & performance Core Offline System Development RDBMS

31 Manpower Estimate Available (FTE) +1 waiting for funds

32 The People Involved Core Offline: – Coordinator: Gatto – Steering Program: Di Benedetto, Gatto – FastMC: Mazzacane, Tassielli – TS Implementation: Chiri – Interface to raw data: Chiri, Di Benedetto, Tassielli – Calibration Classes & Interface to RDBMS: Barbareschi* – Trigger: Siragusa *Will start on Mar Detector experts: – LXe: Signorelli, Yamada, Savada – TS: Schneebeli (hit), Hajime (Pattern), Lecce (Pattern) – TOF: Pavia/Genova – Trigger: Nicolo’ (Pisa) – Magnet: Ootani All <100% All 100% (except for Chiri) Montecarlo: – LXe: Cattaneo, Cei, Yamada All <100%

33 Milestones Offline 1. Start-up: October Start deploying the minifarm: January First working version of the framework (FastMC + Reconstruction): April 2005 (badly needed to proceed with the Computing Model) 4. Initial estimate of CPU/storage needed: June Computing Model: November MDC: 4 th quarter 2005 Montecarlo 1. Include the calorimeter in gem 2. Keep the existing MC in the Geant3 framework. Form a panel to decide if and how to migrate to ROOT: 4 th quarter 2005

34 Milestones Offline 1. Start-up: October Start deploying the minifarm: January First working version of the framework (FastMC + Reconstruction): April 2005 (badly needed to proceed with the Computing Model) 4. Initial estimate of CPU/storage needed: June Computing Model: November MDC: 4 th quarter 2005 Montecarlo 1. Include the calorimeter in gem 2. Keep the existing MC in the Geant3 framework. Form a panel to decide if and how to migrate to ROOT: 4 th quarter 2005

35 Conclusions Offline project approved by MEG and the INFN Offline group has consolidated (mostly in Lecce) Work has started Minifarm for testing the software is being setup Definition of the Computing Model is under way Montecarlo is on its way to a detailed description of the detector

36 Backup Slides

37 Compare to the others

38 Data Model (Monarc) ESD (Event Summary Data) – contain the reconstructed tracks (for example, track pt, particle Id, pseudorapidity and phi, and the like), the covariance matrix of the tacks, the list of track segments making a track etc… – AOD (Analysis Object Data) – contain information on the event that will facilitate the analysis (for example, centrality, multiplicity, number of positrons, number of EM showers, and the like). Tag objects – identify the event by its physics signature (for example, a cosmic ray) and is much smaller than the other objects. Tag data would likely be stored into a database and be used as the source for the event selection. DPD ( Derived Physics Data) – are constructed from the physics analysis of AOD and Tag objects. – They will be specific to the selected type of physics analysis (ex: mu->e gamma) – Typically consist of histograms or nt-uple-like objects. – These objects will in general be stored locally on the workstation performing the analysis, thus not add any constraint to the overall data-storage resources

39 Recostruction Structure Detector Class Global Reco Detector tasks Run Class Run Manager Detector Class DCH Detector tasks ROOT Data Base Tree Branches One or more Global Objects execute a list of tasks involving objects from several detectors Detector Class EMC Detector tasks Run Class MC Run Manager Detector Class TOF Detector tasks The Run manager executes the detector objects in the order of the list Each detector executes a list of detector tasks On demand actions are possible but not the default

40 Responsibilities & Tasks (all software) Detector experts: – LXe: Signorelli, Yamada, Savada – DC: Schneebeli (hit), Hajime (Pattern), Lecce (Pattern) – TC: Pavia/Genova – Trigger: Nicolo’ (Pisa) – Magnet: Ootani

41 Manpower Estimate (framework only) Available at Lecce Job posted: apply now 4 from Naples from Lecce +1 waiting for funds

42