The June Software Review David Lawrence, JLab Feb. 16, 2012 2/16/121Preparations for June Software Review David Lawrence.

Slides:



Advertisements
Similar presentations
WHAT IS AN OPERATING SYSTEM? An interface between users and hardware - an environment "architecture ” Allows convenient usage; hides the tedious stuff.
Advertisements

Software Process Models
Software Summary Database Data Flow G4MICE Status & Plans Detector Reconstruction 1M.Ellis - CM24 - 3rd June 2009.
1: Operating Systems Overview
CompuNet Grid Computing Milena Natanov Keren Kotlovsky Project Supervisor: Zvika Berkovich Lab Chief Engineer: Dr. Ilana David Spring, /
The Software Product Life Cycle. Views of the Software Product Life Cycle  Management  Software engineering  Engineering design  Architectural design.
Trigger and online software Simon George & Reiner Hauser T/DAQ Phase 1 IDR.
May 2010 Graham Heyes Data Acquisition and Analysis group. Physics division, JLab Data Analysis Coordination, Planning and Funding.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Offline Software Status Jan. 30, 2009 David Lawrence JLab 1.
The June Software Review David Lawrence, JLab Feb. 16, 2012.
Computer System Architectures Computer System Software
Software Overview David Lawrence, JLab Oct. 26, 2007 David Lawrence, JLab Oct. 26, 2007.
Status of Hall C 6 GeV Analysis Software Robust Fortran/CERNLIB code, “ENGINE”, for analysis of HMS/SOS coincidence and single arm experiments that has.
Introduction to Hall-D Software February 27, 2009 David Lawrence - JLab.
Online Data Challenges David Lawrence, JLab Feb. 20, /20/14Online Data Challenges.
REVIEW OF NA61 SOFTWRE UPGRADE PROPOSAL. Mandate The NA61 experiment is contemplating to rewrite its fortran software in modern technology and are requesting.
LOGO OPERATING SYSTEM Dalia AL-Dabbagh
Operating System Review September 10, 2012Introduction to Computer Security ©2004 Matt Bishop Slide #1-1.
An Introduction to Software Architecture
The GlueX Collaboration Meeting October 4-6, 2012 Jefferson Lab Curtis Meyer.
Software Engineering Chapter 8 Fall Analysis Extension of use cases, use cases are converted into a more formal description of the system.Extension.
Status of Hall C 6 GeV Analysis Software Robust Fortran/CERNLIB code, “ENGINE”, for analysis of HMS/SOS coincidence and single arm experiments that has.
Offline Coordinators  CMSSW_7_1_0 release: 17 June 2014  Usage:  Generation and Simulation samples for run 2 startup  Limited digitization and reconstruction.
GlueX Software Status April 28, 2006 David Lawrence, JLab.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Assessing the influence on processes when evolving the software architecture By Larsson S, Wall A, Wallin P Parul Patel.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
Computers Operating System Essentials. Operating Systems PROGRAM HARDWARE OPERATING SYSTEM.
JANA and Raw Data David Lawrence, JLab Oct. 5, 2012.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Page 1 May 10, 2011 IT for the 12 GeV Era 2011 Review Review Closing Summary.
November 2013 Review Talks Morning Plenary Talk – CLAS12 Software Overview and Progress ( ) Current Status with Emphasis on Past Year’s Progress:
Hall-D/GlueX Software Status 12 GeV Software Review III February 11[?], 2015 Mark Ito.
1: Operating Systems Overview 1 Jerry Breecher Fall, 2004 CLARK UNIVERSITY CS215 OPERATING SYSTEMS OVERVIEW.
GDB Meeting - 10 June 2003 ATLAS Offline Software David R. Quarrie Lawrence Berkeley National Laboratory
Thomas Jefferson National Accelerator Facility Page Hall B:User Software Contributions Gerard Gilfoyle University of Richmond 12 GeV Upgrade Software Review.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
Application Software System Software.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
The JANA Reconstruction Framework David Lawrence - JLab May 25, /25/101JANA - Lawrence - CLAS12 Software Workshop.
Monitoring Update David Lawrence, JLab Feb. 20, /20/14Online Monitoring Update -- David Lawrence1.
Thomas Jefferson National Accelerator Facility Page 1 Overview Talk Content Break-out Sessions Planning 12 GeV Upgrade Software Review Jefferson Lab November.
General requirements for BES III offline & EF selection software Weidong Li.
DAQ Status & Plans GlueX Collaboration Meeting – Feb 21-23, 2013 Jefferson Lab Bryan Moffit/David Abbott.
November 1, 2004 ElizabethGallas -- D0 Luminosity Db 1 D0 Luminosity Database: Checklist for Production Elizabeth Gallas Fermilab Computing Division /
A computer contains two major sets of tools, software and hardware. Software is generally divided into Systems software and Applications software. Systems.
GlueX Software Status + Framework Development David Lawrence JLab September 19, /19/081Software Status -- David Lawrence, JLab.
PCAP Close Out Feb 2, 2004 BNL. Overall  Good progress in all areas  Good accomplishments in DC-2 (and CTB) –Late, but good.
SoLID simulation thoughts Zhiwen Zhao 2015/04/02.
Background Computer System Architectures Computer System Software.
Status of Hall C 6 GeV Analysis Software Robust Fortran/CERNLIB code, “ENGINE”, for analysis of HMS/SOS coincidence and single arm experiments that has.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
VI/ CERN Dec 4 CMS Software Architecture vs Hybrid Store Vincenzo Innocente CMS Week CERN, Dec
Update on CHEP from the Computing Speaker Committee G. Carlino (INFN Napoli) on behalf of the CSC ICB, October
October 19, 2010 David Lawrence JLab Oct. 19, 20101RootSpy -- CHEP10, Taipei -- David Lawrence, JLab Parallel Session 18: Software Engineering, Data Stores,
Analysis Model Zhengyun You University of California Irvine Mu2e Computing Review March 5-6, 2015 Mu2e-doc-5227.
Please fill in my session feedback form available on each chair. SPSCairo Welcome.
MAUS Status A. Dobbs CM43 29 th October Contents MAUS Overview Infrastructure Geometry and CDB Detector Updates CKOV EMR KL TOF Tracker Global Tracking.
1 GlueX Software Oct. 21, 2004 D. Lawrence, JLab.
David Lawrence JLab May 11, /11/101Reconstruction Framework -- GlueX Collab. meeting -- D. Lawrence.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
David Adams ATLAS Hybrid Event Store Integration with Athena/StoreGate David Adams BNL March 5, 2002 ATLAS Software Week Event Data Model and Detector.
GLAST LAT ProjectNovember 18, 2004 I&T Two Tower IRR 1 GLAST Large Area Telescope: Integration and Test Two Tower Integration Readiness Review SVAC Elliott.
CMS High Level Trigger Configuration Management
MICE Collaboration Meeting Saturday 22nd October 2005 Malcolm Ellis
Conditions Data access using FroNTier Squid cache Server
OO-Design in PHENIX PHENIX, a BIG Collaboration A Liberal Data Model
Chapter 2: Building a System
Presentation transcript:

The June Software Review David Lawrence, JLab Feb. 16, /16/121Preparations for June Software Review David Lawrence

… the short summary… Software review is scheduled for June 7-8, 2012 – Review committee not yet formed Purpose is to assess the overall ability of the software to analyze the data at the time data taking begins – No official charge yet David Lawrence and Matt Shepherd are lead organizers of Hall-D preparations (with considerable help from Mark Ito and Curtis Meyer) Information being gathered on wiki page: – 2/16/12Preparations for June Software Review David Lawrence2

(some history) May 10, 2011 IT Review An internal JLab review of IT readiness was done on May 10 th, This was intended as a “warm up” for the review coming this summer. Charge to the review panel: We request that the review panel address the following points for IT in the 12 GeV era: – An assessment of the state of current software and systems developments An assessment of planning for bringing all software to a suitable level of maturity, including software testing for correctness and performance – An assessment of planning for an evolution of computing, networking and storage capacity and performance to address the needs of detector simulation and data analysis – An assessment of the IT infrastructure to meet requirements including support for other areas, e.g. accelerator, light source, theory, operations – An assessment of the quality and effectiveness of the management of the major efforts to prepare – As assessment of the resources, budget and staffing, to meet the needs of the program one day review afternoon session focused on non-ENP* software Management Information Systems Networking and Infrastructure Accelerator Controls Hall-D had one 25-minute talk given by Mark Ito. *ENP=Experimental Nuclear Physics 2/16/123Preparations for June Software Review David Lawrence

From May IT Review closeout Software: No common process for defining requirements, no common management structure 4 halls not sharing much software Hall D: – D’s requirements not as well defined as other halls – Software head seems to have insufficient authority to direct software development priorities (i.e. software architect) – 2 FTE seems too small for 40% of effort planned for Jefferson Lab – Hall D Offsite computing & networking requirements nebulous Halls do not yet have robust plans for testing and reviewing readiness to operate. Identification of risks, and addressing risks, still needs to be done … these items were specific to the experimental halls… -- We disagreed with some of the findings in the close-out and first draft of committee report. These were communicated to the committee but a final draft of the report was never publically released -- 2/16/124Preparations for June Software Review David Lawrence

Translation Reconstruction DST Generator primary reconstruction program Raw Data Analysis DST Translation Table DB Translation Table DB CalibrationDB DST DB Configuration DB Sim. Data Simulation + Digitization Calibration Programs Calibration Programs Calibration Programs Calibration Programs Calibration Programs Calibration Programs Calibration Programs Calibration Programs Web Interface scripts The software effort centers around reconstruction where a large fraction of the effort is spent. Reconstruction requires inputs from databases to 1.) translate the geographic DAQ info into detector ids and 2.) convert digitized values into physical units. Software is used to fill these databases with the bulk of that effort focused on the calibration programs. Simulated data is required for the Amplitude Analysis. It does not require translation, but will be processed by the same reconstruction software as is used for the real data. The configuration DB will hold information used to configure the online systems prior to data taking. The conditions DB will have values read from the online systems streamed into it during data taking. Conditions DB For simplicity, not all connections are shown. (e.g. arrow from “Raw Data” to “Calibration Programs” ) Rough Diagram of GlueX Software 2/16/125Preparations for June Software Review David Lawrence

Hall-D Software Activity Schedule Activity schedule adopted for BIA (Baseline Improvement Activity) schedule. Tracking of BIA stopped in 2009 Minor tweaks including addition of a couple of lines (e.g. Data Format under DST Generation) % complete column added and based purely on my “engineering judgment” Responsible Persons column added 2/16/126Preparations for June Software Review David Lawrence

Overall status of Hall-D Software Activities 2/16/127Preparations for June Software Review David Lawrence

Hall-D Software Development 2 years Steady development over the last few years. (Repository restructuring 2 years ago limits the reach of this plot). Includes GlueX simulation and reconstruction only (no CalibDB or JANA) 2/16/128Preparations for June Software Review David Lawrence * First entry into original cvs Hall-D repository in May, 1998 by Joe Manak

Hall-D reconstruction factory call graph auto-generated by JANA Requests for objects originate here Objects read from file enter here 2/16/12 9 Preparations for June Software Review David Lawrence

Detailed profiling of entire reconstruction chain provided by JANA framework for diagnosing where time is spent. Profiling at subroutine level can be done using standard C/C++ debugging utilities. 2/16/12 10 Preparations for June Software Review David Lawrence

Estimated Resource Requirements (from Mark’s spreadsheet prepared for May 2011 IT review) 2/16/1211

Preparations for June Software Review David Lawrence Estimated Resource Requirements (from Mark’s spreadsheet prepared for May 2011 IT review) About 115 computers with 32 cores each will be needed just to keep up with time-averaged data acquisition + calibration Another 170 will be needed for simulation (+recon.) 2/16/1212

Software Sharing Among Halls Meeting was held on Jan. 26 th to discuss areas where halls could share software, minimizing duplication of effort. All halls were represented Rolf Ent asked halls to get together to discuss specific topics and explore sharing opportunities Two items were given to Halls B and D to discuss (a few others for all halls) : – Tracking Algorithms Multiple, organized discussions have taken place between primaries Hall-B has read-access to our repository and is using it as a reference as they develop their own tracking package – CLARA and JANA (see next few slides) 2/16/1213Preparations for June Software Review David Lawrence

“Loosely Coupled”: Allows multiple languages to be combined since each module is a separate process Data passed between modules by value Built-in ability to distribute reconstruction job over multiple computers (cloud) “Tightly Coupled”: Single language, all modules contained within a single process Data passed between modules by reference Utilizes external distributed computing mechanisms like the GRID and Auger Primary Differences between JANA and CLARA CLARA is designed to provide interactive access to a system of services hosted either on a single node or distributed over a cloud JANA is designed to make maximal use of a local, multi-core resource CLARAJANA 2/16/1214Preparations for June Software Review David Lawrence

Framework for event reconstruction Modular: allow easy replacement of one or more algorithms allow independent development of modules by separate groups Provides mechanism to parallelize reconstruction using multiple cores on the same computer Plugin mechanism to allow extension of existing functionality at run time Functionality common to both JANA and CLARA 2/16/1215Preparations for June Software Review David Lawrence

How JANA and CLARA might used in conjunction JANA could be used to implement CLARA services that need to be highly efficient. CLARA could be used to deploy JANA applications as shared services in a network distributed cloud computing environment. The primary benefit to CLAS12 users of integrating JANA-based components into a CLARA-based system could be overall faster reconstruction for a fixed set of resources. The primary benefit to Hall-D users of wrapping JANA-based programs as CLARA services would be gaining an interactive distributed computing environment that could provide a faster simulation/analysis cycle for specific studies. 2/16/1216Preparations for June Software Review David Lawrence

2/16/12 Preparations for June Software Review David Lawrence Manpower Use standard COCOMO model to estimate man-years put into CLAS offline software Estimate was ~53 man-years for core CLAS offline GlueX was estimated to need ~40 man-years to be ready for start of operations from GlueX-doc-767 (2007) ~28 FTE-years ~23 FTE-years It is estimated that we will need approx. 40 FTE-years of offline software effort total for GlueX* Estimate is that we have done ~ 50% of work for offline software* Remaining 20 FTE-years of estimated work is well matched with manpower commitments from collaboration* *every one of these estimates could be completely wrong 17

Summary Software review is scheduled for early June Focus will be on having offline software development on track to be ready for analysis by the start of data taking Integrated GlueX manpower seems to be well- matched with what is needed to meet this goal 2/16/1218Preparations for June Software Review David Lawrence

Backup Slides 2/16/1219Preparations for June Software Review David Lawrence

top Memory usage is between 3 GB and 4 GB for single process running with 48 processing threads CPU completely utilized in user space (like we want it!) Negligible time spent sleeping in mutex locks or system calls 6/15/1020GlueX Offline Software meeting - D. Lawrence

Testing on a 48-core “Magny Cours” Oct. 19, 2010 RootSpy -- CHEP10, Taipei -- David Lawrence, JLab21 Event reconstruction using 48 processing threads on a CPU with 48 cores generally scales quite well. Eventually, an I/O limit will be encountered Occasionally some problems with inexplicably lower rates. Program appears to simply run slower while not operating any differently. Unclear if this is due to hardware or Linux kernel Memory Usage vs. time while repeatedly running the 35 thread test. The marked area indicates one test where the program ran slower.