Computing Overview Topics here: CSA lessons (briefly) PADA

Slides:



Advertisements
Similar presentations
Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
Advertisements

DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
December Pre-GDB meeting1 CCRC08-1 ATLAS’ plans and intentions Kors Bos NIKHEF, Amsterdam.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
LHCb Quarterly Report October Core Software (Gaudi) m Stable version was ready for 2008 data taking o Gaudi based on latest LCG 55a o Applications.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
Claudio Grandi INFN Bologna CMS Operations Update Ian Fisk, Claudio Grandi 1.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
ATLAS Metrics for CCRC’08 Database Milestones WLCG CCRC'08 Post-Mortem Workshop CERN, Geneva, Switzerland June 12-13, 2008 Alexandre Vaniachine.
CMS STEP09 C. Charlot / LLR LCG-DIR 19/06/2009. Réunion LCG-France, 19/06/2009 C.Charlot STEP09: scale tests STEP09 was: A series of tests, not an integrated.
Offline Coordinators  CMSSW_7_1_0 release: 17 June 2014  Usage:  Generation and Simulation samples for run 2 startup  Limited digitization and reconstruction.
1 M. Paganoni, HCP2007 Computing tools and analysis architectures: the CMS computing strategy M. Paganoni HCP2007 La Biodola, 23/5/2007.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
LCG CCRC’08 Status WLCG Management Board November 27 th 2007
All CMS 30 May07 TSV1 All CMS 3 30 May 2007 LHC machine CMS Progress Overall Schedule.
Claudio Grandi INFN Bologna CERN - WLCG Workshop 13 November 2008 CMS - Plan for shutdown and data-taking preparation Claudio Grandi Outline: Global Runs.
The CMS Computing System: getting ready for Data Analysis Matthias Kasemann CERN/DESY.
Report from GSSD Storage Workshop Flavia Donno CERN WLCG GDB 4 July 2007.
OPERATIONS REPORT JUNE – SEPTEMBER 2015 Stefan Roiser CERN.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
LCG Service Challenges SC2 Goals Jamie Shiers, CERN-IT-GD 24 February 2005.
Victoria, Sept WLCG Collaboration Workshop1 ATLAS Dress Rehersals Kors Bos NIKHEF, Amsterdam.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
1 September 2007WLCG Workshop, Victoria, Canada 1 WLCG Collaboration Workshop Victoria, Canada Site Readiness Panel Discussion Saturday 1 September 2007.
21 Mar 2007ALICE - Paul Dauncey1 ALICE CR07: Installation and Commissioning Paul Dauncey Material from: L. Leistam: “ALICE Status” W. Riegler: “Commissioning.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
Monthly video-conference, 18/12/2003 P.Hristov1 Preparation for physics data challenge'04 P.Hristov Alice monthly off-line video-conference December 18,
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
Operations Coordination Team Maria Girone, CERN IT-ES GDB, 11 July 2012.
1 M. Paganoni, 17/1/08 Modello di calcolo di CMS M. Paganoni Workshop Storage T2 - 17/01/08.
Maria Girone, CERN CMS Status Report Maria Girone, CERN David Lange, LLNL.
Project Execution Methodology
The CMS Experiment at LHC
WP18, High-speed data recording Krzysztof Wrona, European XFEL
CLAS12 DAQ & Trigger Status
LCG Service Challenge: Planning and Milestones
August ICA Agenda Time Topic 8:00 – 8:15
Technical Coordination Parallel Session November 2011, FNAL
DPG Activities DPG Session, ALICE Monthly Mini Week
LHC experiments Requirements and Concepts ALICE
Data Challenge with the Grid in ATLAS
INFN-GRID Workshop Bari, October, 26, 2004
Database Readiness Workshop Intro & Goals
for the Offline and Computing groups
Main next computing activities
Update on Plan for KISTI-GSDC
Status and Prospects of The LHC Experiments Computing
CMS transferts massif Artem Trunov.
LHCb Software & Computing Status
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
Olof Bärring LCG-LHCC Review, 22nd September 2008
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Francesco Forti University and INFN, Pisa
T1 visit to IN2P3 Computing
US ATLAS Physics & Computing
Preparations for the CMS-HI Computing Workshop in Bologna
N. De Filippis - LLR-Ecole Polytechnique
R. Graciani for LHCb Mumbay, Feb 2006
UK Link Technology Refresh
WLCG Collaboration Workshop: Outlook for 2009 – 2010
ATLAS DC2 & Continuous production
The ATLAS Computing Model
The LHCb Computing Data Challenge DC06
Presentation transcript:

Computing Overview Topics here: CSA lessons (briefly) PADA DESY CMS meeting 20.12.07 Topics here: CSA lessons (briefly) PADA CCRC’08 = CSA08 Matthias Kasemann December 2007 December 20, 2007 DESY CMS meeting

CSA07 Goals Test and validate the components the CMS Computing Model in a simultaneous exercise the Tier-0, Tier-1 and Tier-2 workflows Test the CMS software: particularly the reconstruction and HLT packages Test the CMS production systems at 50% scale of expected 2008 operation workflow management, data management, facilities, transfers Test the computing facilities and mass storage systems. Demonstrate that data will transfer between production and analysis sites in a timely way. Test the Alignment and Calibration stream (AlcaReco) Produce, deliver and store AODs + skims for analysis by physics groups December 20, 2007 DESY CMS meeting

Prompt Reconstruction CSA07 Workflows Prompt Reconstruction CASTOR HLT TIER-0 CAF Calibration Express-Stream Analysis 300MB/s Re-Reco Skims TIER-1 TIER-1 TIER-1 TIER-1 20-200MB/s ~10MB/s Simulation Analysis TIER-2 TIER-2 TIER-2 TIER-2 December 20, 2007 DESY CMS meeting

Preparing for the CSA07 (Jul-Sep) CMSSW - software releases organized by offline team Releases are tested by data operations teams Distributed and installed to the sites (This is not an easy process.) Steps for preparing data for physics (pre-challenge workflows): Generation and Simulation with Geant4 (at the Tier-2 centers) Digitization Digi2RAW - format change to look like data input to HLT HLT processing Data are split into 7 Primary Data Sets (PDS) based on the HLT information This was a big addition in CSA07. The data samples more accurately reflect what will come from the detector, but are harder to produce. December 20, 2007 DESY CMS meeting

Preparing for the CSA07 (Jul-Sep) Planned Workflows for the Challenge: Reconstruction - HLT + RECO output (~1 MB) AOD production - (200 kB) Skims for physics analysis at the Tier-1 centers Re-Reco (and redoing AOD production/skims) at the Tier-1 centers Analysis at the Tier-2 centers Lessons from CSA07 preparations: It turned out that there was insufficient time for testing the components since some of the components were coming at the latest moment.. CSA08: We have to devote more time for testing the components December 20, 2007 DESY CMS meeting

December 20, 2007 DESY CMS meeting

MC Production summary Substantial (… more resources used) December 20, 2007 DESY CMS meeting

CSA07 Issues and Lessons There are clearly areas that are going to need development Need to work on the CMSSW application Reduce the number of workflows (in 170, 180 and 200) Reduce the memory footprint to increase the number of events we can run and increase the available resources Goal: CMSSW applications should stay in 1 GB memory Several area should be improved Access and manipulation of IOV constant (over Xmas) HLT data model (on going) New huge increase in memory seen in 170 to be address immediately (mainly in DPG’s code) December 20, 2007 DESY CMS meeting

CSA07 Issues and Lessons Increase the speed of IO on mass storage test using new ROOT version Improve our testing and validation procedures for the applications and workflows. Reduce event size RAW/DIGI and RECO size AOD size Mini-workshop with physics and DPG Groups on Jan 28/29 (CERN) Two task forces has been created in order to prepare this workshop RECO Task Force chair (Shahram Rahatlou) Analysis Task Force chair (Roberto Tenchini) FW support for handing of RAW, RECO versus FEVT (This is foreseen for version 2_0_0) December 20, 2007 DESY CMS meeting

CSA07 Issues and Lessons Need to work on the CMS Tools Augment the Production tools to be able to better handle continuous operations Roll back to known good points. Modify workflows more simply Increase the speed of Bookkeeping System under specific load conditions Optimize the data transfers in PhEDEx for data availability Improve the analysis Tool (CRAB) Planning a Workshop in January 21-25 2008 (Lyon) http://indico.cern.ch/conferenceTimeTable.py?confId=24877 Goals: Review Data and Workload management components Improvement integration (communication) between operation and development teams Will include also Tier0 components Define work plan for 2008 December 20, 2007 DESY CMS meeting

CSA07 Issues and Lessons Facility Lessons: We learned a lot about operating Castor and dCache under load Need to improve the rate of file opens Need to decrease the rate of errors. Need to improve the scalability of some components Need to work on the stability of services at CERN and Tier-1 centers Need to work on the transfer quality when the farms are under heavy processing load General lessons: Much work is needed to achieve simultaneous, sustainable and stable operations December 20, 2007 DESY CMS meeting

PADA: processing and data access taskforce Draft Mandate: Integrate developments and services to bring our centers and services to production quality for processing and analysis The Processing And Data Access Task Force is an initiative in the Integration Program Designed to transition services developed in Offline to Operations Elements of integration and testing for Production, Analysis, and Data Management tools Designed to ensure services and sites used in operations are production quality Elements in the commissioning program for links and sites Verify that items identified in the CSA07 are solved Development work is primarily in offline, but verification in Integration Plan is: To build on the expertise of the distributed MC production teams, extend scope We need the expertise in proximity of the centers to help us here For 2008 we want to make this a recognized service contribution in the MoA scheme, Initial time frame: 1 year until we have seen the first data We need to define steps, milestones, recruit people, hope for MC-OPS, DDT, .... December 20, 2007 DESY CMS meeting

PADA tasks + schedule December 20, 2007 DESY CMS meeting

Final Check before Data taking starts: CCRC’08 = CSA08CMS A combined challenge by all Experiments must be used to demonstrate the readiness of the WLCG Computing infrastructure before start of data taking at a scale comparable to the data taking in 2008. CMS fully supports the plan, to execute this CCRC in two phases: a set of functional tests in February 2008 the final challenge in May 2008 at 100% scale, starting with the readout of the experiment We must do this challenge as WLCG collaboration: Centers and Experiments together Combined planning has started: Mailing list created: wlcg-ccrc08@cern.ch Agenda pages: Phone conference every Monday afternoon Monthly session in pre-GDB meeting December 20, 2007 DESY CMS meeting

CMS CCRC’08 Schedule Phase 1 - February 2008: (proposed: 4-29.2.2008) Possible scenario: blocks of functional tests, Try to reach 2008 scale for tests at… Phase 2: - May 2008: (proposed: 5-30.5.2008) Full workflows at all centers executed simultaneously by all 4 LHC experiments Use data from cosmics data run, add artificial load to reach 100% Duration of challenge: 1 week setup, 4 weeks challenge December 20, 2007 DESY CMS meeting

1) Detector Installation, Commissioning & Operation 2) Preparation of Software, Computing & Physics Analysis Aug Sep Oct Nov Dec Jan Feb Mar Apr May S/w Release 1_6 (CSA07) V36 Schedule (Nov’07) CSA07 Cooldown of Magnet: Test S/w Release 1_7 (CCR_0T, HLT Validation) Tracker Insertion 2007 Physics Analyses First Results Out CMS Cosmic Run CCR_0T Several short periods Dec-Mar) Last Heavy Element Lowered Test Magnet at low current S/w Release 1_8 (Lessons of ‘07) Functional Tests CSA08 (CCRC) Beam-pipe Closed and Baked-out S/w Release 2_0 (CCR_4T, Production startup MC samples) 1 EE endcap Installed, Pixels installed MC Production for Startup Cosmic Run CCR_4T CSA08 (CCRC) Combined Computing Readiness Challenge Master Contingency 2nd ECAL Endcap Ready for Installation end Jun’08 December 20, 2007 DESY CMS meeting

CCRC’08 Phase 1: February 2008 Goals for CMS: Verify solutions to CSA07 issues and lessons - don’t repeat CSA07 where solution is not ready Attempt to reach ‘08 scale on individual tests - don’t repeat CSA07 where no increase in scale possible Guiding principles: CCRC’08-Phase 1 will be a Computing&Software challenge no coupling of other deliverables to CCRC’08/1 tests Cosmics run, MC production and Physics Analysis have priority and cannot be interrupted by CCRC’08 tests for long We defined blocks of tests, which stress a specific service or workflow Tests should be as independent from each other as possible Tests should be done in parallel where possible Individual test is considered successful if sustained for (3-5) days Where full ‘08 scale is not possible (hardware) tests are scaled down to hardware limit December 20, 2007 DESY CMS meeting

Status of preparation Need to provide requirements to sites Draft plan of of tests available Goals of tests agreed by computing Area or responsibility and coordination is defined Tests are to be coordinated and agreed within CMS Many tests are to be scheduled together with other VO’s (ATLAS, …) to reach scale of stress test CCRC’08 meeting next Thursday, 20.12, 16:00 Plan details of tests Propose metrics Define constraints and specify resources required Propose schedule of test Need to provide requirements to sites December 20, 2007 DESY CMS meeting

Planned blocks of tests Data recording at CERN 1a) readout from P5, use HLT, w. stream definition, use Storage Manager, transfer to T0, perform repacking, write to CASTOR Goal: verify dataflow for CMS 1b) CASTOR data archiving test Goal: verify CASTOR performance at full CMS and ATLAS rate 2. Processing at T0 at high rate Goal: verify T0 performance under CMS + ATLAS load 3. CERN data export to T1 at full / high rate Goal: verify CERN export and T1 import performance under CMS + ATLAS load 4. T1 data handling and processing Goal: verify full CMS T1 re-processing workflow in presence of ATLAS load at T1’s December 20, 2007 DESY CMS meeting

Planned blocks of tests 5 Data Transfer performance tests Goal: verify T1/T2 export + import performance under CMS +ATLAS load 5.1 T1 - T1 data transfer at real rates 5.2 T1 - T2 data transfer at real rates 5.3 T2 - T1 data transfer at real rates Data Transfer tests (5.1-3) should be done individually and then together Monte Carlo Production and Analysis at Tier-2’s Goal: Verify Pile-up and FastSim MC Production Goal: Scale tests of Analysis Jobs Data transfer request system tests Goal: verify Data Transfer and SE’s at Tier-2 centers 8. CAF tests: (depending on CAF infrastructure schedule) Goal: verify basics CMS use cases at scale December 20, 2007 DESY CMS meeting

Summary (1/2) In CSA07 a lot was learned and a lot was achieved.. We hit most of metrics - but separately and intermittently Several steps accomplished simultaneously Many workflow steps hit metric routinely Now work on accomplishing all steps simultaneously, and providing stability in a sustainable way. Global connectivity between T1-T2 sites is still an important issue. The DDT task force has been successful in increasing the # of working links. This effort must continue and work must be done to automate the process of testing/commissioning the links. We still have to increase the number of people involved in facilities, commissioning and operations. Some recent actions: New (2nd) L2 appointed to lead facility operations (based at CERN) New Production And Data Access (PADA) Task Force starting - will include some of the people from DDT task force and MC production teams. December 20, 2007 DESY CMS meeting

Summary (2/2) ~ 200M Events processed and re-processed Calibration, MC production, Reconstruction, skimming, merging all tested successfully. Still need time to test the analysis model. CSA07 Goals for providing data for physics will be accomplished … albeit delayed due to schedule slips Processing continues to complete the data samples for physics and detector studies. We are keeping the challenge infrastructure alive and trying to keep it stable, going forward... Continue to support global detector commissioning and physics studies. We have to prepare for the ‘Combined Computing Readiness Challenge’, CCRC’08  CSA08 Without testing the software and infrastructure we are not prepared… December 20, 2007 DESY CMS meeting