The Worldwide LHC Computing Grid WLCG Service Ramp-Up LHCC Referees’ Meeting, January 2007.

Slides:



Advertisements
Similar presentations
Exporting Raw/ESD data from Tier-0 Tier-1s Wrap-up.
Advertisements

CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
Les Les Robertson WLCG Project Leader WLCG – Worldwide LHC Computing Grid Where we are now & the Challenges of Real Data CHEP 2007 Victoria BC 3 September.
December Pre-GDB meeting1 CCRC08-1 ATLAS’ plans and intentions Kors Bos NIKHEF, Amsterdam.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
Les Les Robertson LCG Project Leader LCG - The Worldwide LHC Computing Grid LHC Data Analysis Challenges for 100 Computing Centres in 20 Countries HEPiX.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
SC4 Workshop Outline (Strong overlap with POW!) 1.Get data rates at all Tier1s up to MoU Values Recent re-run shows the way! (More on next slides…) 2.Re-deploy.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
SRM 2.2: tests and site deployment 30 th January 2007 Flavia Donno, Maarten Litmaath IT/GD, CERN.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
WLCG Service Schedule June 2007.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
WLCG Service Report ~~~ WLCG Management Board, 24 th November
2 Sep Experience and tools for Site Commissioning.
CCRC08-1 report WLCG Workshop, April KorsBos, ATLAS/NIKHEF/CERN.
John Gordon STFC-RAL Tier1 Status 9 th July, 2008 Grid Deployment Board.
LCG Service Challenges: Planning for Tier2 Sites Update for HEPiX meeting Jamie Shiers IT-GD, CERN.
LCG Service Challenges: Planning for Tier2 Sites Update for HEPiX meeting Jamie Shiers IT-GD, CERN.
Jürgen Knobloch/CERN Slide 1 A Global Computer – the Grid Is Reality by Jürgen Knobloch October 31, 2007.
Stefano Belforte INFN Trieste 1 Middleware February 14, 2007 Resource Broker, gLite etc. CMS vs. middleware.
LCG CCRC’08 Status WLCG Management Board November 27 th 2007
SC4 Planning Planning for the Initial LCG Service September 2005.
WLCG Planning Issues GDB June Harry Renshall, Jamie Shiers.
WLCG Service Report ~~~ WLCG Management Board, 7 th September 2010 Updated 8 th September
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
LCG Report from GDB John Gordon, STFC-RAL MB meeting February24 th, 2009.
The LHC Computing Environment Challenges in Building up the Full Production Environment [ Formerly known as the LCG Service Challenges ]
The CMS Computing System: getting ready for Data Analysis Matthias Kasemann CERN/DESY.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
DJ: WLCG CB – 25 January WLCG Overview Board Activities in the first year Full details (reports/overheads/minutes) are at:
ATLAS Computing Requirements LHCC - 19 March ATLAS Computing Requirements for 2007 and beyond.
Report from GSSD Storage Workshop Flavia Donno CERN WLCG GDB 4 July 2007.
LCG Service Challenges SC2 Goals Jamie Shiers, CERN-IT-GD 24 February 2005.
The Worldwide LHC Computing Grid Introduction & Housekeeping Collaboration Workshop, Jan 2007.
14/03/2007A.Minaenko1 ATLAS computing in Russia A.Minaenko Institute for High Energy Physics, Protvino JWGC meeting 14/03/07.
Victoria, Sept WLCG Collaboration Workshop1 ATLAS Dress Rehersals Kors Bos NIKHEF, Amsterdam.
SL5 Site Status GDB, September 2009 John Gordon. LCG SL5 Site Status ASGC T1 - will be finished before mid September. Actually the OS migration process.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
SRM v2.2 Production Deployment SRM v2.2 production deployment at CERN now underway. – One ‘endpoint’ per LHC experiment, plus a public one (as for CASTOR2).
Operations model Maite Barroso, CERN On behalf of EGEE operations WLCG Service Workshop 11/02/2006.
GDB, 07/06/06 CMS Centre Roles à CMS data hierarchy: n RAW (1.5/2MB) -> RECO (0.2/0.4MB) -> AOD (50kB)-> TAG à Tier-0 role: n First-pass.
8 August 2006MB Report on Status and Progress of SC4 activities 1 MB (Snapshot) Report on Status and Progress of SC4 activities A weekly report is gathered.
The Grid Storage System Deployment Working Group 6 th February 2007 Flavia Donno IT/GD, CERN.
WLCG Service Report Jean-Philippe Baud ~~~ WLCG Management Board, 24 th August
LCG Service Challenges: Progress Since The Last One –
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
Summary of SC4 Disk-Disk Transfers LCG MB, April Jamie Shiers, CERN.
1 September 2007WLCG Workshop, Victoria, Canada 1 WLCG Collaboration Workshop Victoria, Canada Site Readiness Panel Discussion Saturday 1 September 2007.
Top 5 Experiment Issues ExperimentALICEATLASCMSLHCb Issue #1xrootd- CASTOR2 functionality & performance Data Access from T1 MSS Issue.
1 S. JEZEQUEL- First chinese-french workshop 13 December 2006 Grid: An LHC user point of vue S. Jézéquel (LAPP-CNRS/Université de Savoie)
ATLAS Computing Model Ghita Rahal CC-IN2P3 Tutorial Atlas CC, Lyon
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
The Worldwide LHC Computing Grid WLCG Milestones for 2007 Focus on Q1 / Q2 Collaboration Workshop, January 2007.
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
Computing Operations Roadmap
“A Data Movement Service for the LHC”
The LHC Computing Environment
LCG Service Challenge: Planning and Milestones
Jan 12, 2005 Improving CMS data transfers among its distributed Computing Facilities N. Magini CERN IT-ES-VOS, Geneva, Switzerland J. Flix Port d'Informació.
Data Challenge with the Grid in ATLAS
Database Readiness Workshop Intro & Goals
An introduction to the ATLAS Computing Model Alessandro De Salvo
LHC Data Analysis using a worldwide computing grid
Overview & Status Al-Ain, UAE November 2007.
Presentation transcript:

The Worldwide LHC Computing Grid WLCG Service Ramp-Up LHCC Referees’ Meeting, January 2007

Ramp-Up Outline  The clear goal for 2007 is to be ready for first data taking ahead of the machine itself This translates to: –Dress Rehearsals in the 2 nd half of the year –Preparation for these in the 1 st half –Continuous service operation and hardening –Continual (quasi-continuous) experiment production Different views: –Experiment, site, Grid-specific, WLCG… Will focus on first and (mainly) last of these… –Other views, in particular site views, will come shortly

3 WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing of full data chain from DAQ to T-2 chain was major item from last CR Timely testing of full data chain from DAQ to T-2 chain was major item from last CR DAQ  T-0 still largely untested DAQ  T-0 still largely untested

Service Ramp-Up As discussed at last week’s WLCG Collaboration Workshop, much work has already been done on service hardening –Reliable hardware, improved monitoring & logging, middleware enhancements  Much still remains to be done – this will be an on-going activity during the rest of 2007 and probably beyond The need to provide as much robustness in the services themselves – as opposed to constant baby-sitting – is well understood There are still new / updated services to deploy in full production (see previous slide)  It is unrealistic to expect that all of these will be ready prior to the start of the Dress Rehearsals Foresee a ‘staged approach’ – focussing on maintaining and improving both service stability and functionality (‘residual services’)  Must remain in close contact with both experiments and sites on schedule and service requirements – these will inevitably change with time Draft of experiment schedule (from December 2006) attached to agenda Updated schedules presented last Friday during WLCG w/s (pointer)

5  Running continously throughout the year (increasing rate)  Simulation production  Cosmic ray data-taking (detector commissioning)  January to June:  Data streaming tests  February and May:  Intensive Tier0 tests  From February onwards:  Data Distribution tests  From March onwards:  Distributed analysis (intensive tests)  May to July:  Calibration Data Challenge  June to October  Full Dress Rehearsal  November:  GO! ATLAS 2007 Timeline

Stefano Belforte INFN Trieste 6 Timeline February  Deploy PhEDEx 2.5  T0-T1, T1-T1, T1-T2 independent transfers  Restart job robot  Start work on SAM  FTS full deployment March  SRM v2.2 tests start  T0-T1(tape)-T2 coupled transfers (same data)  Measure data serving at sites (esp. T1)  Production/analysis share at sites verified April  Repeat transfer tests with SRM v2.2, FTS v2  Scale up job load  gLite WMS test completed (synch. with Atlas) May  Start ramping up to CSA07 June

WLCG Milestones These high-level milestones are complementary to the experiment-specific milestones and more detailed goals and objectives listed in the WLCG Draft Service Plan (see attachment to agenda) –Similar to that prepared and maintained in previous years –Regularly reviewed and updated through LCG ECM –Regular reports on status and updates to WLCG MB / GDB  Focus is on real production scenarios & (moving rapidly to) end to end testing –Time for component testing is over – we learnt a lot but not enough! –Time before data taking is very short – let alone the dress rehearsals All data rates refer to the Megatable and to pp running Any ‘factors’, such as accelerator and/or service efficiency, are mentioned explicitly –N.B. ‘catch-up’ is a proven feature of the end-end FTS service

Q – Tier0 / Tier1s 1.Demonstrate Tier0-Tier1 data export at 65% of full nominal rates per site using experiment-driven transfers –Mixture of disk / tape endpoints as defined by experiment computing models, i.e. 40% tape for ATLAS; transfers driven by experiments –Period of at least one week; daily VO-averages may vary (~normal) 2.Demonstrate Tier0-Tier1 data export at 50% of full nominal rates (as above) in conjunction with T1-T1 / T1-T2 transfers –Inter-Tier transfer targets taken from ATLAS DDM tests / CSA06 targets 3.Demonstrate Tier0-Tier1 data export at 35% of full nominal rates (as above) in conjunction with T1-T1 / T1-T2 transfers and Grid production at Tier1s –Each file transferred is read at least once by a Grid job –Some explicit targets for WMS at each Tier1 need to be derived from above 4.Provide SRM v2.2 endpoint(s) that implement(s) all methods defined in SRM v2.2 MoU, all critical methods pass tests –See attached list; Levels of success: threshold, pass, success, (cum laude) –This is a requirement if production deployment is to start in Q2!

Q – Tier0 / Tier1s As Q1, but using SRM v2.2 services at Tier0 and Tier1, gLite 3.x-based services and SL(C)4 as appropriate, (higher rates? (T1 T1/2)) Provide services required for Q3 dress rehearsals –Includes, for example, production Distributed Database Services at required sites & scale Full detail to be provided in coming weeks…

Measuring Our Level of Success Existing tools and metrics, such as CMS PhEDEx quality plots, ATLAS DDM transfer status, provide clear and intuitive views  These plots are well known to the sites and provide a good measure of current status as well as showing evolution with time Need metrics for WMS related to milestone 3 –CMS CSA06 metrics are a good model

12 DDM Functional Test 2006 (9 Tier-1s, 40 Tier-2s) Tier-1Tier-2sSept 06Oct 06Nov 06 ASGC IPAS, Uni MelbourneFailed within the cloud Failed for Melbourne T1-T1 not testd BNL GLT2, NET2,MWT2,SET2, WT2done 2+GB & DPM CNAF LNF,Milano,Napoli,Roma165% failure rate done FZK CSCS, CYF, DESY-ZN, DESY-HH, FZU, WUPFailed from T2 to FZK dCache problem T1-T1 not testd LYON BEIIJING, CPPM, LAPP, LPC, LPHNE, SACLAY, TOKYO donedone, FTS conn =< 6 NG not tested PIC IFAE, IFIC, UAMFailed within the cloud done RAL CAM, EDINBOURGH, GLASGOW, LANCS, MANC, QMUL Failed within the cloud Failed for Edinbrg. done SARA IHEP, ITEP, SINPFailedIHEP not tested IHEP in progress TRIUMF ALBERTA, TORONTO, UniMontreal, SFU, UVICFailed within the cloud FailedT1-T1 not testd New DQ2 release (0.2.12) After SC4 test

Summary 2007 will be an extremely busy and challenging year!  For those of us who have been working on LHC Computing for 15+ years (and others too…) it will nonetheless be extremely rewarding ¿Is there a more important Computing Challenge on the planet this year ?  The ultimate goal – to enable the exploitation of the LHC’s physics discovery potential – is beyond measure

Megatable Extract Tier1 CentreALICE (x4)ATLASCMSLHCbTarget IN2P3, Lyon GridKA, Germany CNAF, Italy FNAL, USA BNL, USA RAL, UK NIKHEF, NL ASGC, Taipei PIC, Spain Nordic Data Grid Facility TRIUMF, Canada US ALICE TOTALS