Presentation is loading. Please wait.

Presentation is loading. Please wait.

Ian Bird LCG Project Leader Jamie Shiers Grid Support Group, CERN WLCG Project Status Report NEC 2009 September 2009.

Similar presentations


Presentation on theme: "Ian Bird LCG Project Leader Jamie Shiers Grid Support Group, CERN WLCG Project Status Report NEC 2009 September 2009."— Presentation transcript:

1 Ian Bird LCG Project Leader Jamie Shiers Grid Support Group, CERN WLCG Project Status Report NEC 2009 September 2009

2 Introduction The sub-title of this talk is “Grids step-up to a set of new records: Scale Testing for the Experiment Programme (STEP’09)” STEP’09 means different things to different people A two week period during June 2009 when there was intense testing – particularly by ATLAS & CMS – of specific (overlapping) workflows A several month period, starting around CHEP’09, and encompassing the above I would like to “step back” and take a much wider viewpoint – with a reference to my earlier “HEP SSC” talk: Are we ready to “successfully and efficiently exploit the scientific and discovery potential of the LHC”? 2

3 “The Challenge” This challenge was clearly posed by Fabiola Gianotti during her CHEP 2004 plenary talk “Fast forward” 3 years – to CHEP 2007 – when some people were asking whether it was wise to travel to Vancouver when the LHC startup was imminent  At that time we clearly had not tested key Use Cases – sometimes not even by individual experiments, let alone all experiments (and at all concerned sites) together This led to the Common Computing Readiness Challenge (CCRC’08) which advanced the state of play significantly >> to CHEP’09 – “ready but there will be problems” 3

4 CCRC’08 Once again, this was supposed to be a final production test prior to real collisions between accelerated beams in the LHC It certainly raised the bar considerably – and much of our operations infrastructure was completed as a result of that exercise – but it still left some components untested  These were the focus of STEP’09 The bottom line: we were not fully ready for data in 2007 – nor even 2008. The impressive results must be considered in the light of this sobering thought 4

5 So What Next? Whilst there is no doubt that the service has “stepped up” considerably since e.g. one year ago, can We (providers) live with this level of service and the operations load that it generates? The experiments live with this level of service and the problems that it causes? (Loss of useful work, significant additional work, …) Where are wrt “the challenge” of CHEP 2004? 5

6 An Aside Over the past few years there were a number of technical problems related to the LHC machine itself For me, a particularly large slice of “humble pie” came with the “IT problem” This was not about Indico being down or slow, or Twiki being inaccessible, it was about the (LHC) Inner Triplets To many, the collaboration is perceived to be “LHC machine + detectors” – “computing” is either an afterthought or more likely not a thought at all! 6

7 LHC + Experiments + WLCG??? In reality, IT is needed from the very beginning – to design the machine, the detectors, to build and operate them... And – by the way – there would today be no physics discovery without major computational, network and storage facilities  We call this (loosely) WLCG – as you know! But the only way to get on the map is through the provision of reliable, stable, predictable services And a service is determined as much by what happens when things go wrong as by the “trivial” situation of smooth running… 7

8 STEP’09: Service Advances For CCRC’08 we had to put in place new or upgraded service / operations infrastructure Some elements were an evolution of what had been used for previous Data and Service challenges but key components were basically new Not only did these prove their worth in CCRC’08 but basically no major changes have been needed to date The operations infrastructure worked smoothly – sites were no longer in “hero” (unsustainable) mode – previously a major concern  Rather light-weight but collaborative procedures proved their worth But most importantly, our ability to handle / recover from / circumvent even major disasters! 8

9 What Has Gone Wrong? Loss of LHC OPN to US – cables cut by fishing trawler This happened during an early Service Challenge and at the time we thought it was “unusual” Loss of LHC OPN within Europe – construction work near Madrid, motorway construction between Zurich and Basle (you can check the GPS coordinates with Google Earth), Tsunami in Asia, fire in Taipei, tornadoes, hurricanes, collapse of machine room floor due to municipal construction underneath(!), bugs in tape robot firmware taking drives offline, human errors, major loss of data due to s/w bugs, … Some of the above occurred during STEP’09 – but the exercise was still globally a success! 9

10 FZK FNAL TRIUMF NGDF CERN Barcelona/PIC Lyon/CCIN2P3 Bologna/CAF Amsterdam/NIKHEF-SARA BNL RAL Taipei/ASGC

11 FZK FNAL TRIUMF NGDF CERN Barcelona/PIC Lyon/CCIN2P3 Bologna/CAF Amsterdam/NIKHEF-SARA BNL RAL Taipei/ASGC

12 STEP’09: What Were The Metrics? Those set by the experiments: based on the main “functional blocks” that Tier1s and Tier2s support Primary (additional) Use Cases in STEP’09: 1.(Concurrent) reprocessing at Tier1s – including recall from tape 2.Analysis – primarily at Tier2s (except LHCb) In addition, we set a single service / operations site metric, primarily aimed at the Tier1s (and Tier0) Details: ATLAS (logbook, p-m w/s), CMS (p-m), blogsATLASlogbookp-m w/sCMSp-mblogs Daily minutes: week1, week2week1week2 WLCG Post-mortem workshop 12

13 WLCG Tier1 [ Performance ] Metrics ~~~ Points for Discussion Jamie.Shiers@cern.ch ~~~ WLCG GDB, 8 th July 2009

14 The Perennial Question During this presentation and discussion we will attempt to sharpen and answer the question: How can a Tier1 know that it is doing OK? We will look at: What we can (or do) measure (automatically); What else is important – but harder to measure (at least today); How to understand what “OK” really means… 14

15 Resources In principle, we know what resources are pledged, can determine what are actually installed(?) and can measure what is currently being used; If installed capacity is significantly(?) lower than pledged, this is an anomaly and site in question “is not doing ok” But actual utilization may vary – and can even exceed – “available” capacity for a given VO (particularly CPU – less or unlikely for storage(?))  This should also be signaled as an anomaly to be understood (it is: poor utilization over prolonged periods impacts future funding, even if there are good reasons for it…) 15

16 Services Here we have extensive tests (OPS, VO) coupled with production use A “test” can pass, which does not mean that experiment production is not (severely) impacted…) Some things are simply not realistic or too expensive to test… But again, significant anomalies should be identified and understood Automatic testing is one measure: GGUS tickets another (# tickets, including alarm, time taken for their resolution) This can no doubt be improved iteratively; additional tests / monitoring added (e.g. tape metrics) A site which is “green”, has few or no tickets open for > days | weeks, and no “complaints” at operations meeting is doing ok, surely? Can things be improved for reporting and long-term traceability? (expecting the answer YES) 16

17 The Metrics… For STEP’09 – as well as at other times – explicit metrics have been set against sites and for well defined activities Can such metrics allow us to “roll-up” the previous issues into a single view? If not, what is missing from what we currently do? Is it realistic to expect experiments to set such targets: During the initial period of data taking? (Will it be known at all what the “targets” actually are?) In the longer “steady state” situation? Processing & reprocessing? MC production? Analysis??? (largely not T1s…) Probable answer: only if it is useful for them to monitor their own production (which it should be..) 17

18 #Metric 1Site is providing (usable) resources that match those pledged & requested; 2The services are running smoothly, pass the tests and meet reliability and availability targets; 3“WLCG operations” metrics on handling scheduled and unscheduled service interruptions and degradations are met; 4Site is meeting or exceeding metrics for “functional blocks”. WLCG Site Metrics

19 Critical Service Follow-up Targets (not commitments) proposed for Tier0 services Similar targets requested for Tier1s/Tier2s Experience from first week of CCRC’08 suggests targets for problem resolution should not be too high (if ~achievable) The MoU lists targets for responding to problems (12 hours for T1s) ¿Tier1s: 95% of problems resolved <1 working day ? ¿Tier2s: 90% of problems resolved < 1 working day ?  Post-mortem triggered when targets not met! 19 Time IntervalIssue (Tier0 Services)Target End 2008Consistent use of all WLCG Service Standards100% 30’Operator response to alarm / call to x5011 / alarm e-mail99% 1 hourOperator response to alarm / call to x5011 / alarm e-mail100% 4 hoursExpert intervention in response to above95% 8 hoursProblem resolved90% 24 hoursProblem resolved99%

20 GGUS summary (2 weeks) VOUserTeamAlarmTotal ALICE8008 ATLAS2156178 CMS2013 LHCb0120 Totals31682101 20

21 21

22 22

23 What Were The Results? The good news first: Most Tier1s and many of the Tier2s met – and in some cases exceeded by a significant margin – the targets that were set In addition, this was done with reasonable operational load at the site level and with quite a high background of scheduled and unscheduled interventions and other problems – including 5 simultaneous LHC OPN fibre cuts!  Operationally, things went really rather well Experiment operations – particularly ATLAS – overloaded  The not-so-good news: Some Tier1s and Tier2s did not meet one or more of the targets 23

24 Tier2s The results from Tier2s are somewhat more complex to analyse – an example this time from CMS: Primary goal: use at least 50% of pledged T2 level for analysis backfill ongoing analysis activity go above 50% if possible Preliminary results: In aggregate: 88% of pledge was used. 14 sites with > 100% 9 sites below 50% The number of Tier2s is such that it does not make sense to go through each by name, however:  Need to understand primary causes for some sites to perform well and some to perform relatively badly  Some concerns on data access performance / data management in general at Tier2s: this is an area which has not been looked at in (sufficient?) detail 24

25 Summary of Tier2s Detailed reports written by a number of Tier2s MC conclusion “solved since a long time” (Glasgow) Also some numbers on specific tasks, e.g. GangaRobotGangaRobot Some specific areas of concern (likely to grow IMHO) Networking: internal bandwidth and/or external Data access: aside from constraints above, concern that data access will met the load / requirements from heavy end-user analysis “Efficiency” – # successful analysis jobs – varies from 94% down to 56% per (ATLAS) cloud, but >99% down to 0% (e.g. 13K jobs failed, 100 succeed) (error analysis also exists) IMHO, the detailed summaries maintained by the experiments together with site reviews demonstrate that the process is under control, not withstanding concerns 25

26 Ian.Bird@cern.ch 26 STEP key points  General:  Multi-VO aspects never tested before at this scale  Almost all sites participated successfully  CERN tape writing well above required level  Most Tier1s showed impressive operation  Demonstrated scale and sustainability of loads  Some limitations were seen; to be re-checked  OPN suffered double fibre cut!... But continued and recovered...  Data rates well above required rates...

27 CCRC 2008 vs STEP 2009 CCRC08 MB/s STEP09 2 weeks vs. 2 days 4 GB/sec vs. 1 GB/sec

28 Recommendations 1.Resolution of major problems with in-depth written reports 2.Site visits to Tier1s that gave problems during STEP’09 (at least DE-KIT & NL-T1) [ ASGC being setup for October? ] 3.Understanding of Tier2 successes and failures 4.Rerun of “STEP’09” – perhaps split into reprocessing and analysis before a “final” re-run – on timescale of September 2009 [ Actually done as a set of sub-tasks ] 5.Review of results prior to LHC restart 28

29 General Conclusions STEP’09 was an extremely valuable exercise and we all learned a great deal! Progress – again – has been significant The WLCG operations procedures / meetings have proven their worth Good progress since (see experiment talks) on understanding and resolving outstanding issues!  Overall, STEP’09 was a big step forward! 29

30 Outstanding Issues & Concerns 30 IssueConcern NetworkT0 – T1 well able to handle traffic that can be expected from normal data taking with plenty of headroom for recovery. Redundancy?? T1 – T1 traffic – less predictable (driven by re-processing) – actually dominates. Concerns about use of largely star network for this purpose. Tn – T2 traffic – likely to become a problem, as well internal T2 bandwidth StorageWe still do not have our storage systems under control. Significant updates to both CASTOR and dCache have been recommended by providers post- STEP’09. Upgrade paths unclear, untested or both. DataData access – particularly “chaotic” access patterns typical of analysis can be expected to cause problems – many sites configured for capacity, not optimized for many concurrent streams, random access etc. UsersAre we really ready to handle a significant increase in the number of (blissfully) grid-unaware users?

31 Summary We are probably ready for data taking and analysis and have a proven track record of resolving even major problems and / or handling major site downtimes in a way that lets production continue Analysis will surely bring some new challenges to the table – not only the ones that we expect! If funded, the HEP SSC and Service Deployment projects described this morning will help us get through the first years of LHC data taking Expect some larger changes – particularly in the areas of storage and data handing – after that 31


Download ppt "Ian Bird LCG Project Leader Jamie Shiers Grid Support Group, CERN WLCG Project Status Report NEC 2009 September 2009."

Similar presentations


Ads by Google