John Gordon STFC-RAL Tier1 Status 9 th July, 2008 Grid Deployment Board.

Slides:



Advertisements
Similar presentations
Exporting Raw/ESD data from Tier-0 Tier-1s Wrap-up.
Advertisements

Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks LHCOPN Operations update Guillaume Cessieux.
SC4 Workshop Outline (Strong overlap with POW!) 1.Get data rates at all Tier1s up to MoU Values Recent re-run shows the way! (More on next slides…) 2.Re-deploy.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
ATLAS Metrics for CCRC’08 Database Milestones WLCG CCRC'08 Post-Mortem Workshop CERN, Geneva, Switzerland June 12-13, 2008 Alexandre Vaniachine.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
CMS STEP09 C. Charlot / LLR LCG-DIR 19/06/2009. Réunion LCG-France, 19/06/2009 C.Charlot STEP09: scale tests STEP09 was: A series of tests, not an integrated.
WLCG Service Report ~~~ WLCG Management Board, 24 th November
CCRC08-1 report WLCG Workshop, April KorsBos, ATLAS/NIKHEF/CERN.
CERN IT Department CH-1211 Genève 23 Switzerland t Tier0 Status - 1 Tier0 Status Tony Cass LCG-LHCC Referees Meeting 18 th November 2008.
CERN IT Department CH-1211 Genève 23 Switzerland t Tier0 Status - 1 Tier0 Status Tony Cass LCG-LHCC Referees Meeting 6 th July 2009.
Ian Bird LCG Project Leader LHCC Referee Meeting Project Status & Overview 22 nd September 2008.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks LHCOPN Ops WG Act 5 Guillaume Cessieux (CNRS/IN2P3-CC,
Project Management Sarah Pearce 3 September GridPP21.
11 March 2008 GridPP20 Collaboration meeting David Britton - University of Glasgow GridPP Status GridPP20 Collaboration Meeting, Dublin David Britton,
LCG Introduction John Gordon, STFC-RAL GDB September 9 th, 2008.
HEPiX FNAL ‘02 25 th Oct 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 25 th October 2002 HEPiX 2002, FNAL.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
WLCG Grid Deployment Board, CERN 11 June 2008 Storage Update Flavia Donno CERN/IT.
Ian Bird LCG Project Leader WLCG Collaboration Issues WLCG Collaboration Board 24 th April 2008.
ATLAS Bulk Pre-stageing Tests Graeme Stewart University of Glasgow.
SC4 Planning Planning for the Initial LCG Service September 2005.
WLCG Tier1 [ Performance ] Metrics ~~~ Points for Discussion ~~~ WLCG GDB, 8 th July 2009.
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
CCRC’08 Monthly Update ~~~ WLCG Grid Deployment Board, 14 th May 2008 Are we having fun yet?
LCG Sep. 26thLHCC comprehensive review 2006 Volker Guelzow 1 Tier 1 status, a summary based upon a internal review Volker Gülzow DESY.
LCG Report from GDB John Gordon, STFC-RAL MB meeting February24 th, 2009.
Network Connections for the Worldwide LHC Computing Grid Tony Cass Leader, Communication Systems Group Information Technology Department 11 th December.
Tier-1 Andrew Sansum Deployment Board 12 July 2007.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
4 March 2008CCRC'08 Feb run - preliminary WLCG report 1 CCRC’08 Feb Run Preliminary WLCG Report.
LCG WLCG Accounting: Update, Issues, and Plans John Gordon RAL Management Board, 19 December 2006.
Author - Title- Date - n° 1 Partner Logo WP5 Status John Gordon Budapest September 2002.
Report from GSSD Storage Workshop Flavia Donno CERN WLCG GDB 4 July 2007.
LCG Accounting Update John Gordon, CCLRC-RAL WLCG Workshop, CERN 24/1/2007 LCG.
LCG Service Challenges SC2 Goals Jamie Shiers, CERN-IT-GD 24 February 2005.
Ian Bird LCG Project Leader WLCG Status Report CERN-RRB th April, 2008 Computing Resource Review Board.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Ian Bird All Activity Meeting, Sofia
SL5 Site Status GDB, September 2009 John Gordon. LCG SL5 Site Status ASGC T1 - will be finished before mid September. Actually the OS migration process.
LCG Issues from GDB John Gordon, STFC WLCG MB meeting September 28 th 2010.
8 August 2006MB Report on Status and Progress of SC4 activities 1 MB (Snapshot) Report on Status and Progress of SC4 activities A weekly report is gathered.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Grid Deployment Board 5 December 2007 GSSD Status Report Flavia Donno CERN/IT-GD.
Summary of SC4 Disk-Disk Transfers LCG MB, April Jamie Shiers, CERN.
LCG Tier1 Reliability John Gordon, STFC-RAL CCRC09 November 13 th, 2008.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
SRM 2.2: experiment requirements, status and deployment plans 6 th March 2007 Flavia Donno, INFN and IT/GD, CERN.
Ian Bird LCG Project Leader WLCG Status Report 7 th May, 2008 LHCC Open Session.
LCG Accounting Update John Gordon, CCLRC-RAL 10/1/2007.
LCG Introduction John Gordon, STFC-RAL GDB November 7th, 2007.
Top 5 Experiment Issues ExperimentALICEATLASCMSLHCb Issue #1xrootd- CASTOR2 functionality & performance Data Access from T1 MSS Issue.
ATLAS Computing Model Ghita Rahal CC-IN2P3 Tutorial Atlas CC, Lyon
Pledged and delivered resources to ALICE Grid computing in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
Availability of ALICE Grid resources in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
T0-T1 Networking Meeting 16th June Meeting
Grid Operations Centre Progress to Aug 03
Computing Operations Roadmap
Collaboration Meeting
Flavia Donno CERN GSSD Storage Workshop 3 July 2007
Data Challenge with the Grid in ATLAS
Update on Plan for KISTI-GSDC
Summary from last MB “The MB agreed that a detailed deployment plan and a realistic time scale are required for deploying glexec with setuid mode at WLCG.
John Gordon, STFC GDB October 12th 2011
LHC Data Analysis using a worldwide computing grid
February 2007 Note: Source:.
The LHCb Computing Data Challenge DC06
Presentation transcript:

John Gordon STFC-RAL Tier1 Status 9 th July, 2008 Grid Deployment Board

2 Overview  T1 Procurements  Reliability  Tape Efficiency  24x7 & VO Boxes  CCRC08  Readiness  Welcome feedback from T1s and experiments

3 Procurement  All Tier0/Tier 1 had problems in 2008 procurements  Could become a problem in future years...  Funding not always clear before procurements need to start  Added Milestone in Sept08 to report status and prognosis  All Tier0/Tier 1 had problems in 2008 procurements  Could become a problem in future years...  Funding not always clear before procurements need to start  Added Milestone in Sept08 to report status and prognosis

Procurements  Had T1s installed their pledged hardware for 2008 by 1 April?  Had T1s installed their pledged hardware for 2008 by 5 May? For CCRC08(May).  Had T1s installed sufficient capacity to meet the experiments plans for CCRC08(May) April08?May08? CCRC Requirements met? ASGCNo Yes BNLNo Yes CNAFNo Yes FNALNo Yes FZKYes IN2P3No Yes NDGFNo Yes NIKHEFNo Yes PICNo Yes RALNoYes TriumfYes Disk and Tape OK, 80% of CPU CPU OK Tape OK, 63% disk CPU 110%, disk 80%, tape 25% Tape 30% CPU 30%, Disk 60%

5 Harry’s Table  A 4 th T1 met pledges in June (FNAL)  But a further 3 meet their CPU pledge  IN2P3, PIC, BNL  and 2 of those, tape too.  More sites fail to deliver disk  Disk is the biggest shortfall  A 4 th T1 met pledges in June (FNAL)  But a further 3 meet their CPU pledge  IN2P3, PIC, BNL  and 2 of those, tape too.  More sites fail to deliver disk  Disk is the biggest shortfall

6 Site Procurement Comments All in Place ASGCMid September BNL CPU by June 20th, disk (less 1PB) June 20th, remaining PB after October when new machine room open. CNAFCPU by July, disk by September, tape July FNALIn place before start of collisions FZKAlways planned to meet part of pledge in October IN2P3Disk by September NDGFDisk by September NIKHEFLater dates PICCPU start of June, disk by end of July RAL Triumf Lessons for Tier2s?

7 Reliability  Definite improvements in reliability  11/12 sites > 93% in May  10/12 sites > 93% in June  8/12 sites > 95% (new target) in June  Average of ALL sites > 95% in May and June  Milestones completed:  Average of 8 best sites above June target in May  Milestones completed:  Average of 8 best sites above June target in May

8  T1 avail apr.jpg T1 avail apr.jpg

9 Reliability – 2  New Tier 1 reliability milestones:  June – improved overall values  December – All sites to be above target  New Tier 1 reliability milestones:  June – improved overall values  December – All sites to be above target

10 Tape Metrics – MSS Efficiency Tier-0 SiteLast Update CERN Tier-1 SitesData Available CA-TRIUMF DE-KIT ES-PIC FR-CCIN2P3 - IT-INFN-CNAF NDGF NL-T TW-ASGC - UK-T1-RAL US-FNAL-CMS US-T1-BNL  9/11 T1 Publishing Efficiency Metrics  Conclusions? Issues?

11

12  TRIUMF has better rates than most sites.  Writing almost full tapes

13 Outstanding Milestones All 12(10) sites have tested their 24 X 7 support, and 10(7) have put the support into operation 7(6) sites have implemented a VO BOX SLA No change in acceptance by experiments 7(6) sites have implemented a VO BOX SLA No change in acceptance by experiments

14 Happiness with CCRC08  Tier1s declared themselves generally happy with their performance in CCRC08  Issues included:  Information, Information, Information.  unsure what was expected of them at any given time  need a site-centric view of the world  Need tools to monitor storage  Storage tokens defined late, data rates not at all  Storage – robustness and quality issues  Both dCache and Castor  Job Mix  Floods of jobs  High i/o  User analysis tape mounts

15 General Issues of Readiness  All Tier1s considered themselves ready for data  Within the limitations of the middleware.  Remaining doubts but Tier1s cannot solve alone.  Need good storage monitoring  Observe that reconstruction and bulk tape recall have not been tested to the required level  Human intervention level may still be high  I think many ignored their lack of installed capacity  Bring it on!  Tier2s?  Mixed responses from T1s about ‘their’ T2s.  Some happy  Some mention communication issues  Still ramping up hardware  Most have now passed functional tests but few have been stressed.