M.C. Vetterli – LHCC review, CERN; Feb.’09 – #1 Simon Fraser Status of the WLCG Tier-2 Centres M.C. Vetterli Simon Fraser University and TRIUMF LHCC mini-review,

Slides:



Advertisements
Similar presentations
Exporting Raw/ESD data from Tier-0 Tier-1s Wrap-up.
Advertisements

LCG-France Project Status Fabio Hernandez Frédérique Chollet Fairouz Malek Réunion Sites LCG-France Annecy, May
M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #1 Simon Fraser Status of the WLCG Tier-2 Centres M.C. Vetterli Simon Fraser University and TRIUMF WLCG.
1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
Ian Bird LHCC Referees’ meeting; CERN, 18 th November 2014.
ATLAS computing in Geneva Szymon Gadomski, NDGF meeting, September 2009 S. Gadomski, ”ATLAS computing in Geneva", NDGF, Sept 091 the Geneva ATLAS Tier-3.
Jan 2010 Current OSG Efforts and Status, Grid Deployment Board, Jan 12 th 2010 OSG has weekly Operations and Production Meetings including US ATLAS and.
Les Les Robertson WLCG Project Leader WLCG – Worldwide LHC Computing Grid Where we are now & the Challenges of Real Data CHEP 2007 Victoria BC 3 September.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Global Science experiment Data hub Center Oct. 13, 2014 Seo-Young Noh Status Report on Tier 1 in Korea.
Resources and Financial Plan Sue Foffano WLCG Resource Manager C-RRB Meeting, 12 th October 2010.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
16 October 2005 Collaboration Meeting1 Computing Issues & Status L. Pinsky Computing Coordinator ALICE-USA.
Status of the DESY Grid Centre Volker Guelzow for the Grid Team DESY IT Hamburg, October 25th, 2011.
Preparation of KIPT (Kharkov) computing facilities for CMS data analysis L. Levchuk Kharkov Institute of Physics and Technology (KIPT), Kharkov, Ukraine.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
CCRC’08 Weekly Update Jamie Shiers ~~~ LCG MB, 1 st April 2008.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team 12 th CERN-Korea.
LCG Introduction John Gordon, STFC-RAL GDB September 9 th, 2008.
DDM Monitoring David Cameron Pedro Salgado Ricardo Rocha.
GDB March User-Level, VOMS Groups and Roles Dave Kant CCLRC, e-Science Centre.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
Ian Bird LCG Project Leader WLCG Collaboration Issues WLCG Collaboration Board 24 th April 2008.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team ATHIC2012, Busan,
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Procedure to follow for proposed new Tier 1 sites Ian Bird CERN, 27 th March 2012.
The following is a collection of slides from a few recent talks on computing for ATLAS in Canada, plus a few new ones. I might refer to all of them, I.
CCRC’08 Monthly Update ~~~ WLCG Grid Deployment Board, 14 th May 2008 Are we having fun yet?
US-CMS T2 Centers US-CMS Tier 2 Report Patricia McBride Fermilab GDB Meeting August 31, 2007 Triumf - Vancouver.
Procedure for proposed new Tier 1 sites Ian Bird WLCG Overview Board CERN, 9 th March 2012.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
LCG Accounting Update John Gordon, CCLRC-RAL WLCG Workshop, CERN 24/1/2007 LCG.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
U.S. ATLAS Facility Planning U.S. ATLAS Tier-2 & Tier-3 Meeting at SLAC 30 November 2007.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 1 st March 2011 Visit of Dr Manuel Eduardo Baldeón.
Jiri Chudoba for the Pierre Auger Collaboration Institute of Physics of the CAS and CESNET.
WLCG Operations Coordination report Maria Alandes, Andrea Sciabà IT-SDC On behalf of the WLCG Operations Coordination team GDB 9 th April 2014.
M.C. Vetterli – WLCG-CB, March ’09 – #1 Simon Fraser WLCG Collaboration Board Meeting Praha, March 22 nd, 2009 Thanks to Milos for hosting us.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
1 September 2007WLCG Workshop, Victoria, Canada 1 WLCG Collaboration Workshop Victoria, Canada Site Readiness Panel Discussion Saturday 1 September 2007.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
WLCG Accounting Task Force Update Julia Andreeva CERN GDB, 8 th of June,
LCG Accounting Update John Gordon, CCLRC-RAL 10/1/2007.
LCG Introduction John Gordon, STFC-RAL GDB June 11 th, 2008.
ATLAS Computing Model Ghita Rahal CC-IN2P3 Tutorial Atlas CC, Lyon
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
Pledged and delivered resources to ALICE Grid computing in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
Operations Coordination Team Maria Girone, CERN IT-ES GDB, 11 July 2012.
Availability of ALICE Grid resources in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
Alice Operations In France
LCG Service Challenge: Planning and Milestones
Data Challenge with the Grid in ATLAS
Update on Plan for KISTI-GSDC
Readiness of ATLAS Computing - A personal view
Presentation transcript:

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #1 Simon Fraser Status of the WLCG Tier-2 Centres M.C. Vetterli Simon Fraser University and TRIUMF LHCC mini-review, CERN, February 16 th 2009

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #2 Simon Fraser Communications for Tier-2s  Many lines of communication do indeed exist.  Some examples are: CMS has two Tier-2 coordinators: Ken Bloom (Nebraska) Giuseppe Bagliesi (INFN) - attend all operations meetings - feed T2 issues back to the operations group - write T2-relevant minutes - organize T2 workshops ALICE has designated 1 Core Offline person in 3 to have privileged contact with a given T2 site manager - weekly coordination meetings - Tier-2 federations provide a single contact person - A Tier-2 coordinates with its regional Tier-1

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #3 Simon Fraser Communications for Tier-2s ATLAS uses its cloud structure for communications - Every Tier-2 is coupled to a Tier national clouds; others have foreign members (e.g. “Germany” includes Krakow, Prague, Switzerland; Netherlands includes Russia, Israel, Turkey) - Each cloud has a Tier-2 coordinator Regional organizations, such as: + France Tier-2/3 technical group: - coordinates with Tier-1 and with experiments - monthly meetings - coordinates procurement and site management + GRIF: Tier-2 federation of 5 labs around Paris + Canada: Weekly teleconferences of technical personnel (T1 & T2) to share information and prepare for upgrades, large production, etc. + Many others exist; e.g. in the US and the UK

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #4 Simon Fraser Communications for Tier-2s  Tier-2 Overview Board reps: Michel Jouvin and Atul Gurtu were appointed in October to the OB to give the Tier-2s a voice there.  Tier-2 mailing list: Actually exists and is being reviewed for completeness & accuracy  Tier-2 GDB: The October GDB was dedicated to Tier-2 issues + reports from experiments: role of the T2s; communications + talks on regional organizations + discussion of accounting + technical talks on storage, batch systems, middleware  Seems to have been a success; repeat a couple of times per year?

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #5 Simon Fraser Tier-2 Reliability September ‘08 May ‘08 January ‘09

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #6 Simon Fraser Tier-2 Reliability 41 of 62 sites are now green; 8 more are >80% Average is now ≈90% All but 1 site are reporting; in particular the situation in the US has been resolved. Still some ”one-off” issues such as a few sites with green relia- bility, but yellow availability (i.e. significant declared downtime). Tier-2 specific tests exist: - CMS has Tier-2 commissioning - ATLAS has Tier-2 specific functional tests

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #7 Simon Fraser

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #8 Simon Fraser  But how much of this is a problem of under-use rather than under-contribution?  a task force was set up to extract installed capacities from the Glue schema  Monthly APEL reports still undergo significant modifications from first draft.  Good because communication with T2s better  Bad because APEL accounting still has problems  However, the task force’s work is nearing completion; the MB has approved the document outlining the solution (actually it is solutions: EGEE vs OSG, CPU vs storage) Tier-2 Installed Resources

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #9 Simon Fraser Installed vs Pledged Capacities at U.S. Tier-2s NET2North East Tier-2 Center at Boston University and Harvard University SWT2Southwest Tier-2 Center at University at Texas – Arlington and Oklahoma University MWT2Midwest Tier-2 Center at University of Chicago and Indiana University AGLT2ATLAS Great Lakes Tier-2 Center at University of Michigan and Michigan State University WT2Western Tier-2 Center at SLAC (From M. Ernst)

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #10 Simon Fraser How are the Tier-2s being used? From APEL accounting page for the last 6 months

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #11 Simon Fraser Tier-2s in Production From APEL accounting portal for Aug.’08 to Jan.’09; #s in MSI2k AliceATLASCMSLHCbTotal Tier-1s % Tier-2s % Total Warning: These numbers vary depending on what you put in your query

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #12 Simon Fraser Analysis jobs last month 20,000 Pending 5,000 Running Note: We do not have stats for jobs that do not report to dashboard. We know that such jobs exist.  Need WLCG dashboard comparison ! From F. Wuerthwein (UCSD-CMS)

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #13 Simon Fraser CMS Summary 80% of analysis activity at T2 & T3. 1/4 of collaboration submitted jobs in ~1 Million hours consumed per week. 30 T2 & 3 T3 with CMS-SAM availability > 80% for the last month. 1.0 PB placed, and accounted for by “groups” at T2. Additional 8 PB placed outside group accounting: –5.5PB at T1 and T0 –136TB at T3 Note: #s based on CMS dashboard and PhEDEx accounting. From F. Wuerthwein (UCSD-CMS)

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #14 Simon Fraser Placement Accounting Examples Placement includes T0,T1,T2,T3 The same dataset may be “owned” by different groups at different sites.

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #15 Simon Fraser  ATLAS has started an organized program of file deletion. Data Issues at ATLAS

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #16 Simon Fraser  “10M files” exercise: - stress the data distribution system by transferring a huge number of files in a short time (10k datasets transferred in 10 days; 1M files to each T1) - Brought to light some issues with RTT for file registra- tion; these should apply to large-scale T2 transfers too  need bulk registration capabilities on the LFC Data Issues at ATLAS

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #17 Simon Fraser 10M files ATLAS (From S. Campana)

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #18 Simon Fraser  How does the LHC delay affect the requirements and pledges for 2009? + We have heard about this earlier  We need to use something other than SpecInt2000! + this benchmark is totally out-of-date & useless for new CPUs + SpecHEP06 will be used from now on; welcomed development Tier-2 Hardware Questions

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #19 Simon Fraser  Networking to the nodes is now an issue. + with 8 cores per node, 1 GigE connection ≈ 16.8 MB/sec/core + Tier-2 analysis jobs run on reduced data sets and can do rather simple operations  see M. Schott slide (next) Tier-2 Hardware Questions

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #20 Simon Fraser Data processed per second Data read per second, as measured by root. All files cached on disk. Data format, programReading speed [MB/s] AOD (7 container), Athena AOD (7 containers), ARA3.75 – 5.0 AOD (trk.particles), Athena2.75 Vector >, ROOT4.93 Simple ntuple, ROOT6.99 ALICE esd file18 ROOT example47 (From M. Schott)

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #21 Simon Fraser  Networking to the nodes is now an issue. + with 8 cores per node, 1 GigE connection ≈ 16.8 MB/sec/core + Tier-2 analysis jobs run on reduced data sets and can do rather simple operations  see M. Schott slide (next) + Do we need to go to Infiniband? + We certainly need increased capability for the uplinks; we should have a minimum of fully non-blocking GigE the worker nodes.  We need more guidance from the experiments The next round of purchases is soon/now! Tier-2 Hardware Questions

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #22 Simon Fraser User Issues: Load & Support  We saw earlier that the number of users has gone up significantly, but it will go up a lot more. + We must make the Grid easier to use

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #23 Simon Fraser User Issues: It’s all still a little complicated

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #24 Simon Fraser User Issues: Load & Support  We saw earlier that the number of users has gone up significantly, but it will go up a lot more + We must make the Grid easier to use  User stress tests are being done regularly :  Hammercloud tests  Work continues to make the “backend” invisible  Much progress has been made on user support + A distributed-analysis user support group has been formed + Four people in the EU, four in the US; uses hypernews & gmail + Quite successful but we need better documentation + User2user support is starting to happen; encourage this.

M.C. Vetterli – LHCC review, CERN; Feb.’09 – #25 Simon Fraser Summary  The role of the Tier-2 centres continues to increase  Communication issues have been addressed but need more work  Reliability continues to be generally good, but needs to be better  Automatic resource & usage monitoring tools are almost ready  Stress testing of data distribution and user analysis has ramped up  The number of users continues to increase but we still need to make the Grid easier to use for beginners  Organized user support is becoming a reality  The Tier-2 layer of WLCG continues to improve!