M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #1 Simon Fraser Status of the WLCG Tier-2 Centres M.C. Vetterli Simon Fraser University and TRIUMF WLCG.

Slides:



Advertisements
Similar presentations
ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
Advertisements

LCG-France Project Status Fabio Hernandez Frédérique Chollet Fairouz Malek Réunion Sites LCG-France Annecy, May
M.C. Vetterli – LHCC review, CERN; Feb.’09 – #1 Simon Fraser Status of the WLCG Tier-2 Centres M.C. Vetterli Simon Fraser University and TRIUMF LHCC mini-review,
Sue Foffano LCG Resource Manager WLCG – Resources & Accounting LHCC Comprehensive Review November, 2007 LCG.
ATLAS computing in Geneva Szymon Gadomski, NDGF meeting, September 2009 S. Gadomski, ”ATLAS computing in Geneva", NDGF, Sept 091 the Geneva ATLAS Tier-3.
Jan 2010 Current OSG Efforts and Status, Grid Deployment Board, Jan 12 th 2010 OSG has weekly Operations and Production Meetings including US ATLAS and.
Overview of LCG-France Tier-2s and Tier-3s Frédérique Chollet (IN2P3-LAPP) on behalf of the LCG-France project and Tiers representatives CMS visit to Tier-1.
Project Status Report Ian Bird Computing Resource Review Board 30 th October 2012 CERN-RRB
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Resources and Financial Plan Sue Foffano WLCG Resource Manager C-RRB Meeting, 12 th October 2010.
Status of the DESY Grid Centre Volker Guelzow for the Grid Team DESY IT Hamburg, October 25th, 2011.
Preparation of KIPT (Kharkov) computing facilities for CMS data analysis L. Levchuk Kharkov Institute of Physics and Technology (KIPT), Kharkov, Ukraine.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
John Gordon STFC-RAL Tier1 Status 9 th July, 2008 Grid Deployment Board.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
ATLAS in LHCC report from ATLAS –ATLAS Distributed Computing has been working at large scale Thanks to great efforts from shifters.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
LCG Introduction John Gordon, STFC-RAL GDB September 9 th, 2008.
Progress in Computing Ian Bird ICHEP th July 2010, Paris
Dan Tovey, University of Sheffield User Board Overview Dan Tovey University Of Sheffield.
…building the next IT revolution From Web to Grid…
Tony Doyle - University of Glasgow 8 July 2005Collaboration Board Meeting GridPP Report Tony Doyle.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Ian Bird LCG Project Leader WLCG Collaboration Issues WLCG Collaboration Board 24 th April 2008.
Site Validation Session Report Co-Chairs: Piotr Nyczyk, CERN IT/GD Leigh Grundhoefer, IU / OSG Notes from Judy Novak WLCG-OSG-EGEE Workshop CERN, June.
Procedure to follow for proposed new Tier 1 sites Ian Bird CERN, 27 th March 2012.
The following is a collection of slides from a few recent talks on computing for ATLAS in Canada, plus a few new ones. I might refer to all of them, I.
CCRC’08 Monthly Update ~~~ WLCG Grid Deployment Board, 14 th May 2008 Are we having fun yet?
US-CMS T2 Centers US-CMS Tier 2 Report Patricia McBride Fermilab GDB Meeting August 31, 2007 Triumf - Vancouver.
Procedure for proposed new Tier 1 sites Ian Bird WLCG Overview Board CERN, 9 th March 2012.
CMS Usage of the Open Science Grid and the US Tier-2 Centers Ajit Mohapatra, University of Wisconsin, Madison (On Behalf of CMS Offline and Computing Projects)
CERN IT Department CH-1211 Genève 23 Switzerland t Experiment Operations Simone Campana.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks APEL CPU Accounting in the EGEE/WLCG infrastructure.
Interfacing Grid-Canada to LCG M.C. Vetterli, R. Walker Simon Fraser Univ. Grid Deployment Area Mtg August 2 nd, 2004.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
SL5 Site Status GDB, September 2009 John Gordon. LCG SL5 Site Status ASGC T1 - will be finished before mid September. Actually the OS migration process.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 1 st March 2011 Visit of Dr Manuel Eduardo Baldeón.
WLCG Operations Coordination report Maria Alandes, Andrea Sciabà IT-SDC On behalf of the WLCG Operations Coordination team GDB 9 th April 2014.
M.C. Vetterli – WLCG-CB, March ’09 – #1 Simon Fraser WLCG Collaboration Board Meeting Praha, March 22 nd, 2009 Thanks to Milos for hosting us.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
1 September 2007WLCG Workshop, Victoria, Canada 1 WLCG Collaboration Workshop Victoria, Canada Site Readiness Panel Discussion Saturday 1 September 2007.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
WLCG Accounting Task Force Update Julia Andreeva CERN GDB, 8 th of June,
LCG Introduction John Gordon, STFC-RAL GDB June 11 th, 2008.
ATLAS Computing Model Ghita Rahal CC-IN2P3 Tutorial Atlas CC, Lyon
Pledged and delivered resources to ALICE Grid computing in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
Operations Coordination Team Maria Girone, CERN IT-ES GDB, 11 July 2012.
Availability of ALICE Grid resources in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
Alice Operations In France
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
Computing Operations Roadmap
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Grid site as a tool for data processing and data analysis
Data Challenge with the Grid in ATLAS
Update on Plan for KISTI-GSDC
RDIG for ALICE today and in future
John Gordon, STFC GDB October 12th 2011
LHC Data Analysis using a worldwide computing grid
Collaboration Board Meeting
Presentation transcript:

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #1 Simon Fraser Status of the WLCG Tier-2 Centres M.C. Vetterli Simon Fraser University and TRIUMF WLCG Overview Board, CERN, October 27 th 2008

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #2 Simon Fraser Sources of Information  Discussions with experiment representatives in July  APEL monitoring portal  WLCG reliability reports  October GDB mtg; dedicated to Tier-2 issues  Talks from the last OB & LHCC Slides labeled with a * are from MV’s LHCC rapporteur talk

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #3 Simon Fraser Tier-2 Performance Summary*  Overall, the Tier-2s are contributing much more now  Significant fractions of the Monte Carlo simulations are being done in the T2s for all experiments  Reliability is better, but still needs to improve  CCRC’08 exercise is generally considered a success for the Tier2s

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #4 Simon Fraser  Overall, the Tier-2s and the experiments considered the CCRC’08 exercise to be a success  The networking/data transfers were tested extensively; some FTS tuning was needed, but it worked out  Experiments tended to continue other activities in parallel which is a good test of the system, although the load was not as high as anticipated  While CMS did include significant user analysis activities, the chaotic use of the Grid by a large number of inexperienced people is still to be tested Tier-2 Centres in CCRC’08 – General*

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #5 Simon Fraser Tier-2 Issues/Concerns As of CB and meetings with experiments this summer  Communications: Do Tier-2s have a voice? Is there a good mechanism for disseminating information?  Better monitoring: Pledges vs actual vs used  Hardware acquisitions: What should be bought? kSI2006?  Tier-2 capacity : Size of datasets? Effect of LHC delay?  …

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #6 Simon Fraser Tier-2 Issues/Concerns  Upcoming onslaught of users: Some user analysis tests have been done but scaling is a concern  User Support: Ticketing system exists but it is not really used for user support issues. This affects Tier-2s especially.  Federated Tier-2s: Tools to federate? Monitoring? (averaging)  Interoperability of EGEE, OSG, and NDGF should be improved  Software/Middleware updates: Could be smoother; too frequent

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #7 Simon Fraser Communications for Tier-2s  Identified by the T2s at the last CB as a serious problem. Interesting to me that many in experiment computing management did not share this concern.  Should communication be organized according to experiment or to Tier-1 association? There are also differing opinions on this.  There are two issues: Grid middleware/operations Experiment software  My view after studying this is that the situation is OK for “tightly coupled” Tier-2s, but not for remote and smaller Tier-2s that are not well coupled to a Tier-1.

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #8 Simon Fraser Communications for Tier-2s  Many lines of communication do indeed exist.  Some examples are: CMS has two Tier-2 coordinators: Ken Bloom (Nebraska) Giuseppe Bagliesi (INFN) - attend all operations meetings - feed T2 issues back to the operations group - write T2-relevant minutes - organize T2 workshops ALICE has designated 1 Core Offline person in 3 to have privileged contact with a given T2 site manager - weekly coordination meetings - Tier-2 federations provide a single contact person - A Tier-2 coordinates with its regional Tier-1

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #9 Simon Fraser Communications for Tier-2s ATLAS uses its cloud structure for communications - Every Tier-2 is coupled to a Tier national clouds; others have foreign members (e.g. “Germany” includes Krakow, Prague, Switzerland; Netherlands includes Russia, Israel, Turkey) - Each cloud has a Tier-2 coordinator Regional organizations, such as: + France Tier-2/3 technical group: - coordinates with Tier-1 and with experiments - monthly meetings - coordinates procurement and site management + GRIF: Tier-2 federation of 5 labs around Paris + Canada: Weekly teleconferences of technical personnel (T1 & T2) to share information and prepare for upgrades, large production, etc. + Many others exist; e.g. in the US and the UK

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #10 Simon Fraser Communications for Tier-2s  Tier-2 Overview Board reps: Michel Jouvin and Atul Gurtu have just been appointed to the OB to give the Tier-2s a voice there.  Tier-2 mailing list: Actually exists and is being reviewed for completeness & accuracy  Tier-2 GDB: The October GDB was dedicated to Tier-2 issues + reports from experiments: role of the T2s; communications + talks on regional organizations + discussion of accounting + technical talks on storage, batch systems, middleware  Seems to have been a success; repeat a couple of times per year?

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #11 Simon Fraser

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #12 Simon Fraser

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #13 Simon Fraser  But how much of this is a problem of under-use rather than under-contribution?  a task force has been set up to extract installed capacities from the Glue schema  Monthly APEL reports still undergo significant modifications from first draft.  Good because communication with T2s better  Bad because APEL accounting still has problems Accounting seems to be very finicky; breaks when the CE or MON box is upgraded  How are jobs distributed to the Tier-2s? Tier-2 Installed Resources

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #14 Simon Fraser  How does the LHC delay affect the requirements and pledges for 2009? + We are told to go ahead and buy what was planned but we have already seen some under-use of CPU capacity and we have seen this for storage as well Tier-2 Hardware Questions

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #15 Simon Fraser  How does the LHC delay affect the requirements and pledges for 2009? + We are told to go ahead and buy what was planned but we have already seen some under-use of CPU and we are now starting to see this for storage as well  We need to use something other than SpecInt2000! + this benchmark is totally out-of-date & useless for new CPUs + continued delays in SpecHEP can cause sub-optimal decisions Tier-2 Hardware Questions

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #16 Simon Fraser  Networking to the nodes is now an issue. + with 8 cores per node, 1 GigE connection ≈ 16.8 MB/sec/core + Tier-2 analysis jobs run on reduced data sets and can do rather simple operations  have seen 7.5 MB/sec at ATLAS and much more (x10?) + Do we need to go to Infiniband? + We certainly need increased capability for the uplinks; we should have a minimum of fully non-blocking GigE the worker nodes.  We need more guidance from the experiments The next round of purchases is now! Tier-2 Hardware Questions

M.C. Vetterli – WLCG-OB, CERN; October 27, 2008 – #17 Simon Fraser Summary  The role of the Tier-2 centres has increased markedly in the last year  >50% of Monte Carlo simulation is done in the T2s now.  The CCRC’08 exercise is considered a success by the Tier2s and by the experiments.  Availability and reliability are up, but still need improvement.  Resource acquisition vs pledges is better but still needs work  Issues for Tier2s: - communication should be (& is being) improved - work should ramp up on chaotic user analysis - reporting actual resources should be established - improved user support is needed