James Casey, IT-GD, CERN CERN, 5th September 2005

Slides:



Advertisements
Similar presentations
Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
Advertisements

EGEE is a project funded by the European Union under contract IST Using SRM: DPM and dCache G.Donvito,V.Spinoso INFN Bari
Service Challenge Meeting “Service Challenge 2 update” James Casey, IT-GD, CERN IN2P3, Lyon, 15 March 2005.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
Sharing Information across Congestion Windows CSE222A Project Presentation March 15, 2005 Apurva Sharma.
LHC Data Challenges and Physics Analysis Jim Shank Boston University VI DOSAR Workshop 16 Sept., 2005.
Data management for ATLAS, ALICE and VOCE in the Czech Republic L.Fiala, J. Chudoba, J. Kosina, J. Krasova, M. Lokajicek, J. Svec, J. Kmunicek, D. Kouril,
Data transfer over the wide area network with a large round trip time H. Matsunaga, T. Isobe, T. Mashimo, H. Sakamoto, I. Ueda International Center for.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
WLCG Service Report ~~~ WLCG Management Board, 1 st September
BNL Service Challenge 3 Site Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Andrea Sciabà CERN CMS availability in December Critical services  CE, SRMv2 (since December) Critical tests  CE: job submission (run by CMS), CA certs.
Testing the UK Tier 2 Data Storage and Transfer Infrastructure C. Brew (RAL) Y. Coppens (Birmingham), G. Cowen (Edinburgh) & J. Ferguson (Glasgow) 9-13.
Optimisation of Grid Enabled Storage at Small Sites Jamie K. Ferguson University of Glasgow – Jamie K. Ferguson – University.
CERN Using the SAM framework for the CMS specific tests Andrea Sciabà System Analysis WG Meeting 15 November, 2007.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
BNL Wide Area Data Transfer for RHIC & ATLAS: Experience and Plans Bruce G. Gibbard CHEP 2006 Mumbai, India.
End-to-End performance tuning Brian Davies Gridpp28 Manchester 2012.
BNL Facility Status and Service Challenge 3 HEPiX Karlsruhe, Germany May 9~13, 2005 Zhenping Liu, Razvan Popescu, and Dantong Yu USATLAS/RHIC Computing.
HEPiX FNAL ‘02 25 th Oct 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 25 th October 2002 HEPiX 2002, FNAL.
Alberto Aimar CERN – LCG1 Reliability Reports – May 2007
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Using Heterogeneous Paths for Inter-process Communication in a Distributed System Vimi Puthen Veetil Instructor: Pekka Heikkinen M.Sc.(Tech.) Nokia Siemens.
ATLAS Bulk Pre-stageing Tests Graeme Stewart University of Glasgow.
Status SC3 SARA/Nikhef 20 juli Status & results SC3 throughput phase SARA/Nikhef Mark van de Sanden.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
SC4 Planning Planning for the Initial LCG Service September 2005.
GGUS Slides for the 2012/07/24 MB Drills cover the period of 2012/06/18 (Monday) until 2012/07/12 given my holiday starting the following weekend. Remove.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Busy Storage Services Flavia Donno CERN/IT-GS WLCG Management Board, CERN 10 March 2009.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
TCP continued. Discussion – TCP Throughput TCP will most likely generate the saw tooth type of traffic. – A rough estimate is that the congestion window.
Report from GSSD Storage Workshop Flavia Donno CERN WLCG GDB 4 July 2007.
WLCG Latency Mesh Comments + – It can be done, works consistently and already provides useful data – Latency mesh stable, once configured sonars are stable.
SRM-2 Road Map and CASTOR Certification Shaun de Witt 3/3/08.
LCG Storage Workshop “Service Challenge 2 Review” James Casey, IT-GD, CERN CERN, 5th April 2005.
Service Challenge Meeting “Review of Service Challenge 1” James Casey, IT-GD, CERN RAL, 26 January 2005.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
HEPiX Rome – April 2006 The High Energy Data Pump A Survey of State-of-the-Art Hardware & Software Solutions Martin Gasthuber / DESY Graeme Stewart / Glasgow.
WLCG Service Report Jean-Philippe Baud ~~~ WLCG Management Board, 24 th August
CASTOR in SC Operational aspects Vladimír Bahyl CERN IT-FIO 3 2.
Summary of SC4 Disk-Disk Transfers LCG MB, April Jamie Shiers, CERN.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
CERN IT Department CH-1211 Genève 23 Switzerland t Towards end-to-end debugging for data transfers Gavin McCance Javier Conejero Banon Sophie.
Baseline Services Group Status of File Transfer Service discussions Storage Management Workshop 6 th April 2005 Ian Bird IT/GD.
Kevin Thaddeus Flood University of Wisconsin
“A Data Movement Service for the LHC”
LCG Service Challenge: Planning and Milestones
Cross-site problem resolution Focus on reliable file transfer service
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Service Challenge 3 CERN
3D Application Tests Application test proposals
Jan 12, 2005 Improving CMS data transfers among its distributed Computing Facilities N. Magini CERN IT-ES-VOS, Geneva, Switzerland J. Flix Port d'Informació.
Database Readiness Workshop Intro & Goals
BNL FTS services Hironori Ito.
Abhishek Singh Rana UC San Diego
Farida Fassi, Damien Mercie
WLCG Management Board, 16th July 2013
Introduction to Networks
Data Management cluster summary
Grid Canada Testbed using HEP applications
lundi 25 février 2019 FTS configuration
Network performance issues recently raised at IN2P3-CC
Performance-Robust Parallel I/O
Dirk Duellmann ~~~ WLCG Management Board, 27th July 2010
The LHCb Computing Data Challenge DC06
Presentation transcript:

James Casey, IT-GD, CERN CERN, 5th September 2005 Grid Deployment Board “SC3 Debugging” James Casey, IT-GD, CERN CERN, 5th September 2005

Overview Open Issues from SC3 First results from debugging phase Long latency network issues Procedures Transfer issues - best practices for #streams & rates First results from debugging phase Summary CERN IT-GD

Problem 1 Performance on transatlantic networks Very slow per-file transfer rate (~1-2MB/s) Even when multi stream (10/20) Solution is to put a lot of files onto the network at once BNL achieved 150MB/s but with 75 concurrent files We see a lot of timeouts happening FTS retries and the transfers have a high success rate but we lose effective bandwidth These sites have a lot of bandwidth that we don’t use e.g. ASCC have 2G/s but it’s hard to fill even with TCP based iperf Q: How do we up the single file transfer rate on transatlantic sites? Do we need to go back to per site network tuning? CERN IT-GD

Solution 1 (1/2) Sites have verbally reported that 2.6 kernels perform better for them than 2.4 kernels FNAL, … We are placing a SL4 node in the WAN area to start to test this Initially IA32, and will then test with IA64 too We will test the new TCP stacks that come by default with the RHEL4 kernel BIC, Westwood We will investigate the usage of FAST in this kernel With support from the FAST team at CalTech and CERN CS and ADC groups CERN IT-GD

Solution 1 (2/2) Proposal is to place a “reference” node at each T-1 site ‘standard’ OS installed – i.e. SL3/4 The DPM software would be installed on it This can be used for system debugging, network tuning and regression tests This would be used to run background iperf tests and file replication for regression analysis Can remove a link from the transfer chain to eliminate source or destination SRM as cause of problem Does not have to be most up-to-date hardware But good network connectivity and NIC essential CERN IT-GD

Problem 2 SRM cleanup procedures are not understood Often we see something going wrong on the transfers and we diagnose and solve the problem e.g. all allocated transfers have timed but movers not cleaned up But the effect tends to go on longer We see degraded performance afterwards and often the sites ends up just rebooting everything Q: How can we create, document and share standard procedures, so we don’t have to reinvent the wheel 11 times? CERN IT-GD

Solution 2 dCache workshop held last week at DESY to share knowledge Maarten Litmaath will report on it Plans to hold similar event at CHEP covering all the SRM systems CERN IT-GD

Problem 3 During SC2, we tended to run with few transfers and a single stream per transfer INFN – 10 single stream file transfers 100MB/s FZK – 3 single stream file transfers – 150MB/s Now we don’t see this INFN has good file transfer rates (~10-15MB/s) but we only get 60% utilization of the network FZK sees very low file transfer rate (~1-2MB/s) for many file transfers (but some seem to run much faster) PIC (& IN2P3/SARA) work best when doing 10 concurrent streams Q: How can we reduce number of streams and get individual file rate higher (and more stable) ? CERN IT-GD

Debugging Phase Tackled the third problem: How can we get higher and more reliable file transfer rates? Looked to answer several questions : What is an ideal node kernel tuning? How many streams are best? What is effect of using SRM Copy? Restricted to low-latency sites since network issues seem to play bigger role in high latency network routes Tested with DESY to see how a well-tuned system should behave Comparative results against INFN for CASTOR CERN IT-GD

CERN-DESY with FTS (10 files) With dCache transfer rate does not seem to scale with no. streams. “# streams x #files ~ 50” CERN IT-GD

CERN-INFN with FTS (10 files) Slight increase with no of streams (fixed to 10 concurrent files) But total bandwidth did not translate to ~20MB x 10 – was in the range of 60-80MB/s. CERN IT-GD

CERN-DESY with srmcp We tended to fill bandwidth but single file bandwidth inv. prop. to # streams CASTOR returns all TURLs immediately, so dCache transfers them Resource management needs to be done on both sides CERN IT-GD

Srmcp distribution by dest host Note effect of different TCP buffer sizes 22+29 had 64K buffers, the rest had 2M buffers CERN IT-GD

Monitoring ongoing transfers FTS used gridftp performance markers Has 120 seconds marker-to-marker timeout Has global transfer time set much higher (~1hr) dCache does not send the performance markers This initially caused all long-hop transfers to time out Have to disable this feature Had the effect of if any problem occurs, it takes 1hr to fail ! Bad for channel utilization dCache developers have promised to implement this feature CERN IT-GD

Failure of DPM pool node 1 DPM pool node out of 6 started to fail on gridftp SRM kept scheduling to that node Reminiscent of Globus gridftp black holes from SC2 Rate drops from 150MB/s to 80MB/s CERN IT-GD

Misc Issues dCache 1.6.5 had a problem with pool balancing Fixed in 1.6.6 This reduced rates, and explain some effects we saw at SARA, IN2P3 FZK had problems with ext3 file systems Moved to GPFS file systems – now can run at up to 250MB/s SARA incresed transfer rate to 160MB/s using 3 nodes by throttling #transfers in dCache Allow a large number of FTS file transfers (~20) but throttle each pool node to a small number of movers (~3) This leads to transfers “bunching up” before gridftp Overcomes some of the latencies involved with SRM CERN IT-GD

Post-debugging Now can achieve same rate as before with fewer sites Still need to add in other sites, and see how what the new upper limit it CERN IT-GD

Summary Started to tackle the problems Added some knowledge Especially in regards to rates to individual sites Added some knowledge 5 streams is a good number for 10 concurrent files with FTS But dCache does seem capable of running high speed single stream transfers Srmcp gives better load balancing over door nodes With FTS, all pool nodes were used for storage, but door node usage wasn’t balanced But throttling needed in other SRM implementations to stop dCache overloading them #files x file transfer rate != throughput Significant lossage, due to SRM overhead and FTS scheduling CERN IT-GD