Data Distribution Performance Hironori Ito Brookhaven National Laboratory.

Slides:



Advertisements
Similar presentations
SafetyFirsts Customer-Only Web Site Resources A Green Light to Safety Research, Articles, Specialty Reports & Training Presentations.
Advertisements

Toolbox Mirror -Overview Effective Distributed Learning.
Business Continuity and DR, A Practical Implementation Mich Talebzadeh, Consultant, Deutsche Bank
Performance Evaluation
31/08/05Network Traffic Management1 Network Traffic Management LtCdr Samit Mehra (05IT 6018) Guided By Dr. SK Ghosh.
Web Proxy Server Anagh Pathak Jesus Cervantes Henry Tjhen Luis Luna.
(part 3).  Switches, also known as switching hubs, have become an increasingly important part of our networking today, because when working with hubs,
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
ALICE DATA ACCESS MODEL Outline ALICE data access model - PtP Network Workshop 2  ALICE data model  Some figures.
1 Microsoft Distributed File System (Dfs) Brett O’Neill CSE 8343 – Group A6.
FZU participation in the Tier0 test CERN August 3, 2006.
ALICE data access WLCG data WG revival 4 October 2013.
1. There are different assistant software tools and methods that help in managing the network in different things such as: 1. Special management programs.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
Data management at T3s Hironori Ito Brookhaven National Laboratory.
Take on messages from Lecture 1 LHC Computing has been well sized to handle the production and analysis needs of LHC (very high data rates and throughputs)
Tier 3 Data Management, Tier 3 Rucio Caches Doug Benjamin Duke University.
Session-8 Data Management for Decision Support
FAX UPDATE 26 TH AUGUST Running issues FAX failover Moving to new AMQ server Informing on endpoint status Monitoring developments Monitoring validation.
DDM-Panda Issues Kaushik De University of Texas At Arlington DDM Workshop, BNL September 29, 2006.
Personal Computer - Stand- Alone Database  Database (or files) reside on a PC - on the hard disk.  Applications run on the same PC and directly access.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
BNL DDM Status Report Hironori Ito Brookhaven National Laboratory.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
The Alternative Larry Moore. 5 Nodes and Variant Input File Sizes Hadoop Alternative.
Update on replica management
EGEE-III INFSO-RI Enabling Grids for E-sciencE Overview of STEP09 monitoring issues Julia Andreeva, IT/GS STEP09 Postmortem.
Graeme Stewart: ATLAS Computing WLCG Workshop, Prague ATLAS Suspension and Downtime Procedures Graeme Stewart (for ATLAS Central Operations Team)
DDM Monitoring David Cameron Pedro Salgado Ricardo Rocha.
Efi.uchicago.edu ci.uchicago.edu FAX status developments performance future Rob Gardner Yang Wei Andrew Hanushevsky Ilija Vukotic.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
Network awareness and network as a resource (and its integration with WMS) Artem Petrosyan (University of Texas at Arlington) BigPanDA Workshop, CERN,
Replication (1). Topics r Why Replication? r System Model r Consistency Models r One approach to consistency management and dealing with failures.
8 th CIC on Duty meeting Krakow /2006 Enabling Grids for E-sciencE Feedback from SEE first COD shift Emanoil Atanassov Todor Gurov.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
Point-to-point Architecture topics for discussion Remote I/O as a data access scenario Remote I/O is a scenario that, for the first time, puts the WAN.
FTS monitoring work WLCG service reliability workshop November 2007 Alexander Uzhinskiy Andrey Nechaevskiy.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
ATLAS Distributed Computing ATLAS session WLCG pre-CHEP Workshop New York May 19-20, 2012 Alexei Klimentov Stephane Jezequel Ikuo Ueda For ATLAS Distributed.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Data Management at Tier-1 and Tier-2 Centers Hironori Ito Brookhaven National Laboratory US ATLAS Tier-2/Tier-3/OSG meeting March 2010.
Any Data, Anytime, Anywhere Dan Bradley representing the AAA Team At OSG All Hands Meeting March 2013, Indianapolis.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
Advanced Taverna Aleksandra Pawlik University of Manchester materials by Katy Wolstencroft, Aleksandra Pawlik, Alan Williams
Campana (CERN-IT/SDC), McKee (Michigan) 16 October 2013 Deployment of a WLCG network monitoring infrastructure based on the perfSONAR-PS technology.
Efi.uchicago.edu ci.uchicago.edu Sharing Network Resources Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago Federated Storage.
LHCONE Workshop Richard P Mount February 10, 2014 Concerns from Experiments ATLAS Richard P Mount SLAC National Accelerator Laboratory.
29/04/2008ALICE-FAIR Computing Meeting1 Resulting Figures of Performance Tests on I/O Intensive ALICE Analysis Jobs.
PanDA & Networking Kaushik De Univ. of Texas at Arlington UM July 31, 2013.
J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April ATLAS Grid Activities Preparing for Data Analysis Jim Shank.
Computing Operations Roadmap
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
Software Architecture in Practice
Future of WAN Access in ATLAS
ALICE Monitoring
Global Catalog and Flexible Single Master Operations (FSMO) Roles
BNL FTS services Hironori Ito.
The ADC Operations Story
XROOTd for cloud storage
Brookhaven National Laboratory Storage service Group Hironori Ito
Hironori Ito Brookhaven National Laboratory
Storage Virtualization
Any Data, Anytime, Anywhere
EE 122: Lecture 22 (Overlay Networks)
Presentation transcript:

Data Distribution Performance Hironori Ito Brookhaven National Laboratory

Old – Cloud centric Hierarchical distribution chain – Well tested and monitored. » Regular tests between BNL and T2s. T0->T1s->T2s – In US, BNL -> US T2s. 400MB/s at minimum with sustained level. – BNL can push at 20Gb/s now, and going 40GB/s. – Some T2s are able to show the sustain transfer rate at 10Gb/s – Pre-placement Data is placed well before the usage regardless of needs New – Non cloud centric T2Ds – T2s becomes part of all clouds – T2s are used as a source of data – Minimum pre-placement PD2P – Data is placed as needed. Change in distribution model

Further change A lot more smaller consumers. – T3s 50 (or more) T3s coming soon world wide – Individual users dq2-get xRootd: new type of requests – Although each of them are small, the combined requests could rival or exceed the level of T1 – Non-experienced users/sites administrators Unexpected, creative (not necessary efficient) use of DDM resources

Consequence of change More point-to-point transfers – T2s to all T1s: PANDA JOBS – T2s to all T1s/T2s: DDM distributions  T2s are no longer protected by cloud FTS.  The number of requests could dramatically increase at certain times (by factor of 10 or more) More on-demand transfers – Timely delivery of a particular data is more crucial Jobs/users won’t wait forever.

Current status Sonar tests – Regular tests by the central operation. – Data are sent from various T1s/T2s to many T1s/T2s. – Results are shown at cached/ cached/ – Results are used to evaluate eligibility of T2s to be promoted to T2D

Some Sonar results (MWT2)

More Sonar results (SWT2CPB)

Other tests US’s own regular tests are expanded for foreign T1s to T2. – Opposite directions is also considered. All T1s to BNL and BNL to all T1s are already conducted. – It is run twice a day using 10 large files to measure the network throughput. – The results are shown at the same as before (

Throughput test results (AGLT2)

Identification of a problem Who will find problematic transfers – Shifters Not capable of identifying any detail – Just looking at Green, Yellow, Red light at some place – Requesters: panda jobs, users, production Requesters always know when they don’t get what they want. – Sometimes, it is quite late to notice the problems. – There is usually some rudimentary effort to debug a problem by some experienced users. – The cause is almost always assumed to be the destination. » Eg. I can get DSN at site A, but not at B. Therefore, it must be B.  However, they don’t realized that it also does not work at sites C, D, E and F. – Site administrators

“Slow” transfer is notoriously difficult to find the cause. Generally, a ticket is sent to the destination simply because shifters are trained that way. And, it is something to do with the fact that DDM dashboard summarizes the transfer failures based on the destinations. Source and destination sites can only eliminate the performance of own sites as sources of problem. – A problem could be anywhere between the source and destination – Someone have to take ownership for quick resolution Who is responsible for fixing problematic transfers

Monitoring/debugging T2 administrators are completely blind in terms of debugging DDM transfers from own sites to foreign T1s/T2s. – CERN DQ2 logs are not visible remotely – T1s FTS log (except BNL) are not visible remotely For foreign T1s/T2s to own T2, you can still see them on the regular BNL DQ2 and FTS log viewers. Network performance between T2s and foreign T1s/T2s are not currently monitored. And, the traffic will compete the bandwidth with general traffic.

How to find, identify and resolve problematic transfers DDM dashboard will show the failed transfers Then… – Analyze the monitor more carefully. Is it failing for all sites or just one (or few) Are there any correlations with other failed sites? – Look at the results of the control data. Sonar and US throughput results? – US throughput test will provides BNL administrators the logs – Initiate a new test on demand. – Network issue. Traceroute, contact the network engineer, etc… Finding the location of problem is currently difficult!

Creative use of DDM clients by users Users generally do not understand physical limitations of getting data. – No network bandwidth limit – No disk IO bandwidth limit at source or destination Since no limitation, the use could be quite creative and destructive. – One user was found to be doing about simultaneous 1500 download from two hosts (150 dq2-get with 10 concurrency option) with one 1Gb/s NIC and writing to NFS area limited to about 70MB/s. Although it did not cause a problem at source sites, the load of two hosts doing dq2-get was very high, making client hosts unusable. He/she would have achieve the same performance with 1/100 th of number of concurrency without crashing the client hosts.

More problems are coming Good luck! Look at the existing information carefully to identify any problems. More tools and information would help. – Tests on demand US throughput page will include the ability for a site administrator to initiate the transfer on demand and look at the transfer logs. – PerfSonoar It would greatly help if the pure network performance results are available. Good and direct communications with network engineers and remote site administrators are needed. – Probably need help from T1s and T0