100G R&D at Fermilab Gabriele Garzoglio (for the High Throughput Data Program team) Grid and Cloud Computing Department Computing Sector, Fermilab Overview.

Slides:



Advertisements
Similar presentations
Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
Advertisements

Storage System Integration with High Performance Networks Jon Bakken and Don Petravick FNAL.
Big Data over a 100G Network at Fermilab Gabriele Garzoglio Grid and Cloud Services Department Computing Sector, Fermilab CHEP 2013 – Oct 15, 2013.
Experience and proposal for 100 GE R&D at Fermilab Interactomes – May 22, 2012 Gabriele Garzoglio Grid and Cloud Computing Department Computing Sector,
High Throughput Data Program at Fermilab R&D Parag Mhashilkar Grid and Cloud Computing Department Computing Sector, Fermilab Network Planning for ESnet/Internet2/OSG.
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
US CMS Tier1 Facility Network Andrey Bobyshev (FNAL) Phil DeMar (FNAL) CHEP 2010 Academia Sinica Taipei, Taiwan.
CHEPREO Tier-3 Center Achievements. FIU Tier-3 Center Tier-3 Centers in the CMS computing model –Primarily employed in support of local CMS physics community.
FNAL Site Perspective on LHCOPN & LHCONE Future Directions Phil DeMar (FNAL) February 10, 2014.
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
Esma Yildirim Department of Computer Engineering Fatih University Istanbul, Turkey DATACLOUD 2013.
ATLAS computing in Geneva Szymon Gadomski, NDGF meeting, September 2009 S. Gadomski, ”ATLAS computing in Geneva", NDGF, Sept 091 the Geneva ATLAS Tier-3.
GridPP meeting Feb 03 R. Hughes-Jones Manchester WP7 Networking Richard Hughes-Jones.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Questionaire answers D. Petravick P. Demar FNAL. 7/14/05 DLP -- GDB2 FNAL/T1 issues In interpreting the T0/T1 document how do the T1s foresee to connect.
CMS Data Transfer Challenges LHCOPN-LHCONE meeting Michigan, Sept 15/16th, 2014 Azher Mughal Caltech.
CD FY09 Tactical Plan Review FY09 Tactical Plan for Wide-Area Networking Phil DeMar 9/25/2008.
Fermi National Accelerator Laboratory 3 Fermi National Accelerator Laboratory Mission Advances the understanding of the fundamental nature of matter.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Tier 3 Data Management, Tier 3 Rucio Caches Doug Benjamin Duke University.
100G R&D at Fermilab Gabriele Garzoglio Grid and Cloud Computing Department Computing Sector, Fermilab Overview Fermilab Network R&D 100G Infrastructure.
100G R&D at Fermilab Gabriele Garzoglio Grid and Cloud Computing Department Computing Sector, Fermilab Overview Fermilab Network R&D 100G Infrastructure.
GlobusWorld 2012: Experience with EXPERIENCE WITH GLOBUS ONLINE AT FERMILAB Gabriele Garzoglio Computing Sector Fermi National Accelerator.
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
DYNES Storage Infrastructure Artur Barczyk California Institute of Technology LHCOPN Meeting Geneva, October 07, 2010.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
BNL Wide Area Data Transfer for RHIC & ATLAS: Experience and Plans Bruce G. Gibbard CHEP 2006 Mumbai, India.
Spectrum of Support for Data Movement and Analysis in Big Data Science Network Management and Control E-Center & ESCPS Network Management and Control E-Center.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
From the Transatlantic Networking Workshop to the DAM Jamboree to the LHCOPN Meeting (Geneva-Amsterdam-Barcelona) David Foster CERN-IT.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CMS Computing Model Simulation Stephen Gowdy/FNAL 30th April 2015CMS Computing Model Simulation1.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Storage and Data Movement at FNAL D. Petravick CHEP 2003.
Slide 1 UCSC 100 Gbps Science DMZ – 1 year 9 month Update Brad Smith & Mary Doyle.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
High Throughput Data Program (HTDP) at FNAL Mission: investigate the impact of and provide solutions for the scientific computing challenges in Big Data.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
CRISP WP18, High-speed data recording Krzysztof Wrona, European XFEL PSI, 18 March 2013.
100G R&D for Big Data at Fermilab Gabriele Garzoglio Grid and Cloud Computing Department Computing Sector, Fermilab ISGC – March 22, 2013 Overview Fermilab.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
Run - II Networks Run-II Computing Review 9/13/04 Phil DeMar Networks Section Head.
100GE Upgrades at FNAL Phil DeMar; Andrey Bobyshev CHEP 2015 April 14, 2015.
Supporting Advanced Scientific Computing Research Basic Energy Sciences Biological and Environmental Research Fusion Energy Sciences High Energy Physics.
100G R&D at Fermilab Gabriele Garzoglio (for the High Throughput Data Program team) Grid and Cloud Computing Department Computing Sector, Fermilab Overview.
Big Data over a 100G Network at Fermilab Gabriele Garzoglio Grid and Cloud Services Department Computing Sector, Fermilab CHEP 2013 – Oct 15, 2013 Overview.
Fermilab Cal Tech Lambda Station High-Performance Network Research PI Meeting BNL Phil DeMar September 29, 2005.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Dynamic Extension of the INFN Tier-1 on external resources
“A Data Movement Service for the LHC”
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Provisioning 160,000 cores with HEPCloud at SC17
DOE Facilities - Drivers for Science: Experimental and Simulation Data
Enabling High Speed Data Transfer in High Energy Physics
Presentation transcript:

100G R&D at Fermilab Gabriele Garzoglio (for the High Throughput Data Program team) Grid and Cloud Computing Department Computing Sector, Fermilab Overview Fermilab’s interest in 100 GE Results from the ESNet 100GE testbed 100 GE infrastructure at Fermilab

Fermilab Users and 100 GE Decades long dependence on sustained, high speed, large and wide-scale distribution of and access to data  High Energy Physics community  Multi-disciplinary communities using grids (OSG, XSEDE) Figures of merit  40 Petabytes on tape, today mostly coming from offsite  140Gbps LAN traffic from archive to local processing farms  LHC peak WAN usage at Gbps 2 Compact Muon Solenoid (CMS) routinely peaks at Gbps. 94 PB of data ever written to the Enstore tape archive – 54 PB available for retrieval 20 Gpbs CMS WAN Traffic (1 of 2 Dec, 2012 Written data: 94 PB 2 yrs G

Goals of 100 GE Program at Fermilab 3 End-to-end experiment analysis systems include a deep stack of software layers and services. Need to ensure these are functional and effective at the 100 GE scale.  Determine and tune the configuration to ensure full throughput in and across each layer/service.  Measure and determine efficiency of the end-to-end solutions.  Monitor, identify and mitigate error conditions.

High Throughput Data Program (HTDP) at Fermilab Mission: prepare the Computing Sector and its stakeholders for the 100GE infrastructure and put Fermilab in a strategic position of leadership. Establish collaborations with stakeholders, computing facilities, scientific communities, and institutions, to coordinate a synergistic program of work on 100GE. The program includes technological investigations, prototype development, and participation on funding agency solicitations. Close collaboration with the OSG network area to instrument the cyber-infrastructure (PerfSONAR) and provide nation-wide network metrics. The ANI has been the major testbed used since 2011 in close partnership with ESNet 4

Ongoing Program of Work 2011: ANI Long Island MAN (LIMAN) testbed.  GO / GridFTP over 3x10GE : Super Computing ’11  Fast access to ~30TB of CMS data in 1h from NERSC to ANL using GridFTP.  15 srv / 28 clnt – 4 gFTP / core; 2 strms; TCP Win. 2MB Currently: 100GE testbed (focus of this talk)  Tuning parameters of middleware for data movement: xrootd, GridFTP, SRM, Globus Online, Squid  Achieved ~97Gbps Spring 2013: 100GE Endpoint at Fermilab  Plan to repeat and extend tests using CMS current datasets Gbps (peak: 75)

GridFTP / SRM / GlobusOnline Tests 6 Data Movement using GridFTP  3 rd party Srv to Srv trans.: Src at NERSC / Dest at ANL  Dataset split into 3 size sets Large files transfer performance ~ 92Gbps Small files transfer performance - abysmally low Issues uncovered on 100G Testbed:  GridFTP Pipelining needs to be fixed on Globus implementation Optimal performance: 97 Gbps w/ GridFTP 2 GB files – 3 nodes x 16 streams / node 97 Gbps GO control latencies GO control channel sent to the VPN through port forwarding

XRootD Tests 7 Data Movement over XRootD, testing LHC experiment (CMS / Atlas) analysis use cases.  Clients at NERSC / Servers at ANL  Using RAMDisk as storage area on the server side Challenges  Tests limited by the size of RAMDisk  Little control over xrootd client / server tuning parameters Calculation of the scaling factor between 1 NIC and an aggregated 12 NIC for datasets too large to fit on the RAM disk

Squid / Frontier Tests 8 Data transfers  Cache 8 MB file on Squid – This size mimics LHC use case for large calib. data  Clients (wget) at NERSC / Servers at ANL  Data always on RAM Setup  Using Squid2: single threaded  Multiple servers per node (4 NIC per node)  Testing core affinity on/off: pin Squid to core i.e. to L2 cache  Testing all clnt noded vs. all servers AND aggregate one node vs. only one server Results  Core-affinity improve performance by 21% in some tests  Increasing the number of servers improves performance  Best performance w/ 9000 clients: ~100 Gbps

100GE production endpoint coming to Fermilab (see next slides)  Expecting 100 GE capabilities by Spring 2013  Connecting local cluster (FermiCloud Integration Testbed) to ESNet testbed  Validating 100GE link at Fermilab running measurements of middleware already tested on ANI. Continue testing of middleware technologies defined by stakeholders  Now planning measurements on NFS v4 for dCache on ESNet testbed 9 Plans

Current Fermilab WAN Capabilities Metropolitan Area Network provides 10GE channels:  Currently 8 deployed Five channels used for circuit traffic  Supports CMS WAN traffic Two used for normal routed IP traffic  Backup 10GE for redundancy  Circuits fail over to routed IP paths * Diagram courtesy of Phil Demar

Upcoming 100GE WAN capability ESnet deploying 100GE MAN as part of ESnet5  One 100GE wave for FNAL  Also 2x10GE channels for routed IP traffic 100GE wave will be used to support circuit traffic Legacy 10GE MAN will remain for diversity  Backup routed IP path  One 10GE circuit path * Diagram courtesy of Phil Demar

Use of 100GE Wave for FNAL R&D 100GE wave will support 4x10GE circuit paths  Excess capacity available for WAN R&D activities Planning ~6 x 10GE link to FNAL R&D network  Network will host 10GE test/development systems  Possibly 40GE systems later Anticipate WAN circuit into ESnet ANI test bed * Diagram courtesy of Phil Demar

Summary The High Throughput Data program at Fermilab is testing 100GE networks for its scientific stakeholders The collaboration with ANI and ESNet has been central to this program Fermilab will have 100GE capability in the Spring 2013 – planning for involvement with 100G ESNet testbed 13

FNAL VPN gateway FNAL test initiator Globus Online GridFTP control GO control Performance Analysis on ANI 100Gb/s Testbed NERSC ANL ANI Testbed VPN GridFT P data