Rationalizing Tier 2 Traffic and Utilizing the Existing Resources (T/A Bandwidth) What We Have Here is a Traffic Engineering Problem LHC Tier 2 Technical.

Slides:



Advertisements
Similar presentations
Ethernet Switch Features Important to EtherNet/IP
Advertisements

Bandwidth Management and Optimisation (BMO): Policy development workshop Overview of Challenges and Solutions.
M A Wajid Tanveer Infrastructure M A Wajid Tanveer
1 EL736 Communications Networks II: Design and Algorithms Class3: Network Design Modeling Yong Liu 09/19/2007.
Ethernet and switches selected topics 1. Agenda Scaling ethernet infrastructure VLANs 2.
Dynamically Provisioned Networks as a Substrate for Science David Foster CERN.
Trial of the Infinera PXM Guy Roberts, Mian Usman.
Traffic Engineering Jennifer Rexford Advanced Computer Networks Tuesdays/Thursdays 1:30pm-2:50pm.
1 Chapter 8 Local Area Networks - Internetworking.
1 Chapter 10 Introduction to Metropolitan Area Networks and Wide Area Networks Data Communications and Computer Networks: A Business User’s Approach.
ESnet On-demand Secure Circuits and Advance Reservation System (OSCARS) Chin Guok Network Engineering Group Thomas Ndousse Visit February Energy.
Questionaire answers D. Petravick P. Demar FNAL. 7/14/05 DLP -- GDB2 FNAL/T1 issues In interpreting the T0/T1 document how do the T1s foresee to connect.
Global Connectivity Joint venture of two workshops Kees Neggers & Dany Vandromme e-IRG Workshop Amsterdam, 13 May 2005.
© 2006 Cisco Systems, Inc. All rights reserved.Cisco Public 1 Version 4.0 4: Addressing in an Enterprise Network Introducing Routing and Switching in the.
NOC Lessons Learned TEIN2 and CERNET Xing Li
Dr. John P. Abraham Professor University of Texas Pan American Internet Routing and Routing Protocols.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
1 ESnet Network Measurements ESCC Feb Joe Metzger
Circuit Services - IPTV Christian Todorov Internet2 Fall Member Meeting October 9, 2007.
LHCOPN & LHCONE Network View Joe Metzger Network Engineering, ESnet LHC Workshop CERN February 10th, 2014.
ESnet Abilene 3+3 Measurements Presented at the Joint Techs Meeting in Columbus July 19 th 2004 Joe Metzger ESnet Network Engineer
Fermi National Accelerator Laboratory, U.S.A. Brookhaven National Laboratory, U.S.A, Karlsruhe Institute of Technology, Germany CHEP 2012, New York, U.S.A.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Jerry Sobieski MAX GigaPoP and TSC Program Manager Mark Johnson MCNC.
Using E2E technology for LHC Apr 3, 2006 HEPiX Spring Meeting 2006
Rick Summerhill Chief Technology Officer, Internet2 Internet2 Fall Member Meeting 9 October 2007 San Diego, CA The Dynamic Circuit.
ACN: RED paper1 Random Early Detection Gateways for Congestion Avoidance Sally Floyd and Van Jacobson, IEEE Transactions on Networking, Vol.1, No. 4, (Aug.
US LHC Tier-1 WAN Data Movement Security Architectures Phil DeMar (FNAL); Scott Bradley (BNL)
InterDomain Dynamic Circuit Network Demo Joint Techs - Hawaii Jan 2008 John Vollbrecht, Internet2
Thoughts on Future LHCOPN Some ideas Artur Barczyk, Vancouver, 31/08/09.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
LHC Open Network Environment LHCONE David Foster CERN IT LCG OB 30th September
Connect communicate collaborate GÉANT3 Services Connectivity and Monitoring Services by and for NRENs Ann Harding, SWITCH TNC 2010.
ASCR/ESnet Network Requirements an Internet2 Perspective 2009 ASCR/ESnet Network Requirements Workshop April 15/16, 2009 Richard Carlson -- Internet2.
TeraPaths TeraPaths: Establishing End-to-End QoS Paths through L2 and L3 WAN Connections Presented by Presented by Dimitrios Katramatos, BNL Dimitrios.
Brookhaven Science Associates U.S. Department of Energy 1 Network Services BNL USATLAS Tier 1 / Tier 2 Meeting John Bigrow December 14, 2005.
1 Network Measurement Summary ESCC, Feb Joe Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
Routing integrity in a world of Bandwidth on Demand Dave Wilson DW238-RIPE
TeraPaths The TeraPaths Collaboration Presented by Presented by Dimitrios Katramatos, BNL Dimitrios Katramatos, BNL.
Summary - Part 2 - Objectives The purpose of this basic IP technology training is to explain video over IP network. This training describes how video can.
NORDUnet Nordic Infrastructure for Research & Education Workshop Introduction - Finding the Match Lars Fischer LHCONE Workshop CERN, December 2012.
LHC Open Network Environment Architecture Overview and Status Artur Barczyk/Caltech LHCONE meeting Amsterdam, September 26 th,
OSCARS Roadmap Chin Guok Feb 6, 2009 Energy Sciences Network Lawrence Berkeley National Laboratory Networking for the Future of.
ATLAS Network Requirements – an Open Discussion ATLAS Distributed Computing Technical Interchange Meeting University of Tokyo.
Internet2 Joint Techs Workshop, Feb 15, 2005, Salt Lake City, Utah ESnet On-Demand Secure Circuits and Advance Reservation System (OSCARS) Chin Guok
Dynamic Circuit Network An Introduction John Vollbrecht, Internet2 May 26, 2008.
LHCONE Point-to-Point Circuit Experiment Authentication and Authorization Model Discussion LHCONE meeting, Rome April 28-29, 2014 W. Johnston, Senior Scientist.
Connect communicate collaborate LHCONE Diagnostic & Monitoring Infrastructure Richard Hughes-Jones DANTE Delivery of Advanced Network Technology to Europe.
Point-to-point Architecture topics for discussion Remote I/O as a data access scenario Remote I/O is a scenario that, for the first time, puts the WAN.
SDN and OSCARS how-to Evangelos Chaniotakis Network Engineering Group ESCC Indianapoilis, July 2009 Energy Sciences Network Lawrence Berkeley National.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
Multiple Protocol Support: Multiprotocol Level Switching.
Brookhaven Science Associates U.S. Department of Energy 1 Network Services LHC OPN Networking at BNL Summer 2006 Internet 2 Joint Techs John Bigrow July.
DICE: Authorizing Dynamic Networks for VOs Jeff W. Boote Senior Network Software Engineer, Internet2 Cándido Rodríguez Montes RedIRIS TNC2009 Malaga, Spain.
Inter-domain Routing Outline Border Gateway Protocol.
Strawman LHCONE Point to Point Experiment Plan LHCONE meeting Paris, June 17-18, 2013.
From the Transatlantic Networking Workshop to the DAM Jamboree David Foster CERN-IT.
A Strawman for Merging LHCOPN and LHCONE infrastructure LHCOPN + LHCONE Meeting Washington, DC, Jan. 31, 2013 W. E. Johnston and Chin Guok.
DICE Diagnostic Service Joe Metzger Joint Techs Measurement Working Group January
Run - II Networks Run-II Computing Review 9/13/04 Phil DeMar Networks Section Head.
100GE Upgrades at FNAL Phil DeMar; Andrey Bobyshev CHEP 2015 April 14, 2015.
Multiprotocol Label Switching (MPLS) Routing algorithms provide support for performance goals – Distributed and dynamic React to congestion Load balance.
1 Deploying Measurement Systems in ESnet Joint Techs, Feb Joseph Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
1 Network Measurement Challenges LHC E2E Network Research Meeting October 25 th 2006 Joe Metzger Version 1.1.
T0-T1 Networking Meeting 16th June Meeting
Lab A: Planning an Installation
Revisiting Ethernet: Plug-and-play made scalable and efficient
InterDomain Dynamic Circuit Network Demo
Networking for the Future of Science
What Are Routers? Routers are an intermediate system at the network layer that is used to connect networks together based on a common network layer protocol.
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
Presentation transcript:

Rationalizing Tier 2 Traffic and Utilizing the Existing Resources (T/A Bandwidth) What We Have Here is a Traffic Engineering Problem LHC Tier 2 Technical Meeting – CERN 13 January 2011 William E. Johnston Chin Guok, Joe Metzger, Kevin Oberman, Chris Tracy, et al

2 Rationalizing Tier 2 Networking The view of the problem from the U.S. is somewhat different than from Europe In the U.S. – The two Tier 1 centers are large: Fermilab holds about 40% of the CMS data and Brookhaven holds 100% of the ATLAS data – Fermi is on a dark fiber ring to StarLight and currently has 50G configured to the ESnet core – BNL currently has 40G to ESnet/MAN LAN, and within two months will be on an ESnet owned dark fiber ring – All of the Tier 2 centers are connected to either StarLight or MAN LAN with dedicated 10G circuits, and several have their own fiber – ESnet handles all Tier 2 traffic (world-wide) to and from Fermi and BNL ESnet sees all transatlantic traffic headed to Fermi and BNL ESnet sees none of the U.S. Tier 2 out-bound traffic as that is university traffic that is handled by the Regionals, Internet2, NLR, etc.

3 Rationalizing Tier 2 Networking Therefore, Tier 2 traffic within the U.S. is not now, nor is it every likely to be, an issue – Tier 3 traffic is still largely uncharacterized However, Tier 2 traffic across the Atlantic (in both directions) requires careful attention – the way the IP networking across the Atlantic is currently structured results in most general traffic (including almost all LHC non-OPN traffic) to use a small number of paths – the same paths used by most other R&E traffic – this situation will get better as the ACE infrastructure comes on-line, but Tier 2 traffic will be ramping up at the same time

4 The Need for Traffic Engineering – Example The LHC community has developed applications and tools that enable very high network data transfer rates over intercontential distances This is necessary in order to accomplish their science On the LHC OPN – a private optical network designed to facilitate data transfers from Tier 0 (CERN) to Tier 1 (National experiment data centers) – the HEP data transfer tools are essential – These tools are mostly parallel data movers – typically GridFTP – The related applications run on hosts that have modern TCP stacks that are appropriately tuned for high latency WAN transfers (e.g. international networks) – the Tier 2 sites use the same highly tuned WAN transfer software

5 The Need for Traffic Engineering – Example Recently, the Tier 2 (mostly physics analysis groups at universities) have abandoned the old hierarchical data distribution model – Tier 0 -> Tier 1 -> Tier 2, with attendant data volume reductions as you move down the hierarchy in favor of a chaotic model – get whatever data you need from wherever it is available This has resulted in enormous site to site data flows on the general IP infrastructure that have never been seen before apart from DDOS attacks

6 The Need for Traffic Engineering – Example GÉANT observed a big spike on their transatlantic peering connection with ESnet (9/2010) – headed for Fermilab – the U.S. CMS Tier 1 data center ESnet observed the same thing on their side Traffic, Gbps, at ESnet-GEANT Peering in New York Scale is 0 – 6.0 Gbps

7 The Need for Traffic Engineering – Example At FNAL is was apparent that the traffic was going to the UK Recalling that moving 10 TBy in 24 hours requires a data throughput of about 1 Gbps, the graph above implies 2.5 to 4+ Gbps of data throughput – which is what was being observed at the peering point

8 The Need for Traffic Engineering – Example Further digging revealed the site and nature of the traffic The nature of the traffic was – as expected – parallel data movers, but with an uncommonly high degree of parallelism: 33 hosts at the UK site and about 170 at FNAL

9 The Need for Traffic Engineering – Example This high degree of parallelism means that the largest host- host data flow rate is only about 2 Mbps, but in aggregate this data mover farm is doing 860 Mbps (seven day average) and has moved 65 TBytes of data – this also makes it hard to identify the sites involved by looking at all of the data flows at the peering point – nothing stands out as an obvious culprit THE ISSUE: This clever physics group is consuming 60% of the available bandwidth on the primary U.S. – Europe general R&E IP network link – for weeks at a time! This is obviously an unsustainable situation and this is the sort of thing that will force the R&E network operators to mark such traffic on the general IP network as scavenger to ensure other uses of the network

10 The Need for Traffic Engineering – Example In this case marking the traffic as scavenger probably would not have made much difference for the UK traffic (from a UK LHC Tier 2 center) as the net was not congested However, this is only one Tier 2 center operating during a period of relative quiet for the LHC - when other Tier 2s start doing this things will fall apart quickly and this will be bad news for everyone: – For the NOCs to identify and mark this traffic without impacting other traffic from the site is labor intensive – The Tier 2 physics groups would not be able to do their physics – It is the mission of the R&E networks to deal with this kind of traffic There are a number of ways to rationalize this traffic, but just marking it all scavenger is not one of them

11 Transatlantic Networking The transatlantic capacity available to the community is probably sufficient in the near term if it used optimally Current R&E T/A circuits (John S. Graham, Global Research NOC) The question is how to optimize the use of the available capacity – satisfy the LHC needs while accommodating all other R&E traffic at the same time this point is critical because the available non-OPN capacity is funded for the benefit of the entire R&E community, not just the LHC

12 Transatlantic Networking The question is how to optimize the use of the available capacity – satisfy the LHC needs while accommodating all other R&E traffic at the same time this point is critical because the available non-OPN capacity is funded for the benefit of the entire R&E community, not just the LHC

13 Roughly Today’s Situation T2 T1 E T3 T1 T2 E T3 T2 T1 E E E T2 T1 E ESnet Internet2 GÉANT NREN1 NREN2 MAN LAN MAX/DC StarLight AMS Paris

14 The Problem T2 T1 E T3 T1 T2 E T3 T2 T1 E E E T2 T1 E ESnet Internet2 GÉANT NREN1 NREN2 MAN LAN MAX/DC StarLight AMS Paris The default routing for both routed IP traffic and circuits overloads certain paths – the ingress load management problem

What you would like to do is spread the load to avoid congestion (e.g. at the ingress of net A) T/A net A Tier 2- A Tier 2-C Tier 2-D Tier 1-A Tier 1-B Net 2 ingr ess B1 T/A net B Tier 2 - B US Europe Loading = 90% Loading = 10% Path 1 Path 2 Loading = 50% ingr ess B2 ingr ess A1 Net 1 egr ess A Net 1 egr ess B Net 1

16 Traffic Engineering – Routed IP Traffic There are several ways that one could address this for IP traffic in a federated infrastructure such as we have now In a single domain such as net B, the operator can use MEDs that are dynamically established from transatlantic path loadings to direct traffic to a less loaded ingress point – e.g. ingress B1 vs. B2 in the figure In the case of balancing across several independent domains (e.g. net A and net B) then the source must redirect traffic away from a congested (though perhaps closer) ingress point – This should be able to be done with BGP local preferences to control the exit path from an edge router the local perfs would have to established and changed dynamically based on the loading of several available paths (e.g. forcing net 1, egress A traffic away from net A, ingress A1 and routing it to net B, ingress B2 – which is not the default route)

17 Traffic Engineering- Circuit Approaches One way to rationalize Tier 2 traffic is to set up virtual circuits that have guaranteed, but at the same time controlled, bandwidth that is isolated from general traffic, from the Tier 2 sites to the Tier 1 data centers – The number of such combinations per Tier 2 is probably relatively small (10s at most) due to the access patterns arising from the nature of the Tier 2 analysis interests and distribution of data in the Tier 1 centers This sort of multi-domain traffic engineering is what is done for almost all of the U.S. Tier 2 centers for accessing the U.S. Tier 1 centers – all of these circuits and all of the U.S. LHC OPN circuits are based on OSCARS virtual circuits With caveats, the DICE IDC protocol has the capability to do this in the international network arena

18 Traffic Engineering- Managing IDC Circuit Paths The situation with circuits is similar to the IP traffic problem: How to avoid “congestion” (fully committed paths is the circuit version of “congestion”) The inter-domain IDCs have a global view of available topology, but not the current state of utilization, so cannot route around congestion In the next figure, with net A at full capacity, a successful circuit request must find and use a longer than normal circuit between T2-A and T1-A, which the current version of the IDC will not do automatically

19 Traffic Engineering- Managing IDC Circuit Paths T/A net A Tier 2- A Tier 2-C Tier 2-D Tier 1-A Tier 1-B Net 2 T/A net B Tier 2 - B US Europe circuit utilization= 90% circuit utization = 10% circuit 2 circuit 3 circuit utization = 50% in gress B2 Net 1 egress B Net 1 in gress A1 ingr ess B1 circuit 1 Tier 2-A and Tier 2-B circuits have exhausted the default IDC route capacity. Any further circuits will have to take non-default paths (e.g. circuit 3). Net 1 egress A

20 Traffic Engineering- Managing IDC Circuit Paths Even though the inter-domain IDCP cannot find a path from T2-A to T1-A, the path exists (“circuit 3”) The hop-by-hop circuit path is defined by an MPLS construct called an Explicit Route Object (ERO) The IDC can return the ERO to the user, and the user can modify it and use the modified version to define a path (assuming it represents a valid path) Currently perfSONAR can return path utilization information on a by-node basis – this information can be used to manually modify an ERO to represent an alternate path that is not “congested” (i.e. has capacity for the requested circuit) – however, perfSONAR cannot report on temporal circuit commitments on the path – this is being worked on This sounds like a “heavy weight” approach, and it is, but not impractically so if the circuit will be long-lived, as almost all production circuits are – DICE group is looking at tools to simplify the process – automation of the process is an active research topic that is seeing some progress

21 Traffic Engineering as a Solution for Tier 2 T/A Traffic At some level, adequate transatlantic R&E capacity is sufficient for the near future, if it can be managed in a federated way that distributes the LHC load across the available capacity Tools exist to accomplish – or at least to prototype – this approach Can a suitable federation be established? – Probably if an acceptable governance model can be agreed to that addresses capacity sharing and operational cooperation