1 ESnet Network Requirements ASCAC Networking Sub-committee Meeting April 13, 2007 Eli Dart ESnet Engineering Group Lawrence Berkeley National Laboratory.

Slides:



Advertisements
Similar presentations
ONE PLANET ONE NETWORK A MILLION POSSIBILITIES Barry Joseph Director, Offer and Product Management.
Advertisements

U.S. Department of Energy’s Office of Science Basic Energy Sciences Advisory Committee Dr. Daniel A. Hitchcock October 21, 2003
High Performance Computing Course Notes Grid Computing.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21) NSF-wide Cyberinfrastructure Vision People, Sustainability, Innovation,
1 Chin Guok ESnet Network Engineer David Robertson DSD Computer Software Engineer Lawrence Berkeley National Laboratory.
1 ESnet Network Requirements ASCAC Networking Sub-committee Meeting April 13, 2007 Eli Dart ESnet Engineering Group Lawrence Berkeley National Laboratory.
1 Chin Guok ESnet Network Engineer David Robertson DSD Computer Software Engineer Lawrence Berkeley National Laboratory.
Marilyn T. Smith, Head, MIT Information Services & Technology DataSpace IS&T Data CenterMIT Optical Network 1.
1© Copyright 2015 EMC Corporation. All rights reserved. SDN INTELLIGENT NETWORKING IMPLICATIONS FOR END-TO-END INTERNETWORKING Simone Mangiante Senior.
Questionaire answers D. Petravick P. Demar FNAL. 7/14/05 DLP -- GDB2 FNAL/T1 issues In interpreting the T0/T1 document how do the T1s foresee to connect.
1 ESnet - Connecting the USA DOE Labs to the World of Science Eli Dart, Network Engineer Network Engineering Group Chinese American Network Symposium Indianapolis,
NORDUnet NORDUnet The Fibre Generation Lars Fischer CTO NORDUnet.
Reliable Data Movement Framework for Distributed Science Environments Raj Kettimuthu Argonne National Laboratory and The University of Chicago.
LHC Tier 2 Networking BOF Joe Metzger Joint Techs Vancouver 2005.
Applications Requirements Working Group HENP Networking Meeting June 1-2, 2001 Participants Larry Price Steven Wallace (co-ch)
GLIF Global Lambda Integrated Facility Maxine Brown Electronic Visualization Laboratory University of Illinois at Chicago.
LAN Switching and Wireless – Chapter 1
1 Nuclear Physics Network Requirements Workshop Washington, DC Eli Dart, Network Engineer ESnet Network Engineering Group May 6, 2008 Energy Sciences Network.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
GrangeNet Dr. Greg Wickham APAN NOC 25 August 2005.
Chapter 5 McGraw-Hill/Irwin Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved.
Office of Science U.S. Department of Energy ESCC Meeting July 21-23, 2004 Network Research Program Update Thomas D. Ndousse Program Manager Mathematical,
1 Network Measurement Summary ESCC, Feb Joe Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
Cyberinfrastructure and Internet2 Eric Boyd Deputy Technology Officer Internet2.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Advanced Networks: The Past and the Future – The Internet2 Perspective APAN 7 July 2004, Cairns, Australia Douglas Van Houweling, President & CEO Internet2.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
Keeping up with the RONses Mark Johnson Internet2 Member Meeting May 3, 2005.
Internet2 Joint Techs Workshop, Feb 15, 2005, Salt Lake City, Utah ESnet On-Demand Secure Circuits and Advance Reservation System (OSCARS) Chin Guok
Use Cases for High Bandwidth Query and Control of Core Networks Greg Bernstein, Grotto Networking Young Lee, Huawei draft-bernstein-alto-large-bandwidth-cases-00.txt.
Point-to-point Architecture topics for discussion Remote I/O as a data access scenario Remote I/O is a scenario that, for the first time, puts the WAN.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
NSF Middleware Initiative Purpose To design, develop, deploy and support a set of reusable, expandable set of middleware functions and services that benefit.
Cyberinfrastructure and Internet2 Eric Boyd Deputy Technology Officer Internet2.
By David P. Schissel and Reza Shakoori Presented at DOE Office of Science High-Performance Network Research PI Meeting Brookhaven National Lab September.
1 (Brief) Introductory Remarks On Behalf of the U.S. Department of Energy ESnet Site Coordinating Committee (ESCC) W.Scott Bradley ESCC Chairman
Strawman LHCONE Point to Point Experiment Plan LHCONE meeting Paris, June 17-18, 2013.
Internet2. Yesterday’s Internet  Thousands of users  Remote login, file transfer  Applications capitalize on underlying technology.
All Hands Meeting 2005 BIRN-CC: Building, Maintaining and Maturing a National Information Infrastructure to Enable and Advance Biomedical Research.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
100GE Upgrades at FNAL Phil DeMar; Andrey Bobyshev CHEP 2015 April 14, 2015.
Internet2 Applications & Engineering Ted Hanss Director, Applications Development.
Supporting Advanced Scientific Computing Research Basic Energy Sciences Biological and Environmental Research Fusion Energy Sciences High Energy Physics.
DutchGrid KNMI KUN Delft Leiden VU ASTRON WCW Utrecht Telin Amsterdam Many organizations in the Netherlands are very active in Grid usage and development,
1 Deploying Measurement Systems in ESnet Joint Techs, Feb Joseph Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
1 Network Measurement Challenges LHC E2E Network Research Meeting October 25 th 2006 Joe Metzger Version 1.1.
Internet2 Greg Wood Director of Communications Internet2 INET’99.
Chapter 1 Characterization of Distributed Systems
Clouds , Grids and Clusters
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
G R N E T-2 Update Tryfon Chiotis, Technical Director
Joint Techs, Columbus, OH
The NSRC cultivates collaboration among a community of peers to build and improve a global Internet that benefits all parties. We facilitate the growth.
DOE Facilities - Drivers for Science: Experimental and Simulation Data
A. Rama Bharathi Regd. No: 08931F0040 III M.C.A
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
Grid Computing.
Cyberinfrastructure and Internet2
GRID COMPUTING PRESENTED BY : Richa Chaudhary.
Introduction to Cloud Computing
ESnet Network Measurements ESCC Feb Joe Metzger
ESnet and Science DMZs: an update from the US
LHC Tier 2 Networking BOF
The roadmap towards the implementation of 5G in Europe
Wide-Area Networking at SLAC
Pfizer Internet2 Day Douglas E. Van Houweling President and CEO, UCAID
The New Internet2 Network: Expected Uses and Application Communities
The UltraLight Program
Internet2 Initiatives at NPS
Presentation transcript:

1 ESnet Network Requirements ASCAC Networking Sub-committee Meeting April 13, 2007 Eli Dart ESnet Engineering Group Lawrence Berkeley National Laboratory

2 Requirements from Instruments and Facilities This is the ‘hardware infrastructure’ of DOE science – types of requirements can be summarized as follows – Bandwidth: Quantity of data produced, requirements for timely movement – Connectivity: Geographic reach – location of instruments, facilities, and users plus network infrastructure involved (e.g. ESnet, Internet2, GEANT) – Services: Guaranteed bandwidth, traffic isolation, etc.; IP multicast Data rates and volumes from facilities and instruments – bandwidth, connectivity, services – Large supercomputer centers (NERSC, NLCF) – Large-scale science instruments (e.g. LHC, RHIC) – Other computational and data resources (clusters, data archives, etc.) Some instruments have special characteristics that must be addressed (e.g. Fusion) – bandwidth, services Next generation of experiments and facilities, and upgrades to existing facilities – bandwidth, connectivity, services – Addition of facilities increases bandwidth requirements – Existing facilities generate more data as they are upgraded – Reach of collaboration expands over time – New capabilities require advanced services

3 Requirements from Examining the Process of Science (1) The geographic extent and size of the user base of scientific collaboration is continuously expanding – DOE US and international collaborators rely on ESnet to reach DOE facilities – DOE Scientists rely on ESnet to reach non-DOE facilities nationally and internationally (e.g. LHC, ITER) – In the general case, the structure of modern scientific collaboration assumes the existence of a robust, high- performance network infrastructure interconnecting collaborators with each other and with the instruments and facilities they use – Therefore, close collaboration with other networks is essential for end-to-end service deployment, diagnostic transparency, etc. Robustness and stability (network reliability) are critical – Large-scale investment in science facilities and experiments makes network failure unacceptable when the experiments depend on the network – Dependence on the network is the general case

4 Requirements from Examining the Process of Science (2) Science requires several advanced network services for different purposes – Predictable latency, quality of service guarantees Remote real-time instrument control Computational steering Interactive visualization – Bandwidth guarantees and traffic isolation Large data transfers (potentially using TCP-unfriendly protocols) Network support for deadline scheduling of data transfers Science requires other services as well – for example – Federated Trust / Grid PKI for collaboration and middleware Grid Authentication credentials for DOE science (researchers, users, scientists, etc.) Federation of international Grid PKIs – Collaborations services such as audio and video conferencing

5 Science Network Requirements Aggregation Summary Science Drivers Science Areas / Facilities End2End Reliability Connectivity2006 End2End Band width 2010 End2End Band width Traffic Characteristics Network Services Advanced Light Source - DOE sites US Universities Industry 1 TB/day 300 Mbps 5 TB/day 1.5 Gbps Bulk data Remote control Guaranteed bandwidth PKI / Grid Bioinformatics- DOE sites US Universities 625 Mbps 12.5 Gbps in two years 250 Gbps Bulk data Remote control Point-to-multipoint Guaranteed bandwidth High-speed multicast Chemistry / Combustion - DOE sites US Universities Industry -10s of Gigabits per second Bulk data Guaranteed bandwidth PKI / Grid Climate Science - DOE sites US Universities International -5 PB per year 5 Gbps Bulk data Remote control Guaranteed bandwidth PKI / Grid High Energy Physics (LHC) % (Less than 4 hrs/year) US Tier1 (DOE) US Tier2 (Universities) International (Europe, Canada) 10 Gbps60 to 80 Gbps (30-40 Gbps per US Tier1) Bulk data Remote control Guaranteed bandwidth Traffic isolation PKI / Grid

6 Science Network Requirements Aggregation Summary Science Drivers Science Areas / Facilities End2End Reliability Connectivity2006 End2End Band width 2010 End2End Band width Traffic Characteristics Network Services Magnetic Fusion Energy % (Impossible without full redundancy) DOE sites US Universities Industry 200+ Mbps 1 Gbps Bulk data Remote control Guaranteed bandwidth Guaranteed QoS Deadline scheduling NERSC- DOE sites US Universities Industry International 10 Gbps20 to 40 Gbps Bulk data Remote control Guaranteed bandwidth Guaranteed QoS Deadline Scheduling PKI / Grid NLCF- DOE sites US Universities Industry International Backbone Band width parity Backbone band width parity Bulk data Nuclear Physics (RHIC) - DOE sites US Universities International 12 Gbps70 Gbps Bulk data Guaranteed bandwidth PKI / Grid Spallation Neutron Source High (24x7 operation) DOE sites640 Mbps2 Gbps Bulk data

7 LHC ATLAS Bandwidth Matrix as of April 2007 Site ASite ZESnet AESnet ZA-Z 2007 Bandwidth A-Z 2010 Bandwidth CERNBNLAofA (NYC)BNL10Gbps20-40Gbps BNLU. of Michigan (Calibration) BNL (LIMAN)Starlight (CHIMAN) 3Gbps10Gbps BNLBoston University BNL (LIMAN) Internet2 / NLR Peerings 3Gbps (Northeastern Tier2 Center) 10Gbps (Northeastern Tier2 Center) BNLHarvard University BNLIndiana U. at Bloomington BNL (LIMAN) Internet2 / NLR Peerings 3Gbps (Midwestern Tier2 Center) 10Gbps (Midwestern Tier2 Center) BNLU. of Chicago BNLLangston University BNL (LIMAN)Internet2 / NLR Peerings 3Gbps (Southwestern Tier2 Center) 10Gbps (Southwestern Tier2 Center) BNLU. Oklahoma Norman BNLU. of Texas Arlington BNLTier3 AggregateBNL (LIMAN)Internet2 / NLR Peerings 5Gbps20Gbps BNLTRIUMF (Canadian ATLAS Tier1) BNL (LIMAN)Seattle1Gbps5Gbps

8 LHC CMS Bandwidth Matrix as of April 2007 Site ASite ZESnet AESnet ZA-Z 2007 Bandwidth A-Z 2010 Bandwidth CERNFNALStarlight (CHIMAN) FNAL (CHIMAN) 10Gbps20-40Gbps FNALU. of Michigan (Calibration) FNAL (CHIMAN) Starlight (CHIMAN) 3Gbps10Gbps FNALCaltechFNAL (CHIMAN) Starlight (CHIMAN) 3Gbps10Gbps FNALMITFNAL (CHIMAN) AofA (NYC)/ Boston 3Gbps10Gbps FNALPurdue UniversityFNAL (CHIMAN) Starlight (CHIMAN) 3Gbps10Gbps FNALU. of California at San Diego FNAL (CHIMAN) San Diego3Gbps10Gbps FNALU. of Florida at Gainesville FNAL (CHIMAN) SOX3Gbps10Gbps FNALU. of Nebraska at Lincoln FNAL (CHIMAN) Starlight (CHIMAN) 3Gbps10Gbps FNALU. of Wisconsin at Madison FNAL (CHIMAN) Starlight (CHIMAN) 3Gbps10Gbps FNALTier3 AggregateFNAL (CHIMAN) Internet2 / NLR Peerings 5Gbps20Gbps

9  Aggregation of Requirements from All Case Studies Analysis of diverse programs and facilities yields dramatic convergence on a well-defined set of requirements – Reliability Fusion – 1 minute of slack during an experiment (99.999%) LHC – Small number of hours (99.95+%) SNS – limited instrument time makes outages unacceptable Drives requirement for redundancy, both in site connectivity and within ESnet – Connectivity Geographic reach equivalent to that of scientific collaboration Multiple peerings to add reliability and bandwidth to interdomain connectivity Critical both within the US and internationally – Bandwidth 10 Gbps site to site connectivity today 100 Gbps backbone by 2010 Multiple 10 Gbps R&E peerings Ability to easily deploy additional 10 Gbps lambdas and peerings Per-lambda bandwidth of 40 Gbps or 100 Gbps should be available by 2010 – Bandwidth and service guarantees All R&E networks must interoperate as one seamless fabric to enable end2end service deployment Flexible rate bandwidth guarantees – Collaboration support (federated trust, PKI, AV conferencing, etc.)

10 ESnet Traffic has Increased by 10X Every 47 Months, on Average, Since 1990 Terabytes / month Log Plot of ESnet Monthly Accepted Traffic, January, 1990 – June, 2006 Oct., TBy/mo. Aug., MBy/mo. Jul., TBy/mo. 38 months 57 months 40 months Nov., TBy/mo. Apr., PBy/mo. 53 months

11 Requirements from Network Utilization Observation In 4 years, we can expect a 10x increase in traffic over current levels without the addition of production LHC traffic – Nominal average load on busiest backbone links is greater than 1 Gbps today – In 4 years that figure will be over 10 Gbps if current trends continue Measurements of this kind are science-agnostic – It doesn’t matter who the users are, the traffic load is increasing exponentially Bandwidth trends drive requirement for a new network architecture – New ESnet4 architecture designed with these drivers in mind

12 Requirements from Traffic Flow Observations Most ESnet science traffic has a source or sink outside of ESnet – Drives requirement for high-bandwidth peering – Reliability and bandwidth requirements demand that peering be redundant – Multiple 10 Gbps peerings today, must be able to add more flexibly and cost-effectively Bandwidth and service guarantees must traverse R&E peerings – “Seamless fabric” – Collaboration with other R&E networks on a common framework is critical Large-scale science is becoming the dominant user of the network – Satisfying the demands of large-scale science traffic into the future will require a purpose-built, scalable architecture – Traffic patterns are different than commodity Internet Since large-scale science will be the dominant user going forward, the network should be architected to serve large-scale science

13 Aggregation of Requirements from Network Observation Traffic load continues to increase exponentially – 15-year trend indicates an increase of 10x in next 4 years – This means backbone traffic load will exceed 10 Gbps within 4 years requiring increased backbone bandwidth – Need new architecture – ESnet4 Large science flows typically cross network administrative boundaries, and are beginning to dominate – Requirements such as bandwidth capacity, reliability, etc. apply to peerings as well as ESnet itself – Large-scale science is becoming the dominant network user

14  Required Network Services Suite for DOE Science We have collected requirements from diverse science programs, program offices, and network analysis – the following summarizes the requirements: – Reliability 99.95% to % reliability Redundancy is the only way to meet the reliability requirements –Redundancy within ESnet –Redundant peerings –Redundant site connections where needed – Connectivity Geographic reach equivalent to that of scientific collaboration Multiple peerings to add reliability and bandwidth to interdomain connectivity Critical both within the US and internationally – Bandwidth 10 Gbps site to site connectivity today 100 Gbps backbone by 2010 Multiple 10+ Gbps R&E peerings Ability to easily deploy additional lambdas and peerings – Service guarantees All R&E networks must interoperate as one seamless fabric to enable end2end service deployment Guaranteed bandwidth, traffic isolation, quality of service Flexible rate bandwidth guarantees – Collaboration support Federated trust, PKI (Grid, middleware) Audio and Video conferencing – Production ISP service