ESnet Status Update ESCC July 18, 2007

Slides:



Advertisements
Similar presentations
1 Chin Guok ESnet Network Engineer David Robertson DSD Computer Software Engineer Lawrence Berkeley National Laboratory.
Advertisements

1 Chin Guok ESnet Network Engineer David Robertson DSD Computer Software Engineer Lawrence Berkeley National Laboratory.
ESnet On-demand Secure Circuits and Advance Reservation System (OSCARS) Chin Guok Network Engineering Group Thomas Ndousse Visit February Energy.
1 ESnet Update Summer 2007 Joint Techs Workshop Joe Burrescia ESnet General Manager July 16,2007 Energy Sciences Network Lawrence Berkeley National Laboratory.
ATLAS Tier 2 Paths Within ESnet Mike O’Connor ESnet Network Engineering Group Lawrence Berkeley National Lab
1 ESnet - Connecting the USA DOE Labs to the World of Science Eli Dart, Network Engineer Network Engineering Group Chinese American Network Symposium Indianapolis,
1 ESnet Network Measurements ESCC Feb Joe Metzger
Circuit Services - IPTV Christian Todorov Internet2 Fall Member Meeting October 9, 2007.
1 ESnet Planning for the LHC T0-T1 Networking William E. Johnston ESnet Manager and Senior Scientist Lawrence Berkeley National Laboratory.
1 ESnet On-demand Secure Circuits and Advance Reservation System (OSCARS) Chin Guok Network Engineering Group ESCC July Energy Sciences Network.
OSCARS Overview Path Computation Topology Reachability Contraints Scheduling AAA Availability Provisioning Signalling Security Resiliency/Redundancy OSCARS.
1 Services to the US Tier-1 Sites LHCOPN April 4th, 2006 Joe Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Jerry Sobieski MAX GigaPoP and TSC Program Manager Mark Johnson MCNC.
Nlr.net © 2004 National LambdaRail, Inc 1 NLR Update Dave Reese Joint Techs February 2007.
Internet2 Performance Update Jeff W. Boote Senior Network Software Engineer Internet2.
1 ESnet Update Winter 2008 Joint Techs Workshop Joe Burrescia ESnet General Manager January 21, 2008 Energy Sciences Network Lawrence Berkeley National.
1 ESnet Update Joint Techs Meeting Minneapolis, MN Joe Burrescia ESnet General Manager 2/12/2007.
1 Nuclear Physics Network Requirements Workshop Washington, DC Eli Dart, Network Engineer ESnet Network Engineering Group May 6, 2008 Energy Sciences Network.
1 ESnet Planning for the LHC T0-T1 Networking William E. Johnston ESnet Manager and Senior Scientist Lawrence Berkeley National Laboratory.
A Framework for Internetworking Heterogeneous High-Performance Networks via GMPLS and Web Services Xi Yang, Tom Lehman Information Sciences Institute (ISI)
Delivering Circuit Services to Researchers: The HOPI Testbed Rick Summerhill Director, Network Research, Architecture, and Technologies, Internet2 Joint.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
Copyright 2004 National LambdaRail, Inc N ational L ambda R ail Update 9/28/2004 Debbie Montano Director, Development & Operations
1 Network Measurement Summary ESCC, Feb Joe Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
Connect. Communicate. Collaborate perfSONAR MDM Service for LHC OPN Loukik Kudarimoti DANTE.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Advanced Networks: The Past and the Future – The Internet2 Perspective APAN 7 July 2004, Cairns, Australia Douglas Van Houweling, President & CEO Internet2.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
1 ESnet Update ESnet/Internet2 Joint Techs Albuquerque, New Mexico February 6, 2005 R. Kevin Oberman ESnet Network Engineer Lawrence Berkeley National.
INDIANAUNIVERSITYINDIANAUNIVERSITY HOPI: Hybrid Packet and Optical Infrastructure Chris Robb and Jim Williams Indiana University 7 July 2004 Cairns, AU.
1 ESnet4 IP Network and Science Data Network Configuration and Roll Out Schedule Projected Schedule as of Sept., 2006 For more information contact William.
1 Revision to DOE proposal Resource Optimization in Hybrid Core Networks with 100G Links Original submission: April 30, 2009 Date: May 4, 2009 PI: Malathi.
-1- ESnet On-Demand Secure Circuits and Advance Reservation System (OSCARS) David Robertson Internet2 Joint Techs Workshop July 18,
DICE Diagnostic Service Joe Metzger Joint Techs Measurement Working Group January
Supporting Advanced Scientific Computing Research Basic Energy Sciences Biological and Environmental Research Fusion Energy Sciences High Energy Physics.
The Internet2 Network and LHC Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 Given by Rich Carlson LHC Meeting 25.
Connecting to the new Internet2 Network What to Expect… Steve Cotter Rick Summerhill FMM 2006 / Chicago.
1 Deploying Measurement Systems in ESnet Joint Techs, Feb Joseph Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
1 ESnet4 IP Network and Science Data Network Configuration and Roll Out Schedule Projected Schedule as of Sept., 2006 For more information contact William.
1 Network Measurement Challenges LHC E2E Network Research Meeting October 25 th 2006 Joe Metzger Version 1.1.
J. Bunn, D. Nae, H. Newman, S. Ravot, X. Su, Y. Xia California Institute of Technology US LHCNet LHCNet WG September 12 th 2006.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 Joint Techs 17 July 2006 University of Wisconsin, Madison,
July 19, 2005-LHC GDB T0/T1 Networking L. Pinsky--ALICE-USA1 ALICE-USA T0/T1 Networking Plans Larry Pinsky—University of Houston For ALICE-USA.
The Internet2 Network and LHC Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 LHC Meeting 23 October 2006 FERMI Lab,
Community Design Workshop - First Day Summary, Issues Rick Summerhill, Director Network Research, Architecture, and Technologies Internet2 Community Design.
1 ESnet Network Requirements ASCAC Networking Sub-committee Meeting April 13, 2007 Eli Dart ESnet Engineering Group Lawrence Berkeley National Laboratory.
Computing at Fermilab D. Petravick Fermilab August 16, 2007.
Copyright 2007 CENIC and PNWGP
T0-T1 Networking Meeting 16th June Meeting
Welcome Network Virtualization & Hybridization Thomas Ndousse
Networking for the Future of Science
Ian Bird GDB Meeting CERN 9 September 2003
Dynamic Network Services In Internet2
Copyright 2007 CENIC and PNWGP
DOE Facilities - Drivers for Science: Experimental and Simulation Data
ESnet Network Engineer Lawrence Berkeley National Laboratory
Networking for the Future of Science
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
Networking for the Future of Large-Scale Science: An ESnet Perspective
Internet2 Performance Update
ESnet Network Measurements ESCC Feb Joe Metzger
Dynamic Circuit Service Hands-On GMPLS Engineering Workshop
ATLAS Tier 2 Paths Within ESnet
ESnet4: Networking for the Future of DOE Science
LHC Tier 2 Networking BOF
Fall 2006 Internet2 Member Meeting
Wide-Area Networking at SLAC
Optical communications & networking - an Overview
OSCARS Roadmap Chin Guok
Abilene Update Rick Summerhill
Presentation transcript:

ESnet Status Update ESCC July 18, 2007 William E. Johnston ESnet Department Head and Senior Scientist Energy Sciences Network Lawrence Berkeley National Laboratory wej@es.net, www.es.net This talk is available at www.es.net/ESnet4 Networking for the Future of Science

DOE Office of Science and ESnet – the ESnet Mission ESnet’s primary mission is to enable the large-scale science that is the mission of the Office of Science (SC) and that depends on: Sharing of massive amounts of data Supporting thousands of collaborators world-wide Distributed data processing Distributed data management Distributed simulation, visualization, and computational steering Collaboration with the US and International Research and Education community ESnet provides network and collaboration services to Office of Science laboratories and many other DOE programs in order to accomplish its mission The point here is that most modern, large-scale, science and engineering is multi-disciplinary, and must use geographically distributed components, and this is what has motivated a $50-75M/yr (approx) investment in Grid technology by the US, UK, European, Japanese, and Korean governments.

I. Current Network Status Talk Outline I. Current Network Status II. Planning and Building the Future Network - ESnet4 III. Science Collaboration Services - 1. Federated Trust IV. Science Collaboration Services - 2. Audio, Video, Data Teleconferencing

ESnet Science Data Network (SDN) core ESnet3 Today Provides Global High-Speed Internet Connectivity for DOE Facilities and Collaborators (Early 2007) ESnet Science Data Network (SDN) core Japan (SINet) Australia (AARNet) Canada (CA*net4 Taiwan (TANet2) Singaren KAREN/REANNZ ODN Japan Telecom America NLR-Packetnet Abilene/I2 SINet (Japan) Russia (BINP) GÉANT - France, Germany, Italy, UK, etc CA*net4 France GLORIAD (Russia, China) Korea (Kreonet2 MREN StarTap Taiwan (TANet2, ASCC) PacificWave CERN (USLHCnet DOE+CERN funded) SEA NSF/IRNC funded LIGO MAN LAN Abilene PNNL KAREN/REANNZ ODN Japan Telecom America Abilene/I2 SINGAREN AU Starlight CHI-SL Abilene NYC PacWave MIT Abilene INEEL ESnet IP core: Packet over SONET Optical Ring and Hubs Lab DC Offices BNL JGI TWC FNAL LLNL LBNL AMES ANL PPPL SNLL NERSC SNV CHI MAE-E SLAC Equinix PAIX-PA Equinix, etc. OSC GTN NNSA Equinix SNV SDN BECHTEL-NV NASA Ames IARC KCP NREL DC JLAB ORNL YUCCA MT OSTI ORAU LANL MAXGPoP ARM SDSC SNLA NOAA AU GA SRS ALB Allied Signal PANTEX UNM ATL AMPATH (S. America) 42 end user sites AMPATH (S. America) DOE-ALB Office Of Science Sponsored (22) ELP International (high speed) 10 Gb/s SDN core 10G/s IP core 2.5 Gb/s IP core MAN rings (≥ 10 G/s) Lab supplied links OC12 ATM (622 Mb/s) OC12 / GigEthernet OC3 (155 Mb/s) 45 Mb/s and less NNSA Sponsored (13) Joint Sponsored (3) Other Sponsored (NSF LIGO, NOAA) Laboratory Sponsored (6) commercial peering points R&E networks Specific R&E network peers Other R&E peering points IP SNV ESnet core hubs Abilene high-speed peering points with Internet2/Abilene

ESnet Availability “4 nines” (>99.95%) “3 nines” (>99.5%) With a goal of “5 nines” for the large science Labs it becomes clear that ESnet will have to deploy dual routers at the site and core-core attachment points in order to avoid down time due to router reloads/upgrades. “5 nines” (>99.995%) “4 nines” (>99.95%) “3 nines” (>99.5%) Dually connected sites Note: These availability measures are only for ESnet infrastructure, they do not include site-related problems. Some sites, e.g. PNNL and LANL, provide circuits from the site to an ESnet hub, and therefore the ESnet-site demarc is at the ESnet hub (there is no ESnet equipment at the site. In this case, circuit outages between the ESnet equipment and the site are considered site issues and are not included in the ESnet availability metric.

Peering Issues ESnet has experienced congestion at both the West Coast and mid-West Equinix commercial peering exchanges

Commercial Peers Congestion Issues: Temporary changes Long-term Fixes The OC3 connection between paix-pa-rt1 and snv-rt1 was very congested, with peaks clipped for most of the day. Temporary mitigation Temporarily forcing West coast Level3 traffic to eqx-chicago - Traffic is now only clipped (if at all) at the peak of the day Long term solution Establish new Level3 peering at eqx-chicago (7/11/07) Working on establishing a second peering with Global Crossing Upgrade current loop (OC3) and fabric (100Mpbs) to 1Gbps Congestion to AT&T Upgraded AT&T peering at eqx-sanjose from OC3 to OC12 (3/15/07) Established OC12 peering with AT&T at eqx-ashburn (1/29/07) and eqx-chicago (07/11/07) The Equinix shared fabric at eqx-ashburn is congested New Level3 peering at eqx-chicago has helped to relieve congestion Additional mitigation Third peering with Google at eqx-chicago, third peering with Yahoo at eqx-chicago Future mitigation Establish a second peering with Global Crossing at eqx-chicago Upgrade equinix-sanjose and equinix-ashburn fabrics connections from 100Mb/s to 1Gbps

Planning and Building the Future Network - ESnet4 II. Planning and Building the Future Network - ESnet4 Requirements are primary drivers for ESnet – science focused Sources of Requirements Office of Science (SC) Program Managers The Program Offices Requirements Workshops BES completed BER in July, 2007 Others to follow at the rate of 3 a year Direct gathering through interaction with science users of the network Example case studies (updated 2005/2006) Magnetic Fusion Large Hadron Collider (LHC) Climate Modeling Spallation Neutron Source Observation of the network Requirements aggregation Convergence on a complete set of network requirements

1. Basic Energy Sciences (BES) Network Requirements Workshop Input from BES facilities, science programs and sites Light Sources SNS at ORNL, Neutron Science program Nanoscience Centers Combustion Research Computational Chemistry Other existing facilities (e.g. National Center for Electron Microscopy at LBL) Facilities currently undergoing construction (e.g. LCLS at SLAC)

Additional discussion – ESnet4 Workshop Process Three inputs Discussion of Program Office – goals, future projects, and science portfolio Discussions with representatives of individual programs and facilities Group discussion about common issues, future technologies (e.g. detector upgrades), etc. Additional discussion – ESnet4 Architecture Deployment schedule Future services

BES Workshop Findings (1) BES facilities are unlikely to provide the magnitude of load that we expect from the LHC However, significant detector upgrades are coming in the next 3 years LCLS may provide significant load SNS data repositories may provide significant load Theory and simulation efforts may provide significant load Broad user base Makes it difficult to model facilities as anything other than point sources of traffic load Requires wide connectivity Most facilities and disciplines expect significant increases in PKI service usage

BES Workshop Findings (2) Significant difficulty and frustration with moving data sets Problems deal with moving data sets that are small by HEP’s standards Currently many users ship hard disks or stacks of DVDs Solutions HEP model of assigning a group of skilled computer people to address the data transfer problem does not map well onto BES for several reasons BES is more heterogeneous in science and in funding User base for BES facilities is very heterogeneous and this results in a large number of sites that must be involved in data transfers It appears that this is likely to be true of the other Program Offices ESnet action item – build a central web page for disseminating information about data transfer tools and techniques Users also expressed interest in a blueprint for a site-local BWCTL/PerfSONAR service (1A)

2. Case Studies For Requirements Advanced Scientific Computing Research (ASCR) NERSC NLCF Basic Energy Sciences Advanced Light Source Macromolecular Crystallography Chemistry/Combustion Spallation Neutron Source Biological and Environmental Bioinformatics/Genomics Climate Science Fusion Energy Sciences Magnetic Fusion Energy/ITER High Energy Physics LHC Nuclear Physics RHIC

(2A) Science Networking Requirements Aggregation Summary Science Drivers Science Areas / Facilities End2End Reliability Connectivity Today End2End Band width 5 years End2End Band width Traffic Characteristics Network Services Magnetic Fusion Energy 99.999% (Impossible without full redundancy) DOE sites US Universities Industry 200+ Mbps 1 Gbps Bulk data Remote control Guaranteed bandwidth Guaranteed QoS Deadline scheduling NERSC and ACLF - International Other ASCR supercomputers 10 Gbps 20 to 40 Gbps Remote file system sharing Deadline Scheduling PKI / Grid NLCF Backbone Band width parity Backbone band width parity Nuclear Physics (RHIC) 12 Gbps 70 Gbps Spallation Neutron Source High (24x7 operation) 640 Mbps 2 Gbps

Science Network Requirements Aggregation Summary Science Drivers Science Areas / Facilities End2End Reliability Connectivity Today End2End Band width 5 years End2End Band width Traffic Characteristics Network Services Advanced Light Source - DOE sites US Universities Industry 1 TB/day 300 Mbps 5 TB/day 1.5 Gbps Bulk data Remote control Guaranteed bandwidth PKI / Grid Bioinformatics 625 Mbps 12.5 Gbps in two years 250 Gbps Point-to-multipoint High-speed multicast Chemistry / Combustion 10s of Gigabits per second Climate Science International 5 PB per year 5 Gbps High Energy Physics (LHC) 99.95+% (Less than 4 hrs/year) US Tier1 (FNAL, BNL) US Tier2 (Universities) International (Europe, Canada) 10 Gbps 60 to 80 Gbps (30-40 Gbps per US Tier1) Coupled data analysis processes Traffic isolation Immediate Requirements and Drivers

TRIUMF (Atlas T1, Canada) (2B) The Next Level of Detail: LHC Tier 0, 1, and 2 Connectivity Requirements Summary TRIUMF (Atlas T1, Canada) Vancouver CANARIE CERN-1 USLHCNet Seattle Toronto BNL (Atlas T1) CERN-2 ESnet SDN Virtual Circuits Boise Chicago CERN-3 New York Denver Sunnyvale KC GÉANT-1 ESnet IP Core FNAL (CMS T1) Wash DC Internet2 / RONs Internet2 / RONs Internet2 / RONs LA Albuq. GÉANT-2 San Diego GÉANT Dallas Atlanta Jacksonville ESnet IP core hubs ESnet SDN/NLR hubs Cross connects ESnet - Internet2 Internet2/GigaPoP nodes USLHC nodes Direct connectivity T0-T1-T2 USLHCNet to ESnet to Abilene Backup connectivity SDN, GLIF, VCs Tier 1 Centers Tier 2 Sites

(2C) The Next Level of Detail: LHC ATLAS Bandwidth Matrix as of April 2007 Site A Site Z ESnet A ESnet Z A-Z 2007 Bandwidth A-Z 2010 Bandwidth CERN BNL AofA (NYC) 10Gbps 20-40Gbps U. of Michigan (Calibration) BNL (LIMAN) Starlight (CHIMAN) 3Gbps Boston University Internet2 / NLR Peerings (Northeastern Tier2 Center) Harvard University Indiana U. at Bloomington (Midwestern Tier2 Center) U. of Chicago Langston University (Southwestern Tier2 Center) U. Oklahoma Norman U. of Texas Arlington Tier3 Aggregate 5Gbps 20Gbps TRIUMF (Canadian ATLAS Tier1) Seattle 1Gbps

LHC CMS Bandwidth Matrix as of April 2007 Site A Site Z ESnet A ESnet Z A-Z 2007 Bandwidth A-Z 2010 Bandwidth CERN FNAL Starlight (CHIMAN) FNAL (CHIMAN) 10Gbps 20-40Gbps U. of Michigan (Calibration) 3Gbps Caltech MIT AofA (NYC)/ Boston Purdue University U. of California at San Diego San Diego U. of Florida at Gainesville SOX / Ultralight at Starlight U. of Nebraska at Lincoln U. of Wisconsin at Madison Tier3 Aggregate Internet2 / NLR Peerings 5Gbps 20Gbps

Large-Scale Data Analysis Systems (Typified by the LHC) have Several Characteristics that Result in Requirements for the Network and its Services The systems are data intensive and high-performance, typically moving terabytes a day for months at a time The system are high duty-cycle, operating most of the day for months at a time in order to meet the requirements for data movement The systems are widely distributed – typically spread over continental or inter-continental distances Such systems depend on network performance and availability, but these characteristics cannot be taken for granted, even in well run networks, when the multi-domain network path is considered The applications must be able to get guarantees from the network that there is adequate bandwidth to accomplish the task at hand The applications must be able to get information from the network that allows graceful failure and auto-recovery and adaptation to unexpected network conditions that are short of outright failure This slide drawn from [ICFA SCIC]

Enabling Large-Scale Science These requirements are generally true for systems with widely distributed components to be reliable and consistent in performing the sustained, complex tasks of large-scale science Networks must provide communication capability that is service-oriented: configurable, schedulable, predictable, reliable, and informative – and the network and its services must be scalable (2D)

3. Observed Evolution of Historical ESnet Traffic Patterns ESnet total traffic passed 2 Petabytes/mo about mid-April, 2007 Terabytes / month top 100 sites to site workflows site to site workflow data not available ESnet Monthly Accepted Traffic, January, 2000 – June, 2007 ESnet is currently transporting more than1 petabyte (1000 terabytes) per month More than 50% of the traffic is now generated by the top 100 sites  large-scale science dominates all ESnet traffic

Log Plot of ESnet Monthly Accepted Traffic, January, 1990 – June, 2007 ESnet Traffic has Increased by 10X Every 47 Months, on Average, Since 1990 Apr., 2006 1 PBy/mo. Nov., 2001 100 TBy/mo. Jul., 1998 10 TBy/mo. 53 months Oct., 1993 1 TBy/mo. 40 months Terabytes / month Aug., 1990 100 MBy/mo. 57 months 38 months Log Plot of ESnet Monthly Accepted Traffic, January, 1990 – June, 2007

Requirements from Network Utilization Observation In 4 years, we can expect a 10x increase in traffic over current levels without the addition of production LHC traffic Nominal average load on busiest backbone links is ~1.5 Gbps today In 4 years that figure will be ~15 Gbps based on current trends Measurements of this type are science-agnostic It doesn’t matter who the users are, the traffic load is increasing exponentially Predictions based on this sort of forward projection tend to be conservative estimates of future requirements because they cannot predict new uses (3A)

Requirements from Traffic Flow Observations Most of ESnet science traffic has a source or sink outside of ESnet Drives requirement for high-bandwidth peering Reliability and bandwidth requirements demand that peering be redundant Multiple 10 Gbps peerings today, must be able to add more bandwidth flexibly and cost-effectively Bandwidth and service guarantees must traverse R&E peerings Collaboration with other R&E networks on a common framework is critical Seamless fabric Large-scale science is now the dominant user of the network Satisfying the demands of large-scale science traffic into the future will require a purpose-built, scalable architecture Traffic patterns are different than commodity Internet (3B) (3C)

Summary of All Requirements To-Date Requirements from SC Programs: 1A) Provide “consulting” on system / application network tuning Requirements from science case studies: 2A) Build the ESnet core up to 100 Gb/s within 5 years 2B) Deploy network to accommodate LHC collaborator footprint 2C) Implement network to provide for LHC data path loadings 2D) Provide the network as a service-oriented capability Requirements from observing traffic growth and change trends in the network: 3A) Provide 15 Gb/s core within four years and 150 Gb/s core within eight years 3B) Provide a rich diversity and high bandwidth for R&E peerings 3C) Economically accommodate a very large volume of circuit-like traffic

ESnet4 - The Response to the Requirements I) A new network architecture and implementation strategy Provide two networks: IP and circuit-oriented Science Data Netework Reduces cost of handling high bandwidth data flows Highly capable routers are not necessary when every packet goes to the same place Use lower cost (factor of 5x) switches to relatively route the packets Rich and diverse network topology for flexible management and high reliability Dual connectivity at every level for all large-scale science sources and sinks A partnership with the US research and education community to build a shared, large-scale, R&E managed optical infrastructure a scalable approach to adding bandwidth to the network dynamic allocation and management of optical circuits II) Development and deployment of a virtual circuit service Develop the service cooperatively with the networks that are intermediate between DOE Labs and major collaborators to ensure and-to-end interoperability III) Develop and deploy service-oriented, user accessable network monitoring systems IV) Provide “consulting” on system / application network performance tuning

ESnet4 Internet2 has partnered with Level 3 Communications Co. and Infinera Corp. for a dedicated optical fiber infrastructure with a national footprint and a rich topology - the “Internet2 Network” The fiber will be provisioned with Infinera Dense Wave Division Multiplexing equipment that uses an advanced, integrated optical-electrical design Level 3 will maintain the fiber and the DWDM equipment The DWDM equipment will initially be provisioned to provide10 optical circuits (lambdas - s) across the entire fiber footprint (80 s is max.) ESnet has partnered with Internet2 to: Share the optical infrastructure Develop new circuit-oriented network services Explore mechanisms that could be used for the ESnet Network Operations Center (NOC) and the Internet2/Indiana University NOC to back each other up for disaster recovery purposes

ESnet4 ESnet will build its next generation IP network and its new circuit-oriented Science Data Network primarily on the Internet2 circuits (s) that are dedicated to ESnet, together with a few National Lambda Rail and other circuits ESnet will provision and operate its own routing and switching hardware that is installed in various commercial telecom hubs around the country, as it has done for the past 20 years ESnet’s peering relationships with the commercial Internet, various US research and education networks, and numerous international networks will continue and evolve as they have for the past 20 years

Internet2 and ESnet Optical Node M320 IP core T640 RON ESnet metro-area networks SDN core switch grooming device Ciena CoreDirector dynamically allocated and routed waves (future) support devices: measurement out-of-band access monitoring security Direct Optical Connections to RONs support devices: measurement out-of-band access monitoring ……. Network Testbeds various equipment and experimental control plane management systems Future access to control plane fiber east fiber west Infinera DTN Internet2/Level3 National Optical Infrastructure fiber north/south

ESnet production IP core hub ESnet Metropolitan Area Network Ring Architecture for High Reliability Sites SDN core east US LHCnet switch US LHCnet switch ESnet production IP core hub IP core east IP core west SDN core switch IP core router SDN core west SDN core switch T320 ESnet IP core hub ESnet SDN core hub MAN fiber ring: 2-4 x 10 Gbps channels provisioned initially, with expansion capacity to 16-64 ESnet managed virtual circuit services tunneled through the IP backbone Large Science Site ESnet managed λ / circuit services ESnet production IP service ESnet MAN switch Independent port cards supporting multiple 10 Gb/s line interfaces MAN site switch Site ESnet switch MAN site switch Virtual Circuits to Site Virtual Circuit to Site Site router Site gateway router SDN circuits to site systems Site LAN Site edge router

ESnet 3 Backbone as of January 1, 2007 Seattle New York City Sunnyvale Chicago Washington DC San Diego Albuquerque Atlanta El Paso Future ESnet Hub ESnet Hub 10 Gb/s SDN core (NLR) 10/2.5 Gb/s IP core (QWEST) MAN rings (≥ 10 G/s) Lab supplied links

ESnet 4 Backbone as of April 15, 2007 Seattle Boston Clev. New York City Sunnyvale Chicago Washington DC San Diego Albuquerque Atlanta El Paso 10 Gb/s SDN core (NLR) 10/2.5 Gb/s IP core (QWEST) 10 Gb/s IP core (Level3) 10 Gb/s SDN core (Level3) MAN rings (≥ 10 G/s) Lab supplied links Future ESnet Hub ESnet Hub

ESnet 4 Backbone as of May 15, 2007 Seattle Boston Boston Clev. Clev. New York City Sunnyvale Chicago Washington DC SNV San Diego Albuquerque Atlanta El Paso 10 Gb/s SDN core (NLR) 10/2.5 Gb/s IP core (QWEST) 10 Gb/s IP core (Level3) 10 Gb/s SDN core (Level3) MAN rings (≥ 10 G/s) Lab supplied links Future ESnet Hub ESnet Hub

ESnet 4 Backbone as of June 20, 2007 Seattle Boston Boston Clev. New York City Sunnyvale Denver Chicago Washington DC Kansas City San Diego Albuquerque Atlanta El Paso 10 Gb/s SDN core (NLR) 10/2.5 Gb/s IP core (QWEST) 10 Gb/s IP core (Level3) 10 Gb/s SDN core (Level3) MAN rings (≥ 10 G/s) Lab supplied links Houston Future ESnet Hub ESnet Hub

ESnet 4 Backbone Target August 1, 2007 Seattle Denver-Sunnyvale-El Paso ring installed July 16, 2007 Boston Boston Clev. Clev. New York City Sunnyvale Denver Chicago Washington DC Kansas City Los Angeles San Diego Albuquerque Atlanta El Paso 10 Gb/s SDN core (NLR) 10/2.5 Gb/s IP core (QWEST) 10 Gb/s IP core (Level3) 10 Gb/s SDN core (Level3) MAN rings (≥ 10 G/s) Lab supplied links Houston Houston Future ESnet Hub ESnet Hub

ESnet 4 Backbone Target August 30, 2007 Seattle Boston Boston Boise Clev. Clev. New York City Sunnyvale Denver Chicago Washington DC Kansas City Los Angeles San Diego Albuquerque Atlanta El Paso 10 Gb/s SDN core (NLR) 10/2.5 Gb/s IP core (QWEST) 10 Gb/s IP core (Level3) 10 Gb/s SDN core (Level3) MAN rings (≥ 10 G/s) Lab supplied links Houston Houston Future ESnet Hub ESnet Hub

ESnet 4 Backbone Target September 30, 2007 Seattle Boston Boston Boise Clev. Clev. New York City Sunnyvale Denver Chicago Washington DC Los Angeles Kansas City Nashville San Diego Albuquerque Atlanta El Paso 10 Gb/s SDN core (NLR) 10/2.5 Gb/s IP core (QWEST) 10 Gb/s IP core (Level3) 10 Gb/s SDN core (Level3) MAN rings (≥ 10 G/s) Lab supplied links Houston Houston Future ESnet Hub ESnet Hub

ESnet4 Roll Out ESnet4 IP + SDN Configuration, mid-September, 2007 All circuits are 10Gb/s, unless noted. Seattle (28) Portland (8) (29) Boise Boston (9) (7) Chicago Clev. (11) (10) NYC (32) Sunnyvale Denver (13) (25) Salt Lake City KC (12) Philadelphia (15) (14) (26) (16) Pitts. Wash DC Indianapolis (21) (27) (23) (0) (0) (22) (30) Raleigh LA Tulsa Nashville Albuq. OC48 (24) (1(3)) (3) (4) San Diego (1) (1) (19) (20) (2) Atlanta Jacksonville El Paso (17) Layer 1 optical nodes at eventual ESnet Points of Presence ESnet IP switch only hubs ESnet IP switch/router hubs ESnet SDN switch hubs Layer 1 optical nodes not currently in ESnet plans Lab site (6) (5) Baton Rouge Houston ESnet IP core ESnet Science Data Network core ESnet SDN core, NLR links Lab supplied link LHC related link MAN link International IP Connections

ESnet4 Metro Area Rings, 2007 Configurations West Chicago MAN Long Island MAN FNAL 600 W. Chicago Starlight ANL USLHCNet Ames BNL 32 AoA, NYC USLHCNet Seattle (28) Portland (8) (29) Boise Boston (9) (7) Chicago Clev. (11) (10) NYC (32) Denver (13) (25) Sunnyvale Salt Lake City KC (12) Philadelphia (15) (14) (26) (16) Pitts. San Francisco Bay Area MAN Wash DC Indianapolis (21) (27) (23) (0) (22) (30) Raleigh LA LBNL SLAC JGI LLNL SNLL NERSC Tulsa Nashville Albuq. OC48 (24) (3) (1(3)) (4) San Diego (1) Newport News - Elite Atlanta (20) (2) JLab ELITE ODU MATP Wash., DC (19) Jacksonville El Paso (17) Layer 1 optical nodes at eventual ESnet Points of Presence ESnet IP switch only hubs ESnet IP switch/router hubs ESnet SDN switch hubs Layer 1 optical nodes not currently in ESnet plans Lab site (6) Atlanta MAN 180 Peachtree 56 Marietta (SOX) ORNL Wash., DC Houston Nashville ESnet IP core ESnet Science Data Network core ESnet SDN core, NLR links (existing) Lab supplied link LHC related link MAN link International IP Connections All circuits are 10Gb/s.

San Francisco Bay Area MAN Note that the major ESnet sites are now directly on the ESnet “core” network West Chicago MAN Long Island MAN FNAL 600 W. Chicago Starlight ANL USLHCNet e.g. the bandwidth into and out of FNAL is equal to, or greater, than the ESnet core bandwidth BNL 32 AoA, NYC USLHCNet Seattle (28) Portland (8) (>1 ) (29) 5 Boise Boston (9) 5 Chicago 4 (7) Clev. (11) 5 (10) NYC 5 Pitts. (32) Denver (13) Sunnyvale 5 (25) KC (12) Salt Lake City (14) Philadelphia (15) 5 5 (26) (16) 4 (21) Wash. DC 4 San Francisco Bay Area MAN LBNL SLAC JGI LLNL SNLL NERSC 5 Indianapolis (27) 4 (23) (0) (22) 3 (30) Raleigh LA 4 Tulsa 5 Nashville Albuq. OC48 (24) 4 4 (3) 3 (4) San Diego (1) 3 Atlanta (19) (20) (2) JLab ELITE ODU MATP Wash., DC 4 Jacksonville Layer 1 optical nodes at eventual ESnet Points of Presence ESnet IP switch only hubs ESnet IP switch/router hubs ESnet SDN switch hubs Layer 1 optical nodes not currently in ESnet plans Lab site El Paso (17) Atlanta MAN 180 Peachtree 56 Marietta (SOX) ORNL Wash., DC Houston Nashville 4 ESnet IP core (1) ESnet Science Data Network core ESnet SDN core, NLR links (existing) Lab supplied link LHC related link MAN link International IP Connections Internet2 circuit number (20) (6) (5) Baton Rouge Houston

The Evolution of ESnet Architecture ESnet IP core ESnet IP core independent redundancy (TBD) ESnet Science Data Network (SDN) core ESnet to 2005: A routed IP network with sites singly attached to a national core ring ESnet from 2006-07: A routed IP network with sites dually connected on metro area rings or dually connected directly to core ring A switched network providing virtual circuit services for data-intensive science Rich topology offsets the lack of dual, independent national cores ESnet sites ESnet hubs / core network connection points Metro area rings (MANs) Other IP networks Circuit connections to other science networks (e.g. USLHCNet)

ESnet 4 Factiods as of July 16, 2007 Installation to date: 10 new 10Gb/s circuits ~10,000 Route Miles 6 new hubs 5 new routers 4 new switches Total of 70 individual pieces of equipment shipped Over two and a half tons of electronics 15 round trip airline tickets for our install team About 60,000 miles traveled so far…. 6 cities 5 Brazilian Bar-B-Qs/Grills sampled

Typical ESnet 4 Hub OWAMP Time Source Power Controllers Secure Term Server Peering Router 10G Performance Tester M320 Router 7609 Switch

(2C) Aggregate Estimated Link Loadings, 2007-08 9 12.5 Seattle 13 13 9 (28) Portland (8) Existing site supplied circuits 2.5 (29) Boise Boston (9) (7) Chicago Clev. (11) (10) NYC (32) (13) (25) Sunnyvale Denver Salt Lake City KC (12) Philadelphia (15) (14) (26) (16) Pitts. Wash DC Indianapolis (21) (27) (23) (0) (0) (22) (30) 8.5 Raleigh LA Tulsa Nashville Albuq. OC48 6 (24) (1(3)) (3) (4) San Diego (1) (1) Atlanta 6 (19) (20) (2) Jacksonville El Paso (17) Layer 1 optical nodes at eventual ESnet Points of Presence ESnet IP switch only hubs ESnet IP switch/router hubs ESnet SDN switch hubs Layer 1 optical nodes not currently in ESnet plans Lab site (6) (5) Baton Rouge Houston ESnet IP core (1) ESnet Science Data Network core ESnet SDN core, NLR links Lab supplied link LHC related link MAN link International IP Connections 2.5 2.5 2.5 Committed bandwidth, Gb/s

(2C) ESnet4 2007-8 Estimated Bandwidth Commitments West Chicago MAN 600 W. Chicago Long Island MAN CMS BNL 32 AoA, NYC USLHCNet CERN 5 Seattle 10 (28) Portland CERN (8) Starlight 13 (29) Boise Boston 29 (total) (9) (7) Chicago Clev. (11) (10) NYC (32) Denver (13) (25) Sunnyvale USLHCNet 10 Salt Lake City KC (12) Philadelphia (15) (14) (26) (16) Pitts. San Francisco Bay Area MAN Wash DC Indianapolis (21) (27) (23) (0) (22) (30) Raleigh LA FNAL LBNL SLAC JGI LLNL SNLL NERSC ANL Tulsa Nashville Albuq. OC48 (24) (1(3)) (3) (4) San Diego (1) Newport News - Elite Atlanta (19) (20) (2) JLab ELITE ODU MATP Wash., DC Jacksonville El Paso (17) Layer 1 optical nodes at eventual ESnet Points of Presence ESnet IP switch only hubs ESnet IP switch/router hubs ESnet SDN switch hubs Layer 1 optical nodes not currently in ESnet plans Lab site (6) (5) Baton Rouge Houston MAX ESnet IP core ESnet Science Data Network core ESnet SDN core, NLR links (existing) Lab supplied link LHC related link MAN link International IP Connections All circuits are 10Gb/s. 2.5 Committed bandwidth, Gb/s

Are These Estimates Realistic? YES! FNAL Outbound CMS Traffic Max= 1064 MBy/s (8.5 Gb/s), Average = 394 MBy/s (3.2 Gb/s)

ESnet4 IP + SDN, 2008 Configuration Seattle (28) Portland (8) (? ) (29) (2) Boise Boston (2) (9) Chicago (1) (7) Clev. (2) (11) (2) NYC (10) (2) Pitts. (32) Sunnyvale Denver (13) (2) (25) (12) Philadelphia Salt Lake City KC (15) (2) (14) (2) (26) (16) (1) (21) (2) Wash. DC Indianapolis (27) (1) (23) (1) (2) (0) (22) (2) (30) Raleigh LA (1) Tulsa Nashville Albuq. OC48 (24) (1) (1) (2) (4) San Diego (1) (1) (3) Atlanta (19) (20) (2) (1) Jacksonville El Paso (1) (17) Layer 1 optical nodes at eventual ESnet Points of Presence ESnet IP switch only hubs ESnet IP switch/router hubs ESnet SDN switch hubs Layer 1 optical nodes not currently in ESnet plans Lab site (6) (5) Baton Rouge ESnet IP core ESnet Science Data Network core ESnet SDN core, NLR links (existing) Lab supplied link LHC related link MAN link International IP Connections Internet2 circuit number (20) Houston

Estimated ESnet4 2009 Configuration (Some of the circuits may be allocated dynamically from shared a pool.) Seattle (28) Portland (8) (? ) (29) 3 Boise Boston (9) 3 Chicago 2 (7) Clev. (11) 3 (10) 3 NYC 3 Pitts. (32) Sunnyvale Denver (13) 3 (25) (12) Philadelphia Salt Lake City KC (14) (15) 3 3 (26) (16) 2 (21) Wash. DC 2 3 Indianapolis (27) 2 (23) (0) (22) 2 (30) 3 Raleigh LA 2 Tulsa Nashville Albuq. OC48 (24) 2 2 2 (4) San Diego (3) (1) 1 Atlanta (19) (20) (2) 2 Jacksonville El Paso 2 (17) ESnet IP switch/router hubs (6) (5) Baton Rouge ESnet IP core ESnet Science Data Network core ESnet SDN core, NLR links (existing) Lab supplied link LHC related link MAN link International IP Connections Internet2 circuit number (20) ESnet IP switch only hubs Houston ESnet SDN switch hubs Layer 1 optical nodes at eventual ESnet Points of Presence Layer 1 optical nodes not currently in ESnet plans Lab site

(2C) Aggregate Estimated Link Loadings, 2010-11 30 45 Seattle 50 (28) 15 20 Portland (8) (>1 ) (29) 5 Boise Boston (9) 5 (7) Chicago 4 Clev. (11) 5 (10) NYC 5 Pitts. (32) Denver (13) (25) Sunnyvale 5 KC (12) Salt Lake City (14) Philadelphia (15) 5 5 (26) (16) 4 (21) Wash. DC 10 4 5 5 Indianapolis (27) 4 (23) (0) (22) 3 (30) 5 Raleigh LA 4 Tulsa 5 Nashville Albuq. 20 OC48 (24) 4 4 3 (4) San Diego (3) (1) 3 Atlanta 20 20 (19) (20) 5 (2) 4 Jacksonville El Paso 4 (17) ESnet IP switch/router hubs (6) 20 (5) Baton Rouge ESnet IP core (1) ESnet Science Data Network core ESnet SDN core, NLR links (existing) Lab supplied link LHC related link MAN link International IP Connections Internet2 circuit number (20) ESnet IP switch only hubs Houston ESnet SDN switch hubs Layer 1 optical nodes at eventual ESnet Points of Presence Layer 1 optical nodes not currently in ESnet plans Lab site

(2C) ESnet4 2010-11 Estimated Bandwidth Commitments 80 FNAL 600 W. Chicago Starlight ANL USLHCNet CERN 40 100 CMS 25 BNL 32 AoA, NYC USLHCNet CERN 65 40 20 Seattle 25 (28) 15 Portland (8) (>1 ) (29) 5 Boise Boston (9) 5 (7) Chicago 4 Clev. (11) 5 (10) NYC 5 Pitts. (32) Denver (13) Sunnyvale 5 (25) Salt Lake City KC (12) (14) Philadelphia (15) 5 5 (26) (16) 4 (21) Wash. DC 4 5 5 Indianapolis (27) 4 (23) (0) (22) 3 (30) 5 Raleigh LA 4 Tulsa 5 Nashville Albuq. 10 OC48 (24) 4 4 (3) 3 (4) San Diego (1) 3 Atlanta 20 20 (19) (20) 5 (2) 4 Jacksonville El Paso 4 (17) ESnet IP switch/router hubs (6) 10 (5) Baton Rouge ESnet IP core (1) ESnet Science Data Network core ESnet SDN core, NLR links (existing) Lab supplied link LHC related link MAN link International IP Connections Internet2 circuit number (20) ESnet IP switch only hubs Houston ESnet SDN switch hubs Layer 1 optical nodes at eventual ESnet Points of Presence Layer 1 optical nodes not currently in ESnet plans Lab site

ESnet4 IP + SDN, 2011 Configuration Seattle (28) Portland (8) (>1 ) (29) 5 Boise Boston 5 (9) (7) Chicago 4 Clev. (11) 5 (10) NYC 5 Pitts. (32) Denver (13) Sunnyvale 5 (25) Salt Lake City KC (12) (14) Philadelphia (15) 5 5 (26) (16) 4 (21) Wash. DC 4 5 Indianapolis (27) 4 (23) (22) 3 (0) (30) Raleigh LA 4 Tulsa 5 Nashville Albuq. OC48 (24) 4 4 3 (4) San Diego (3) (1) 3 Atlanta (19) (20) (2) 4 Jacksonville El Paso 4 Layer 1 optical nodes at eventual ESnet Points of Presence ESnet IP switch only hubs ESnet IP switch/router hubs ESnet SDN switch hubs Layer 1 optical nodes not currently in ESnet plans Lab site (17) (6) (5) Baton Rouge ESnet IP core (1) ESnet Science Data Network core ESnet SDN core, NLR links (existing) Lab supplied link LHC related link MAN link International IP Connections Internet2 circuit number (20) Houston

ESnet4 Planed Configuration Core networks: 40-50 Gbps in 2009-2010, 160-400 Gbps in 2011-2012 Canada (CANARIE) CERN (30 Gbps) Asia-Pacific Canada (CANARIE) Asia Pacific CERN (30 Gbps) GLORIAD (Russia and China) Europe (GEANT) 1625 miles / 2545 km Asia-Pacific Science Data Network Core Seattle Cleveland Boston IP Core Chicago Boise Australia New York Kansas City Denver Washington DC Sunnyvale Atlanta Tulsa LA Albuquerque Australia South America (AMPATH) San Diego IP core hubs Primary DOE Labs SDN (switch) hubs High speed cross-connects with Ineternet2/Abilene Possible hubs South America (AMPATH) Houston Jacksonville Production IP core (10Gbps) ◄ SDN core (20-30-40Gbps) ◄ MANs (20-60 Gbps) or backbone loops for site access International connections Core network fiber path is ~ 14,000 miles / 24,000 km 2700 miles / 4300 km

ESnet Virtual Circuit Service Traffic isolation and traffic engineering Provides for high-performance, non-standard transport mechanisms that cannot co-exist with commodity TCP-based transport Enables the engineering of explicit paths to meet specific requirements e.g. bypass congested links, using lower bandwidth, lower latency paths Guaranteed bandwidth (Quality of Service (QoS)) User specified bandwidth Addresses deadline scheduling Where fixed amounts of data have to reach sites on a fixed schedule, so that the processing does not fall far enough behind that it could never catch up – very important for experiment data analysis Secure The circuits are “secure” to the edges of the network (the site boundary) because they are managed by the control plane of the network which is isolated from the general traffic Provides end-to-end connections between Labs and collaborator institutions

Virtual Circuit Service Functional Requirements Support user/application VC reservation requests Source and destination of the VC Bandwidth, start time, and duration of the VC Traffic characteristics (e.g. flow specs) to identify traffic designated for the VC Manage allocations of scarce, shared resources Authentication to prevent unauthorized access to this service Authorization to enforce policy on reservation/provisioning Gathering of usage data for accounting Provide circuit setup and teardown mechanisms and security Widely adopted and standard protocols (such as MPLS and GMPLS) are well understood within a single domain Cross domain interoperability is the subject of ongoing, collaborative development secure and-to-end connection setup is provided by the network control plane Enable the claiming of reservations Traffic destined for the VC must be differentiated from “regular” traffic Enforce usage limits Per VC admission control polices usage, which in turn facilitates guaranteed bandwidth Consistent per-hop QoS throughout the network for transport predictability

OSCARS Overview OSCARS Guaranteed Bandwidth Virtual Circuit Services On-demand Secure Circuits and Advance Reservation System Web Services APIs OSCARS Guaranteed Bandwidth Virtual Circuit Services Path Computation Topology Reachability Contraints Scheduling AAA Availability Provisioning Signaling Security Resiliency/Redundancy

The Mechanisms Underlying OSCARS Based on Source and Sink IP addresses, route of LSP between ESnet border routers is determined using topology information from OSPF-TE. Path of LSP can be explicitly directed to take SDN network. On the SDN Ethernet switches all traffic is MPLS switched (layer 2.5), which stitches together VLANs VLAN 1 VLAN 2 VLAN 3 On ingress to ESnet, packets matching reservation profile are filtered out (i.e. policy based routing), policed to reserved bandwidth, and injected into a LSP. SDN SDN SDN SDN Link SDN Link RSVP, MPLS enabled on internal interfaces Sink Label Switched Path IP Link Source IP Link IP IP IP high-priority queue standard, best-effort queue MPLS labels are attached onto packets from Source and placed in separate queue to ensure guaranteed bandwidth. Regular production traffic queue. Interface queues

Environment of Science is Inherently Multi-Domain End points will be at independent institutions – campuses or research institutes - that are served by ESnet, Abilene, GÉANT, and their regional networks Complex inter-domain issues – typical circuit will involve five or more domains - of necessity this involves collaboration with other networks For example, a connection between FNAL and DESY involves five domains, traverses four countries, and crosses seven time zones FNAL (AS3152) [US] GEANT (AS20965) [Europe] DESY (AS1754) [Germany] ESnet (AS293) [US] DFN (AS680) [Germany]

Interdomain Virtual Circuit Reservation Control Flow Progress!

OSCARS Status Update ESnet Centric Deployment Prototype layer 3 (IP) guaranteed bandwidth virtual circuit service deployed in ESnet (1Q05) Layer 2 (Ethernet VLAN) virtual circuit service under development Inter-Domain Collaborative Efforts Terapaths Inter-domain interoperability for layer 3 virtual circuits demonstrated (3Q06) Inter-domain interoperability for layer 2 virtual circuits under development HOPI/DRAGON Inter-domain exchange of control messages demonstrated (1Q07) Initial integration of OSCARS and DRAGON has been successful (1Q07) DICE First draft of topology exchange schema has been formalized (in collaboration with NMWG) (2Q07), interoperability test scheduled for 3Q07 Drafts on reservation and signaling messages under discussion UVA Integration of Token based authorization in OSCARS under discussion Measurements Hybrid dataplane testing with ESnet, HOPI/DRAGON, USN, and Tennessee Tech (1Q07) Administrative Vangelis Haniotakis (GRNET) has taken a one-year sabbatical position with ESnet to work on interdomain topology exchange, resource scheduling, and signalling

Monitoring Applications Move Networks Toward Service-Oriented Communications Services perfSONAR is a global collaboration to design, implement and deploy a network measurement framework. Web Services based Framework Measurement Archives (MA) Measurement Points (MP) Lookup Service (LS) Topology Service (TS) Authentication Service (AS) Some of the currently Deployed Services Utilization MA Circuit Status MA & MP Latency MA & MP Bandwidth MA & MP Looking Glass MP Topology MA This is an Active Collaboration The basic framework is complete Protocols are being documented New Services are being developed and deployed.

perfSONAR Collaborators ARNES Belnet CARnet CESnet Dante University of Delaware DFN ESnet FCCN FNAL GARR GEANT2 * Plus others who are contributing, but haven’t added their names to the list on the WIKI. Georga Tech GRNET Internet2 IST POZNAN Supercomputing Center Red IRIS Renater RNP SLAC SURFnet SWITCH Uninett

perfSONAR Deployments 16+ different networks have deployed at least 1 perfSONAR service (Jan 2007)

E2Emon - a perfSONAR application E2Emon provides end-to-end path status in a service-oriented, easily interpreted way a perfSONAR application used to monitor the LHC paths end-to-end across many domains uses perfSONAR protocols to retrieve current circuit status every minute or so from MAs and MPs in all the different domains supporting the circuits is itself a service that produces Web based, real-time displays of the overall state of the network, and it generates alarms when one of the MP or MA’s reports link problems.

E2Emon: Status of E2E link CERN-LHCOPN-FNAL-001 E2Emon generated view of the data for one OPN link [E2EMON]

Path Performance Monitoring Path performance monitoring needs to provide users/applications with the end-to-end, multi-domain traffic and bandwidth availability should also provide real-time performance such as path utilization and/or packet drop Multiple path performance monitoring tools are in development One example – Traceroute Visualizer [TrViz] ,developed by Joe Metzger, ESnet – has been deployed at about 10 R&E networks in the US and Europe that have at least some of the required perfSONAR MA services to support the tool

Traceroute Visualizer Forward direction bandwidth utilization on application path from LBNL to INFN-Frascati (Italy) traffic shown as bars on those network device interfaces that have an associated MP services (the first 4 graphs are normalized to 2000 Mb/s, the last to 500 Mb/s) 1 ir1000gw (131.243.2.1) 2 er1kgw 3 lbl2-ge-lbnl.es.net 4 slacmr1-sdn-lblmr1.es.net (GRAPH OMITTED) 5 snv2mr1-slacmr1.es.net (GRAPH OMITTED) 6 snv2sdn1-snv2mr1.es.net 7 chislsdn1-oc192-snv2sdn1.es.net (GRAPH OMITTED) 8 chiccr1-chislsdn1.es.net 9 aofacr1-chicsdn1.es.net (GRAPH OMITTED) 10 esnet.rt1.nyc.us.geant2.net (NO DATA) 11 so-7-0-0.rt1.ams.nl.geant2.net (NO DATA) 12 so-6-2-0.rt1.fra.de.geant2.net (NO DATA) 13 so-6-2-0.rt1.gen.ch.geant2.net (NO DATA) 14 so-2-0-0.rt1.mil.it.geant2.net (NO DATA) 15 garr-gw.rt1.mil.it.geant2.net (NO DATA) 16 rt1-mi1-rt-mi2.mi2.garr.net 17 rt-mi2-rt-rm2.rm2.garr.net (GRAPH OMITTED) 18 rt-rm2-rc-fra.fra.garr.net (GRAPH OMITTED) 19 rc-fra-ru-lnf.fra.garr.net (GRAPH OMITTED) 20 21 www6.lnf.infn.it (193.206.84.223) 189.908 ms 189.596 ms 189.684 ms link capacity is also provided

Federated Trust Services – Support for Large-Scale Collaboration III. Federated Trust Services – Support for Large-Scale Collaboration Remote, multi-institutional, identity authentication is critical for distributed, collaborative science in order to permit sharing widely distributed computing and data resources, and other Grid services Public Key Infrastructure (PKI) is used to formalize the existing web of trust within science collaborations and to extend that trust into cyber space The function, form, and policy of the ESnet trust services are driven entirely by the requirements of the science community and by direct input from the science community International scope trust agreements that encompass many organizations are crucial for large-scale collaborations ESnet has lead in negotiating and managing the cross-site, cross-organization, and international trust relationships to provide policies that are tailored for collaborative science This service, together with the associated ESnet PKI service, is the basis of the routine sharing of HEP Grid-based computing resources between US and Europe

DOEGrids CA (Active Certificates) Usage Statistics US, LHC ATLAS project adopts ESnet CA service * Report as of July 5, 2007

DOEGrids CA Usage - Virtual Organization Breakdown ** OSG Includes (BNL, CDF, CMS, CompBioGrid,DES, DOSAR, DZero, Engage, Fermilab, fMRI, GADU, geant4, GLOW, GPN, GRASE, GridEx, GROW, GUGrid, i2u2, iVDGL, JLAB, LIGO, mariachi, MIS, nanoHUB, NWICG, OSG, OSGEDU, SBGrid,SDSS, SLAC, STAR & USATLAS) * DOE-NSF collab. & Auto renewals

DOEGrids CA Adopts Red Hat CS 7.1 Motivation: SunOne CMS (Certificate Management System) went End-of-Life 2 years ago RH CS is a continuation of the original CA product and development team, and is fully supported by Red Hat Transition was over a year in negotiation, development, and testing 05 July 2007: Transition from SunONE CMS 4.7 to Red Hat Major transition - Minimal outage of 7 hours Preserved important assets Existing DOEGrids signing key Entire history: over 35000 data objects transferred UI (for subscriber-users and operators) New CA software will allow ESnet to develop more useful applications and interfaces for the user communities

ESnet Conferencing Service (ECS) IV. ESnet Conferencing Service (ECS) An ESnet Science Service that provides audio, video, and data teleconferencing service to support human collaboration of DOE science Seamless voice, video, and data teleconferencing is important for geographically dispersed scientific collaborators Provides the central scheduling essential for global collaborations ESnet serves more than a thousand DOE researchers and collaborators worldwide H.323 (IP) videoconferences (4000 port hours per month and rising) audio conferencing (2500 port hours per month) (constant) data conferencing (150 port hours per month) Web-based, automated registration and scheduling for all of these services Very cost effective (saves the Labs a lot of money)

ESnet Collaboration Services (ECS)

ECS Video Collaboration Service High Quality videoconferencing over IP and ISDN Reliable, appliance based architecture Ad-Hoc H.323 and H.320 multipoint meeting creation Web Streaming options on 3 Codian MCU’s using Quicktime or Real 3 Codian MCUs with Web Conferencing Options 120 total ports of video conferencing on each MCU (40 ports per MCU) 384k access for video conferencing systems using ISDN protocol Access to audio portion of video conferences through the Codian ISDN Gateway

ECS Voice and Data Collaboration 144 usable ports Actual conference ports readily available on the system. 144 overbook ports Number of ports reserved to allow for scheduling beyond the number of conference ports readily available on the system. 108 Floater Ports Designated for unexpected port needs. Floater ports can float between meetings, taking up the slack when an extra person attends a meeting that is already full and when ports that can be scheduled in advance are not available. Audio Conferencing and Data Collaboration using Cisco MeetingPlace Data Collaboration = WebEx style desktop sharing and remote viewing of content Web-based user registration Web-based scheduling of audio / data conferences Email notifications of conferences and conference changes 650+ users registered to schedule meetings (not including guests)

ECS Service Level ESnet Operations Center is open for service 24x7x365. A trouble ticket is opened within15 to 30 minutes and assigned to the appropriate group for investigation. Trouble ticket is closed when the problem is resolved. ECS support is provided Monday to Friday, 8AM to 5 PM Pacific Time excluding LBNL holidays Reported problems are addressed within 1 hour from receiving a trouble ticket during ECS support period ESnet does NOT provide a real time (during-conference) support service

Typical Problems Reported to ECS Support Video Conferencing User E.164 look up Gatekeeper registration problems – forgotten IP address or user network problems Gateway Capacity for ISDN service expanded to 2 full PRI’s = 46 x 64kbps chs For the most part, problems are with user-side network and systems configuration. Voice and Data Collaboration Scheduling Conflicts and Scheduling Capacity has been addressed by expanding overbooking capacity to 100% of actual capacity Future equipment plans will allow for optimal configuration of scheduling parameters. Browser compatibility with Java based data sharing client – users are advised to test before meetings Lost UserID and/or passwords Words of Wisdom We advise users that at least two actions must be taken in advance of conferences to reduce the likelihood of problems: A) testing of the configuration to be used for the audio, video and data conference. B) appropriate setup time must be allocated BEFORE the conference to ensure punctuality and correct local configuration. (at least 15 min recommended)

Limited Human and Financial resources currently prohibit ESnet from: Real Time ECS Support A number of user groups have requested “real-time” conference support (monitoring of conferences while in-session) Limited Human and Financial resources currently prohibit ESnet from: A) Making real time information available to the public on the systems status (network, ECS, etc) This information is available only on some systems to our support personnel B) 24x7x365 real-time support C) Addressing simultaneous trouble calls as in a real time support environment. This would require several people addressing multiple problems simultaneously

Real Time ECS Support Proposed solution A fee-for-service arrangement for real-time conference support Such an arrangement could be made by contracting directly with TKO Video Communications, ESnet’s ECS service provider Service offering would provide: Testing and configuration assistance prior to your conference Creation and scheduling of your conferences on ECS Hardware Preferred port reservations on ECS video and voice systems Connection assistance and coordination with participants Endpoint troubleshooting Live phone support during conferences Seasoned staff and years of experience in the video conferencing industry ESnet community pricing at $xxx per hour (Commercial Price: $yyy/hr)

Summary ESnet is currently satisfying its mission by enabling SC science that is dependant on networking and distributed, large-scale collaboration: “The performance of ESnet over the past year has been excellent, with only minimal unscheduled down time. The reliability of the core infrastructure is excellent. Availability for users is also excellent” - DOE 2005 annual review of LBL ESnet has put considerable effort into gathering requirements from the DOE science community, and has a forward-looking plan and expertise to meet the five-year SC requirements A Lehman review of ESnet (Feb, 2006) has strongly endorsed the plan presented here

References High Performance Network Planning Workshop, August 2002 http://www.doecollaboratory.org/meetings/hpnpw Science Case Studies Update, 2006 (contact eli@es.net) DOE Science Networking Roadmap Meeting, June 2003 http://www.es.net/hypertext/welcome/pr/Roadmap/index.html Science Case for Large Scale Simulation, June 2003 http://www.pnl.gov/scales/ Planning Workshops-Office of Science Data-Management Strategy, March & May 2004 http://www-conf.slac.stanford.edu/dmw2004 For more information contact Chin Guok (chin@es.net). Also see http://www.es.net/oscars [LHC/CMS] http://cmsdoc.cern.ch/cms/aprom/phedex/prod/Activity::RatePlots?view=global [ICFA SCIC] “Networking for High Energy Physics.” International Committee for Future Accelerators (ICFA), Standing Committee on Inter-Regional Connectivity (SCIC), Professor Harvey Newman, Caltech, Chairperson. http://monalisa.caltech.edu:8080/Slides/ICFASCIC2007/ [E2EMON] Geant2 E2E Monitoring System –developed and operated by JRA4/WI3, with implementation done at DFN http://cnmdev.lrz-muenchen.de/e2e/html/G2_E2E_index.html http://cnmdev.lrz-muenchen.de/e2e/lhc/G2_E2E_index.html [TrViz] ESnet PerfSONAR Traceroute Visualizer https://performance.es.net/cgi-bin/level0/perfsonar-trace.cgi

And Ending on a Light Note…. NON SEQUITUR - BY WILEY