Network Developments and Network Monitoring in Internet2 Eric Boyd Director of Performance Architecture and Technologies Internet2.

Slides:



Advertisements
Similar presentations
Connect. Communicate. Collaborate Click to edit Master title style MODULE 1: perfSONAR TECHNICAL OVERVIEW.
Advertisements

Connect. Communicate. Collaborate Towards Multi-domain Monitoring for the Research Networks Nicolas Simar, Dante TNC 2005, Poznan, June 2005.
Performance Update Eric Boyd Director of Performance Architecture and Technologies Internet2.
Connect. Communicate. Collaborate GÉANT2 JRA1 & perfSONAR Loukik Kudarimoti, DANTE 28 th May, 2006 RNP Workshop, Curitiba.
Connect. Communicate. Collaborate Introduction to perfSONAR Loukik Kudarimoti, DANTE 27 th September, 2006 SEEREN2 Summer School, Heraklion.
LHC Monitoring: perfSONAR Overview Eric L. Boyd Director of Performance Architecture and Technologies Internet2.
Network Performance Measurement Atlas Tier 2 Meeting at BNL December Joe Metzger
1 ESnet Network Measurements ESCC Feb Joe Metzger
13 September 2015 The Abilene Observatory and Network Research Rick Summerhill, Director Network Research, Architecture, and Technology, Internet2 Joint.
GEANT Performance Monitoring Infrastructure – Joint Techs meeting July Nicolas Simar GEANT’s Performance Monitoring.
Connect. Communicate. Collaborate perfSONAR and Wavelengths Monitoring LHC meeting, Cambridge, 16 of June 2006 Matthias Hamm - DFN Nicolas Simar - DANTE.
Abilene Update Joint Techs Summer ’05 Vancouver, CA Steve Cotter Director, Network Services Steve Cotter Director, Network Services.
Performance Monitoring - Internet2 Member Meeting -- Nicolas Simar Performance Monitoring Internet2 Member Meeting, Indianapolis.
The Research and Education Network: Platform for Innovation Heather Boyles, Next Generation Network Symposium Malaysia 2007-March-15.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Jerry Sobieski MAX GigaPoP and TSC Program Manager Mark Johnson MCNC.
A Technology Vision for the Future Rick Summerhill, Chief Technology Officer, Eric Boyd, Deputy Technology Officer, Internet2 Joint Techs Meeting 16 July.
27-Jan-2005 Internet2 Activities Toward a Global Measurement Infrastructure Matt Zekauskas Network Performance Measurement and Monitoring APAN19.
 End to End Performance Initiative Russ Hobby HENP Networking Working Group Meeting, Ann Arbor, Michigan 26 October 2001.
The Internet2 Network Observatory Rick Summerhill Director Network Research, Architecture, and Technologies Brian Cashman Network Planning Manager Matt.
Internet2 Performance Update Jeff W. Boote Senior Network Software Engineer Internet2.
1 ESnet Update Joint Techs Meeting Minneapolis, MN Joe Burrescia ESnet General Manager 2/12/2007.
1 Measuring Circuit Based Networks Joint Techs Feb Joe Metzger
Rick Summerhill Chief Technology Officer, Internet2 TIP January 2008 Honolulu, HI Internet2 Update.
A Framework for Internetworking Heterogeneous High-Performance Networks via GMPLS and Web Services Xi Yang, Tom Lehman Information Sciences Institute (ISI)
Delivering Circuit Services to Researchers: The HOPI Testbed Rick Summerhill Director, Network Research, Architecture, and Technologies, Internet2 Joint.
Internet2 E2EPI End-2-End Performance Initiative Jeff W. Boote Internet2 29 May 2006.
Connect. Communicate. Collaborate Implementing Multi-Domain Monitoring Services for European Research Networks Szymon Trocha, PSNC A. Hanemann, L. Kudarimoti,
ASCR/ESnet Network Requirements an Internet2 Perspective 2009 ASCR/ESnet Network Requirements Workshop April 15/16, 2009 Richard Carlson -- Internet2.
Abilene Observatory Chris Robb Indiana University APAN Engineering Workshop 2004 Slides prepared by Chris Small, IU Global NOC Software Engineer.
Measurement on the Internet2 Network: an evolving story Matt Zekauskas Joint Techs, Minneapolis 11-Feb-2007.
1 Network Measurement Summary ESCC, Feb Joe Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
Internet2 Network Observatory Update Matt Zekauskas, Measurement SIG 2006 Fall Member Meeting 4-Dec-2006.
Connect. Communicate. Collaborate perfSONAR MDM Service for LHC OPN Loukik Kudarimoti DANTE.
Advanced Networks: The Past and the Future – The Internet2 Perspective APAN 7 July 2004, Cairns, Australia Douglas Van Houweling, President & CEO Internet2.
Performance Update Eric L. Boyd Director of Performance Architecture and Technologies Internet2 Eric L. Boyd Director of Performance Architecture and Technologies.
PerfSONAR-PS Functionality February 11 th 2010, APAN 29 – perfSONAR Workshop Jeff Boote, Assistant Director R&D.
13-Oct-2003 Internet2 End-to-End Performance Initiative: piPEs Eric Boyd, Matt Zekauskas, Internet2 International.
Jeremy Nowell EPCC, University of Edinburgh A Standards Based Alarms Service for Monitoring Federated Networks.
© 2006 Open Grid Forum Network Monitoring and Usage Introduction to OGF Standards.
Internet2 End-to-End Performance Initiative Eric L. Boyd Director of Performance Architecture and Technologies Internet2.
PerfSONAR-PS Working Group Aaron Brown/Jason Zurawski January 21, 2008 TIP 2008 – Honolulu, HI.
PerfSONAR WG 2006 Spring Member Meeting Jeff W. Boote 24 April 2006.
NORDUnet NORDUnet e-Infrastrucure: Grids and Hybrid Networks Lars Fischer CTO, NORDUnet Fall 2006 Internet2 Member Meeting, Chicago.
DICE: Authorizing Dynamic Networks for VOs Jeff W. Boote Senior Network Software Engineer, Internet2 Cándido Rodríguez Montes RedIRIS TNC2009 Malaga, Spain.
1 LHCOPN Monitoring Directions January 2007 Joe Metzger
05 October 2001 End-to-End Performance Initiative Network Measurement Matt Zekauskas, Fall 2001 Internet2 Member Meeting Network Measurement.
Connect communicate collaborate perfSONAR MDM News Domenico Vicinanza DANTE (UK)
1 Network related topics Bartosz Belter, Wojbor Bogacki, Marcin Garstka, Maciej Głowiak, Radosław Krzywania, Roman Łapacz FABRIC meeting Poznań, 25 September.
Advanced Network Diagnostic Tools Richard Carlson EVN-NREN workshop.
The Internet2 Network and LHC Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 Given by Rich Carlson LHC Meeting 25.
Connecting to the new Internet2 Network What to Expect… Steve Cotter Rick Summerhill FMM 2006 / Chicago.
1 Deploying Measurement Systems in ESnet Joint Techs, Feb Joseph Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
1 Network Measurement Challenges LHC E2E Network Research Meeting October 25 th 2006 Joe Metzger Version 1.1.
Status of perfSONAR Tools Jason Zurawski April 23, 2007 Spring Member Meeting.
HOPI Update Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 Joint Techs 17 July 2006 University of Wisconsin, Madison,
The Internet2 Network and LHC Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 LHC Meeting 23 October 2006 FERMI Lab,
Bob Jones EGEE Technical Director
perfSONAR WG Meeting (06FMM)
Internet2 End-to-End Performance Initiative
Networking for the Future of Science
Robert Szuman – Poznań Supercomputing and Networking Center, Poland
PerfSONAR: Development Status
Internet2 Performance Update
Deployment & Advanced Regular Testing Strategies
ESnet Network Measurements ESCC Feb Joe Metzger
E2E piPEs Overview Eric L. Boyd Internet2 24 February 2019.
Interoperable Measurement Frameworks: Internet2 E2E piPEs and NLANR Advisor Eric L. Boyd Internet2 17 April 2019.
“Detective”: Integrating NDT and E2E piPEs
Internet2 E2E piPEs Project
Presentation transcript:

Network Developments and Network Monitoring in Internet2 Eric Boyd Director of Performance Architecture and Technologies Internet2

Overview Internet2 Network Performance Middleware: Supporting Network-Based Science Internet2 Network Observatory

An Asset for the Community Universities Researchers Regional Networks K-12 Industry International Internet2 Network: An Asset for the Community Universities Researchers Regional Networks K-12 Industry International

Internet2 Network Hybrid optical and IP network Dynamic and static wavelength services Fiber, equipment dedicated to Internet2; Level 3 maintains network and service level Platform supports production services and experimental projects

Internet2 Network - Layer 1 Internet2 Network Optical Switching Node Level3 Regen Site Internet2 Redundant Drop/Add Site ESnet Drop/Add Site

Internet2 Network Deployment Internet2 Network Optical Switching Node Level3 Regen Site Internet2 Redundant Drop/Add Site ESnet Drop/Add Site

Internet2 Network Deployment Internet2 Network Optical Switching Node Level3 Regen Site Internet2 Redundant Drop/Add Site ESnet Drop/Add Site

Internet2 Network Deployment Internet2 Network Optical Switching Node Level3 Regen Site Internet2 Redundant Drop/Add Site ESnet Drop/Add Site

The New Internet2 Network

A New Wrinkle Internet2 exploring a merger with National Lambda Rail (NLR) Goal: Consolidate national higher education and research networking organizations Technical team is exploring what the merged technical infrastructure will look like

Overview Internet2 Network Performance Middleware: Supporting Network-Based Science Internet2 Network Observatory

Network-Based Science Science is a global community Networks links scientists Collaborative research occurs across network boundaries For the scientist, the value of the network is the achieved network performance Scientists should not have to focus on the network; good end-to-end performance should be a given

Large Hadron Collider International Physics facility located in CERN Switzerland Major US involvement 2 major US data repositories (PetaBytes/year) 17 US Institutions provide data analysis and storage 68 Universities and National Laboratories with scientists looking at the data Dedicated transatlantic networks connect US to CERN Advanced network services required over existing campus, connector/regional, and national networks

Cyberinfrastructure Security Middleware Performance End to End Policy Cyberinfrastructure

Achieving Good End-to-End Performance Internet2 consists of: Campuses Corporations Regional networks Internet2 backbone network Our members care about connecting with: Other members Government labs & networks International partners The Internet2 community cares about making all of this work

Identifying the Problem Applications Developer System Administrator LAN Administrator Campus Networking Gigapop BackboneCampus Networking LAN Administrator System Administrator Applications Developer How do you solve a problem along a path? Hey, this is not working right! The computer Is working OK Talk to the other guysEverything is AOK No other complaints The network is lightly loaded All the lights are green We don’t see anything wrong Looks fine Others are getting in ok Not our problem

Status Quo Performance is excellent across backbone networks Performance is a problem end-to-end Problems are concentrated towards the edge and in network transitions We need to: Diagnose: Understand limits of performance Address: Work with members and application communities to address those performance issues

Vision: Performance Information is … Available People can find it (Discovery) “Community of trust” allows access across administrative domain boundaries (AA) Ubiquitous Widely deployed (Paths of interest covered) Reliable (Consistently configured correctly) Valuable Actionable (Analysis suggests course of action) Automatable (Applications act on data)

Goal: No more mystery … Increase network awareness Set user expectations accurately Reduce diagnostic costs Performance problems noticed early Performance problems addressed efficiently Network engineers can see & act outside their turf Transform application design Incorporate network intuition into application behavior

Strategy: Build & Empower the Community Decouple the Problem Space: Analysis and Visualization Performance Data Sharing Performance Data Generation Grow the Footprint: Clean APIs and protocols between each layer Widespread deployment of measurement infrastructure Widespread deployment of common performance measurement tools

Tactics: Leverage position Internet2 is leveraged to help provide diagnostic information for “US backbone” portion of problem Create *some* diagnostic tools (BWCTL, NDT, OWAMP) Make network data as public as is reasonable Work on efforts to more widely make performance data available (perfSONAR) Contribute to ‘base’ perfSONAR development (partnership with ESnet, Europe, and Brazil) Contribute to standards for performance information sharing (Open Grid Forum Network Measurement Working Group) Integrate ‘our’ diagnostic tools as ‘good’ example of perfSONAR services

From the scientist’s perspective On behalf of the scientist, network engineer or application can easily/automatically: Discover additional monitoring resources Authenticate locally Authorized to use remote network resources to a limited extent Acquire performance monitoring data from remote sites via standard protocol Innovate where needed Customize the analysis and visualization

Internet2 End-to-End Performance Initiative (E2Epi) Includes: Internet2 staff Internet2 members Federal partners International partners Building: Performance monitoring tools Performance middleware frameworks Performance improvement tools

Support for E2Epi Funded out of network revenues Partnerships Leveraging GÉANT2, ESnet, and RNP resources through consortium leadership Grants NSF Apps - Targeted Assistance and Instrumentation for Internet2 Applications NSF SGER - Leveraging Internet2 Facilities for the Network Research Community NSF SGER2 - Network Measurement for International Connections NSF BTG - Bridging the Gap: End-to-End Networking for Landmark Applications NLM Pilot - User Experience with the High Performance Internet Infrastructure: Critical Incidents of Success and Failure NLM NDT - Enhancing the Web 100-based Network Diagnostic Tool

Current Activities Analysis/Diagnostic tools Performance tools Software distributions to enable partner network organizations to participate Google Summer of Code New network deployment of measurement infrastructure on new observatory

Software Distributions NPToolkit (Network Performance Toolkit) Will include much of the following eventually NDT (avail now) OWAMP (avail now) BWCTL/Thrulay (May) AMI (Fall?) Regular testing and collection for OWAMP/BWCTL perfSONAR-PS (Earliest - Fall) SNMP collection/archive AMI archive TopoS and L2Status perfSONAR UI’s (Earliest - Winter)

Google Summer of Code 5 Projects NDT enhancements Phoebus protocol enhancements Chrolog (user-space timestamp) OWAMP (Java Client) perfSONAR/cacti interface

OWAMP (3.0c) One-way latencies Full support of RFC 4656 Deployment Status Abilene: all remaining nms4 hosts New network newy and chic (nms-rlat) Software available at:

BWCTL (1.2b) Throughput Test Controller Pending Software release Additional throughput tools Iperf/thrulay/nuttcp More tolerant of questionable clocks Deployment Status Abilene: open TCP testing New network - awaiting new software release

What is perfSONAR? Performance Middleware perfSONAR is an international consortium in which Internet2 and GÉ ANT2 are founders and leading participants perfSONAR is a set of protocol standards for interoperability between measurement and monitoring systems perfSONAR is a set of open source web services that can be mixed-and-matched and extended to create a performance monitoring framework

perfSONAR Design Goals Standards-based Modular Decentralized Locally controlled Open Source Extensible Applicable to multiple generations of network monitoring systems Grows “beyond our control” Customized for individual science disciplines

perfSONAR Integrates Network measurement tools Network measurement archives Discovery Authentication and authorization Data manipulation Resource protection Topology

perfSONAR Credits perfSONAR is a joint effort: ESnet GÉANT2 JRA1 Internet2 RNP ESnet includes: ESnet/LBL staff Fermilab Internet2 includes: University of Delaware Georgia Tech SLAC Internet2 staff GÉANT2 JRA1 includes: Arnes Belnet Carnet Cesnet CYNet DANTE DFN FCCN GRNet GARR ISTF PSNC Nordunet (Uninett) Renater RedIRIS Surfnet SWITCH

perfSONAR Adoption R&E Networks Internet2 ESnet GÉANT2 European NRENs RNP Application Communities LHC Roll-out to other application communities in late 2007 Distributed Development Individual projects (10 before first release) write components that integrate into the overall framework Individual communities (5 before first release) write their own analysis and visualization software

perfSONAR-PS* perfSONAR (Perl Services) Why? Adoption of Java Services difficult Many network administrators don’t do Java, but are fluent in Perl) Services more directly targeted at the data available from Internet2 observatory deployment.

perfSONAR Deployment Status

Demo …

Overview Internet2 Network Performance Middleware: Supporting Network-Based Science Internet2 Network Observatory

History and Motivation Original Abilene racks included measurement devices Included a single (somewhat large) PC Early OWAMP, Surveyor measurements Optical splitters at some locations Motivation was primarily operations, monitoring, and management - understanding the network and how well it operates Data was collected and maintained whenever possible Primarily a NOC function Available to other network operators to understand the network It became apparent that the datasets were valuable as a network research tool

Rick Summerhill The Abilene Upgrade Network

Upgrade of the Abilene Observatory An important decision was made during the Abilene upgrade process (Juniper T-640 routers and OC-192c) Two racks, one of which was dedicated to measurement Potential for research community to collocate equipment Two components to the Observatory Collocation - network research groups are able to collocate equipment in the Abilene router nodes Measurement - data is collected by the NOC, the Ohio ITEC, and Internet2, and made available to the research community

An Abilene router node Power Out-of-band Eth. Switch T-640 (M-5) Power (48VDC) Measurement Machines (nms) Space for Collocation! Measurement (Observatory) Rack

Dedicated servers at each node Houston Router Node - In this picture: Measurement machines Collocated PlanetLab machines

Example Research Projects Collocation projects PlanetLab – Nodes installed in all Abilene Router Nodes. See The Passive Measurement and Analysis Project (PMA) - The Router clamp. See Projects using collected datasets. See projects.html projects.html “Modular Strategies for Internetwork Monitoring” “Algorithms for Network Capacity Planning and Optimal Routing Based on Time-Varying Traffic Matrices” “Spatio-Temporal Network Analysis” “Assessing the Presence and Incidence of Alpha Flows in Backbone Networks”

The New Internet2 Network Expanded Layer 1, 2 and 3 Facilities Includes SONET and Wave equipment Includes Ethernet Services Greater IP Services Requires a new type of Observatory

The New Internet2 Network

The New Internet2 Observatory Seek Input from the Community, both Engineers and Network Researchers Current thinking is to support three types of services Measurement (as before) Collocation (as before) Experimental Servers to support specific projects - for example, Phoebus (this is new) Support different types of nodes: Optical Nodes Router Nodes

The New York Node - First Installment

Existing Observatory Capabilities One way latency, jitter, loss IPv4 and IPv6 (“owamp”) Regular TCP/UDP throughput tests – ~1 Gbps IPv4 and IPv6; On-demand available (“bwctl”) SNMP Octets, packets, errors; collected 1/min Flow data Addresses anonymized by 0-ing the low order 11 bits Routing updates Both IGP and BGP - Measurement device participates in both Router configuration Visible Backbone – Collect 1/hr from all routers Dynamic updates Syslog; also alarm generation (~nagios); polling via router proxy

Observatory Functions DeviceFunctionDetails nms-rthr1MeasurementBWCTL on-demand 1 Gpbs router throughput, Thrulay nms-rthr2MeasurementBWCTL on-demand 10 Gbps router throughput, Thrulay nms-rexpExperimentalNDT/NPAD nms-rpsvMeasurementNetflow collector nms-rlatMeasurementOWAMP with locally attached GPS timing nms-rphoExperimentalPhoebus 2 x 10GE to Multiservice Switch nms-octrManagementControls Multiservice Switch nms-oexpExperimentalNetFPGA nms-othrMeasurementOn-demand Multiservice Switch 10 Gbps throughput

Router Nodes

Optical Nodes

Observatory Hardware Dell 1950 and Dell 2950 servers Dual Core 3.0 GHz Xeon processors 2 GB memory Dual RAID 146 GB disk Integrated 1 GE copper interfaces 10 GE interfaces Hewlett-Packard 10GE switches 9 servers at router sites, 3 at optical only sites

Observatory Databases – Datа Types Data is collected locally and stored in distributed databases Databases Usage Data Netflow Data Routing Data Latency Data Throughput Data Router Data Syslog Data

Uses and Futures Some uses of existing datasets and tools Quality Control Network Diagnosis Network Characterization Network Research Consultation with researchers Open questions

Observatory Deployment (July) NDT/NPAD servers OWAMP/BWCTL deployments perfSONAR services (perfSONAR-PS) LS (Discovery) SNMP collection/archive OWAMP/BWCTL archive TopoS and L2 status (GN2 E2EMon compatible)

Recall: Datasets Usage Data Netflow Data Routing Data Latency Data Throughput Data Router Data Syslog Data ND, NR ND, NC, NR NR QC, ND, NR ND, NR NR And, of course, most used for operations

Quality Control: e-VLBI When starting to connect telescopes, needed to verify inter-site paths Set up throughput testing among sites (using same Observatory tool: bwctl) Kashima, JP Onsala, SE Boston, MA (Haystack) Collect and graph data; distribute via web Quick QC check before applications tests start

Quality Control: e-VLBI Network

Quality Control: eVLBI Result Automated monitoring allowed view of network throughput variation over time Highlights route changes, network outages Automated monitoring also helps to highlight any throughput issues at end points: e.g. Network Interface Card failures, untuned TCP Stacks Integrated monitoring provides overall view of network behavior at a glance

Network Diagnosis: e-VLBI Target at the time: 50Mbps Oops: Onsala-Boston: 1Mbps Divide and Conquer Verify Abilene backbone tests look good Use Abilene test point in Washington DC Eliminated European and trans-Atlantic pieces Focus on problem: found oversubscribed link

Quality Control: IP Backbone Machines with 1GE interfaces, 9000 MTU Full mesh IPv4 and IPv6 Expect > 950 Mbps TCP Keep list of “Worst 10” If any path < 900 Mbps for two successive testing intervals, throw alarm

Quality Control: Peerings Internet2 and ESnet have been watching the latency across peering points for a while. Internet2 and DREN have been preparing to do some throughput and latency testing During the course of this set up, found interesting routing and MTU size issues

Network Diagnosis: End Hosts NDT, NPAD servers Quick check from a host that has a browser Easily eliminate (or confirm) last mile problems (buffer sizing, duplex mismatch, …) NPAD can find switch limitations, provided the server is close enough

Network Diagnosis: Generic Generally looking for configuration & loss Don’t forget security appliances Is there connectivity & reasonable latency? (ping -> OWAMP) Is routing reasonable (traceroute, proxy) Is host reasonable (NDT; NPAD) Is path reasonable (BWCTL)

Network Characterization Flow data collected with flow-tools package All data not used for security alerts and analysis [REN-ISAC] is anonymized Reports from anonymized data available (see truncated addresses) Additionally, some Engineering reports

Network Research Projects Major consumption Flows Routes Configuration Nick Feamster (while at MIT) Dave Maltz (while at CMU) Papers in SIGCOMM, INFOCOM Hard to track folks that just pull data off of web sites

Lots of Work to be Done Internet2 Observatory realization inside racks set for initial deployment, including new research projects (NetFPGA, Phoebus) Software and links easily changed Could add or change hardware depending on costs Researcher tools, new datasets Consensus on passive data

Not Just Research Operations and Characterization of new services Finding problems with stitched together VLANs Collecting and exporting data from Dynamic Circuit Service... Ciena performance counters Control plane setup information Circuit usage (not utilization, although that is also nice) Similar for underlying Infinera equipment And consider inter-domain issues

Sharing Observatory Data We want to make Internet2 Network Observatory Data: Available: Access to existing active and passive measurement data Ability to run new active measurement tests Interoperable: Common schema and semantics, shared across other networks Single format XML-based discovery of what’s available

Internet2 Deployment Status Focus is on development of services for Internet2 new network and integration with Indiana NOC Submitting a proposal to NSF for additional funding Target: July 2007 as new Internet2 network goes operation OWAMP MA BWCTL MA/MP IU-based Topology Service Multi-LS NOC Alarm Transformation Service

More Information Eric Boyd