(OGF NSI-WG co-chairs)

Slides:



Advertisements
Similar presentations
© 2006 Open Grid Forum Network Services Interface OGF30: Connection Services Guy Roberts, 27 th Oct 2010.
Advertisements

Use cases for implementation of the NSI interface Takahiro Miyamoto, Nobutaka Matsumoto KDDI R&D Laboratories Inc. This work is partially supported by.
© 2006 Open Grid Forum Network Services Interface Introduction to NSI Guy Roberts.
Circuit Monitoring July 16 th 2011, OGF 32: NMC-WG Jason Zurawski, Internet2 Research Liaison.
Cloud Computing – a view from the Network cloud Martin Swany.
NSI wg Architecture Elements John Vollbrecht Internet2.
© 2006 Open Grid Forum Network Services Interface OGF30: Working Group Meeting Guy Roberts, Inder Monga, Tomohiro Kudoh 27 th Oct 2010.
NSI Architecture Document Status and Work to be done: A view from a chair Inder Monga ESNet.
Lawrence Berkeley National LaboratoryU.S. Department of Energy | Office of Science Network Service Interface (NSI) Inder Monga Co-chair, Network Services.
NORDUnet Nordic infrastructure for Research & Education NSI in the SDN Environment (from perspective of an NSI fellow) Jerry Sobieski NORDUnet Presented.
© 2006 Open Grid Forum Network Service Interface in a Nut Shell GEC 19, Atlanta, GA Presenter: Chin Guok (ESnet) Contributors: Tomohiro Kudoh (AIST), John.
NORDUnet Nordic infrastructure for Research & Education The OGF Network Services Interface Framework An Overview, Status, and Futures Presented to: OGF.
NORDUnet Nordic infrastructure for Research & Education LHCONE  NSI Adaptation An Applications’ Perspective Jerry Sobieski NORDUnet CERN Dec. 13/
National Science Foundation Arlington, Virginia January 7-8, 2013 Tom Lehman University of Maryland Mid-Atlantic Crossroads.
Automated GOLEs and Fenius: Pragmatic interoperability Winter Joint Techs 2011 Clemson, SC Evangelos Chaniotakis, ESnet Network Engineer Lawrence Berkeley.
Connect. Communicate. Collaborate Afrodite Sevasti, GRNET 8th Annual Global LambdaGrid Workshop Seattle, 1 st October 2008 Developments.
InterDomain Dynamic Circuit Network Demo Joint Techs - Hawaii Jan 2008 John Vollbrecht, Internet2
A Framework for Internetworking Heterogeneous High-Performance Networks via GMPLS and Web Services Xi Yang, Tom Lehman Information Sciences Institute (ISI)
Bandwidth-on-Demand evolution Gerben van Malenstein Fall 2011 Internet2 Member Meeting Raleigh, North Carolina, USA – October 3, 2011.
Hybrid MLN DOE Office of Science DRAGON Hybrid Network Control Plane Interoperation Between Internet2 and ESnet Tom Lehman Information Sciences Institute.
Connect. Communicate. Collaborate BANDWIDTH-ON-DEMAND SYSTEM CASE-STUDY BASED ON GN2 PROJECT EXPERIENCES Radosław Krzywania (speaker) PSNC Mauro Campanella.
OGF DMNR BoF Dynamic Management of Network Resources Documents available at: Guy Roberts, John Vollbrecht.
Dynamic Lightpath Services on the Internet2 Network Rick Summerhill Director, Network Research, Architecture, Technologies, Internet2 TERENA May.
LHC OPEN NETWORK ENVIRONMENT STATUS UPDATE Artur Barczyk/Caltech Tokyo, May 2013 May 14, 2013
© 2006 Open Grid Forum Network Services Interface OGF 32, Salt Lake City Guy Roberts, Inder Monga, Tomohiro Kudoh 16 th July 2011.
LHC Open Network Environment Architecture Overview and Status Artur Barczyk/Caltech LHCONE meeting Amsterdam, September 26 th,
© 2006 Open Grid Forum Network Monitoring and Usage Introduction to OGF Standards.
Internet2 Dynamic Circuit Services and Tools Andrew Lake, Internet2 July 15, 2007 JointTechs, Batavia, IL.
Connect. Communicate. Collaborate NRENs on the AutoBAHN Afrodite Sevasti, GRNET Radek Krzywania, PSNC Guy Roberts, DANTE TERENA Networking Conference
NSI Aggregator: Joint SURFnet/ESnet effort LHCONE Workshop CERN (Geneva, CH) Feb 10-11, 2014 NSI PCE Development Team.
© 2006 Open Grid Forum Network Services Interface Document roadmap, April 2014 Guy Roberts, Chin Guok, Tomohiro Kudoh.
IDCP and NSI: Lessons Learned, Deployments and Gap Analysis Chin Guok, Inder Monga OGF 34 Oxford, UK.
© 2010 Open Grid Forum STP and TF how they work Tomohiro Kudoh.
1 Network Services Interface Connection Service v2.0 Tomohiro Kudoh (AIST) (OGF NSI-WG)
NSI Service Definition Federation of providers A group of network providers get together and decide that they wish to offer a multi-domain connection services.
1 Network Services Interface An Interface for Requesting Dynamic Inter- datacenter Networks Tomohiro Kudoh (AIST) Guy Roberts (DANTE) Inder Monga (ESnet)
ESnet’s Use of OpenFlow To Facilitate Science Data Mobility Chin Guok Inder Monga, and Eric Pouyoul OGF 36 OpenFlow Workshop Chicago, Il Oct 8, 2012.
Topology Issues in Inter-Domain Connection Services Jerry Sobieski (NORDUnet) The Cynic’s Perspective & Jeroen van der Ham (University of Amsterdam) The.
Lawrence Berkeley National LaboratoryU.S. Department of Energy | Office of Science Network Service Interface: Concepts and Architecture Inder Monga Guy.
NSI Topology v2.0 Version 1.2 John MacAuley, ESNET September 22, 2014 Uppsala.
© 2006 Open Grid Forum The Network Services Interface An Overview of the NSI Framework and the GLIF Automated GOLE dynamic network provisioning demonstration.
Service Oriented Architecture (SOA) Prof. Wenwen Li School of Geographical Sciences and Urban Planning 5644 Coor Hall
OGF NSI CS Protocol State Machine Converged single view July 16, 2011.
Inter-Domain Network Provisioning Technology for LHC data transfer
Multi-layer software defined networking in GÉANT
Welcome Network Virtualization & Hybridization Thomas Ndousse
OGF NSI CS Protocol State Machine
Dynamic Circuit Networks Topology Description Issues
Service Provider Requirements for Ethernet Control with GMPLS
Update on Grid User Network Interface (GUNI) Draft
BoD workshop Cambridge
Inder Monga Co-chair, OGF NSI-WG
GÉANT BoD Service Roadmap
GÉANT Multi-Domain Bandwidth-on-Demand Service
InterDomain Dynamic Circuit Network Demo
A Deterministic End to End Performance Verification Architecture
NSI wg Architecture Elements
Grid Resource Allocation Agreement Protocol Working Group
Grid Network Services: Lessons from SC04 draft-ggf-bas-sc04demo-0.doc
Network Services Interface Working Group
Integration of Network Services Interface version 2 with the JUNOS Space SDK
Network Services Interface gateway for future network services
Availability Query / Internal Topology
LHC Open Network Project status and site involvement
Service Provider Requirements for Ethernet Control with GMPLS
Dynamic Circuit Service Hands-On GMPLS Engineering Workshop
AutoGOLE Dashboard presented by Cees de Laat
Network Services Interface Working Group
Interdomain Dynamic Circuits
OSCARS Roadmap Chin Guok
Presentation transcript:

(OGF NSI-WG co-chairs) Network Services Interface An open standard for dynamic circuit service interoperability Tomohiro Kudoh (AIST) Guy Roberts (DANTE) Inder Monga (ESnet) (OGF NSI-WG co-chairs)

OGF Network Services Interface Dynamic circuit services have been recently introduced by many R&E networks. The Open Grid Forum Network Services Interfaces working group (OGF NSI-WG) has been working to define an open interface standard to make such services interoperable among networks. Currently, the NSI-WG is working to finalize a Connection Service document that describes both a service architecture and a protocol for end-to-end circuit provisioning. Additional services, such as a Topology Service, are being proposed as extensions to the NSI interface.

Example: G-lambda project G-lambda project has been defining an interface Joint project of KDDI R&D labs., NTT, NICT and AIST, started in 2005. http://www.g-lambda.net/ The goal of this project is to define a web services interface (GNS-WSI) to request heterogeneous resources (network, computers, storages etc.)

Why NSI is required? Demand for high bandwidth stable end-to-end network circuits is increasing. Applications will include e-science and high definition video transfer. Some R&E and commercial networks are rolling out bandwidth-on-demand services, but there is no open standard interface which makes them interoperable. Networks are starting to be recognized as manageable resources of cloud infrastructure. Dynamic network provisioning will support composition of data-intensive inter-cloud services.

Overview of NSI (1) NSI models a one-to-one relationship between a service provider’s network and a Network Service Agent (NSA). The NSI is a service interface between NSAs. Each NSA can play the role of a requester, a provider, or both. Multiple NSAs forms a recursive framework of requesters and providers. NSI defines representation of service topology which is an abstraction of multi-layer physical network topology.

Overview of NSI (2) NSI service will be composed of multiple services including a connection service (CS) which is currently being defined. CS supports advance reservation of end-to-end circuits. A reserved circuit with properties such as VLANid and/or designated guaranteed bandwidth is dynamically provisioned. Circuit provisioning is done either automatically at the designated start-time (auto start), or when explicitly triggered during the reservation period (manual start).

NSI architecture and connection service

Provisioning Sequences Manual Provisioning Automatic Provisioning Requester Provider Requester Provider reserve.rq reserve.cf provision.rq provision.cf release.rq release.cf terminate.rq terminate.cf reserve.rq reserve.cf Start time provision.rq Start time provision.cf In service release.rq In service release.cf Reserved Reserved provision.rq provision.cf In service terminate.rq In service terminate.cf

Initial NSI state machine Auto Provision >rsv.rq (reservation) (start_time) (provision) >prov.rq Provisioning (provision_cf) <prov.cf Reserving (reservation_cf) <rsv.cf Reserved >prov.rq (provision) (provision_fl) <prov.fl (start_time) (reservation_fl) <rsv.fl <rsv.fl, >term.rq (terminate) Scheduled Provisioned >prov.rq <prov.cf Cleaning >rel.rq (release) (release_fl) <rel.fl (release_cf) <rel.cf (forced_end) <fcd_end <fcd_end, >term.rq (release) (terminate) >rel.rq (release) Releasing (terminate_cf) <term.cf (end_time) (release) (terminate) Terminated (terminate_cf) <term.cf Terminating >term.rq (release)(terminate) Any State* *: excluding “Initial”, “Cleaning”, “ Terminating” and “Terminated” states (terminate_fl) <term.fl (reservation_cf) <rsv.cf (provision_cf) <prov.cf (release_cf) <rel.cf (terminate_cf) <term.cf >prov.rq <prov.fl >rel.rq <rel.fl >term.rq <term.fl (reservation_fl) <rsv.fl (reservation_fl) <prov.fl (release_fl) <rel.fl

STP and TF: topology abstraction Data plane topology is abstracted by STP (Service Termination Point) and TF (Transfer Function) STP is a logical label of a point at the edge of a network STP is used in a connection request to designate a termination point of intra-network connection. TF represents each network’s capability to dynamically connect two STPs of the network

NSI architecture and connection service

Sample configuration Network Boundary network A and B are Connected using one GbE (green) and one 10GbE (red) VLAN 1-8 on the green cable, no VLAN on the red cable Network Boundary Slot 1 Slot 2 Slot 3 GbE, VLAN 1-8 can be used Switch 1 Switch 1 Port 8 Port 2 10GbE, No VLAN Port 13 Switch 2 Port 7 Network A Network B 12

Logical transport view and internal mappings Logical view (topology) Network A internal mapping: Switch 1/Port8 : X1 Switch1/Port8/VLAN1: : X1/1 Switch 1/Port8/VLAN8 : X1/8 Switch 2/Port13 : X2 Network B internal mapping: Switch 1/Slot1/Port2 : YI Switch 1/Slot1/Port2 /VLAN1 : Ya Switch 1/Slot1/Port2 /VLAN8: Yh Switch 1/Slot2/Port7 : YJ STP: X1/1 STP: Ya STP: X1/2 STP: Yb STP: X1 STP: YI STP: X1/8 STP: Yh STP: X2 STP: YJ Network A Network B STP:X1, YI : GbE STP X1/1 – X1/8, Ya-Yh : VLAN on GbE STP X2, YJ: 10GbE Note: STP names are just symbols (labels) and do not necessarily correspond to physical implementation such as VLANs.

Another example The paired STPs, (a2, b1) is an SDP By requesting Network A Network B a1 TF a2 b1 TF b2 SDP The paired STPs, (a2, b1) is an SDP By requesting connect(a1, a2) to network A connect(b1, b2) to network B a1-b2 is connected. NetworkA advertises: I have a1 and a2. a2 is connected to b1 of networkB TF: I can connect a1 and a2 NetworkB advertises: I have b1 and b2. b1 is connected to a2 of networkA TF: I can connect b1 and b2 By gathering these information from networks, topology information sufficient for discovery can be constructed. 14

Current status - plugfest of NSI - CS protocol v1.0draft, which enables provisioning of simple inter-provider connection, has been defined. At GLIF2011, a “plugfest” (protocol interoperability testing) took place with seven organizations participating from all over the world. Rio de Janeiro, Sep.13 We had 7 implementations: OpenNSA (NORUnet) AutoBAHN (GEANT) DRAC (SURFnet) G-LAMBDA (AIST) G-LAMBDA (KDDI Labs) OSCARS (ESnet) DynamicKL (KISTI)

Plugfest Topology For the initial NSI protocol testing that is Rio Plugfest, we elected to use an artificial topology: This is the “Rio” ring topology. Each NSA is allowed to define their internal topology as they see fit as long as it is consistent with this inter-domain view. A1 A2 A3 A4 B1 B2 B3 B4 C1 C2 C3 C4 D1 D2 D3 D4 J2 J3 J4 J1 Jamaica Aruba Bonaire Martinique Grenada Dominica M1 M2 M3 M4 G2 G4 G3 G1 OpenNSA DRAC AutoBAHN OSCARS G-LAMBDA/AIST G-LAMBDA/KDDIL DynamicKL Curacao Courtesy of Jerry Sobieski (NORDUnet)

Interoperation test matrix

NSI documents GFD.173 Network Services Framework The NSF is a framework to support Network Services Supports many services – initial service is Connection Service Possible future services, e.g.: Network Topology Exchange Service Status – NSF v1.0 has been published GFD.XXX Connection Service Protocol Allows an application or network provider to request and automatically reserve and provision circuits from other network providers Designed to support circuits that transit multiple service providers Status – protocol freeze this week to support NSI plugfest Sept 2011

Summary and future plan of NSI NSI is an open interface request connectivity R&E networks around world rolling out dynamic services Commitment by providers to become NSI compliant Including JGN-X According to the lessons learned at the plugfest, the protocol has been improved for the SC11 demonstration, and v1.0 will be officially released after SC11. Issues including topology service, security (authentication and authorization), service definitions, and fault processing are being discussed for v2.0 specification which is to be released in early 2012.