ESnet and Science DMZs: an update from the US

Slides:



Advertisements
Similar presentations
Presentation at WebEx Meeting June 15,  Context  Challenge  Anticipated Outcomes  Framework  Timeline & Guidance  Comment and Questions.
Advertisements

1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21) NSF-wide Cyberinfrastructure Vision People, Sustainability, Innovation,
Netscape Application Server Application Server for Business-Critical Applications Presented By : Khalid Ahmed DS Fall 98.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CF21) IRNC Kick-Off Workshop July 13,
The "Earth Cube” Towards a National Data Infrastructure for Earth System Science Presentation at WebEx Meeting July 11, 2011.
1 GENI: Global Environment for Network Innovations Jennifer Rexford Princeton University
NGNS Program Managers Richard Carlson Thomas Ndousse ASCAC meeting 11/21/2014 Next Generation Networking for Science Program Update.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Lawrence Berkeley National Laboratory Kathy Yelick Associate Laboratory Director for Computing Sciences.
Welcome to HTCondor Week #14 (year #29 for our project)
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Big Data: Movement, Crunching, and Sharing Guy Almes, Academy for Advanced Telecommunications 13 February 2015.
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program CASC, May 3, ADVANCED SCIENTIFIC COMPUTING RESEARCH An.
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program NERSC Users Group Meeting Department of Energy Update September.
National Center for Supercomputing Applications NCSA OPIE Presentation November 2000.
Campus Cyberinfrastructure – Network Infrastructure and Engineering (CC-NIE) Kevin Thompson NSF Office of CyberInfrastructure April 25, 2012.
What is Cyberinfrastructure? Russ Hobby, Internet2 Clemson University CI Days 20 May 2008.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
Cyberinfrastructure for Data Intensive Science (DIS) Follow-on panel to DIS session at Internet2/ESCC Joint Techs Conference Baton Rouge – January 24,
Catawba County Board of Commissioners Retreat June 11, 2007 It is a great time to be an innovator 2007 Technology Strategic Plan *
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
Cyberinfrastructure: An investment worth making Joe Breen University of Utah Center for High Performance Computing.
Minority-Serving Institutions (MSI) Cyberinfrastructure (CI) Institute [MSI-CI 2 ] and CI Empowerment Coalition MSI-CIEC October Geoffrey Fox
1 OFFICE OF ADVANCED SCIENTIFIC COMPUTING RESEARCH The NERSC Center --From A DOE Program Manager’s Perspective-- A Presentation to the NERSC Users Group.
Slide 1 UCI Lightpath A Dedicated Network for Research on UCI Campus Jessica Yu Office of Information Technology, UC Irvine NSF PI Workshop 9/29/2015.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Slide 1 Regional Network Collaborations Marla Meehl, UCAR/FRGP CC*IIE Collaborative Research: CC*IIE Region: Rocky Mountain Cyberinfrastructure Mentoring.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
NSF Middleware Initiative Purpose To design, develop, deploy and support a set of reusable, expandable set of middleware functions and services that benefit.
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
SCIENCE_DMZ NETWORKS STEVE PERRY, DIRECTOR OF NETWORKS UNM PIYASAT NILKAEW, DIRECTOR OF NETWORKS NMSU.
HUIT Cloud Initiative Update November, /20/2013 Ryan Frazier & Rob Parrott.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
1 User-Focused Engagement and Service Development Interactive Panel Discussion on Science Engagement TNC16 June 2016.
Campus Cyberinfrastructure (CC*) Program NSF Larry Gallery Manager, Membership Development and K-20 Program Manager June 16,
The Minority-Serving Institutions (MSI) Cyberinfrastructure (CI) Institute [MSI C(I) 2 ] Providing a scalable mechanism for developing a CI-enabled science.
UNM SCIENCE DMZ Sean Taylor Senior Network Engineer.
A Brief Introduction to NERSC Resources and Allocations
Supporting the NRP with a Lean CI Staff
Computing Clusters, Grids and Clouds Globus data service
Deploying Regional Grids Creates Interaction, Ideas, and Integration
RDA US Science workshop Arlington VA, Aug 2014 Cees de Laat with many slides from Ed Seidel/Rob Pennington.
Tools and Services Workshop
Netscape Application Server
Joslynn Lee – Data Science Educator
Electron Ion Collider New aspects of EIC experiment instrumentation and computing, as well as their possible impact on and context in society (B) COMPUTING.
Innovative Solutions from Internet2
The NSRC cultivates collaboration among a community of peers to build and improve a global Internet that benefits all parties. We facilitate the growth.
of our Partners and Customers
DOE Facilities - Drivers for Science: Experimental and Simulation Data
Grid Computing.
Research Data Transfer Zones
Study course: “Computing clusters, grids and clouds” Andrey Y. Shevel
Regional Software Defined Science DMZ (SD-SDMZ)
XSEDE’s Campus Bridging Project
Canada’s National Research and Education Network (NREN)
State of XSEDE: XSEDE14 John Towns PI and Project Director, XSEDE
PRPv1 Discussion topics
Big-Data around the world
Storing and Accessing G-OnRamp’s Assembly Hubs outside of Galaxy
NTU Presentation Jason Arviso, Director of Information Technology
NTU Presentation Jason Arviso, Director of Information Technology
Interoperable Measurement Frameworks: Internet2 E2E piPEs and NLANR Advisor Eric L. Boyd Internet2 17 April 2019.
The National Grid Service Mike Mineter NeSC-TOE
Data Management Components for a Research Data Archive
NSF Middleware Initiative
The Intelligent Enterprise and SAP Business One
OU BATTLECARD: Oracle WebCenter Training
Presentation transcript:

ESnet and Science DMZs: an update from the US Tony Hey Chief Data Scientist Rutherford Appleton Lab STFC

DOE’s Esnet6 Plans With thanks to Eli Dart, LBNL

ESnet – Connecting Facilities At Scale Eli Dart, Science Engagement Energy Sciences Network (ESnet) Lawrence Berkeley National Laboratory Lawrence Berkeley National Laboratory April 13, 2018

ESnet - the basic facts High-speed international networking facility, optimized for data-intensive science: connecting 50 labs, plants and facilities with >150 networks, universities, research partners globally supporting every science office, and serving as an integral extension of many instruments 400Gbps transatlantic extension in production since Dec 2014 >1.3 Tbps of external connectivity, including high speed access to commercial partners such as Amazon growing number university connections to better serve LHC science (and eventually: Belle II) older than commercial Internet, growing ~twice as fast Areas of strategic focus: software, science engagement. Engagement effort now 12% of staff Software capability critical to next-generation network

ESnet is a dedicated mission network engineered to accelerate a broad range of science outcomes One way to think of ESnet is as it is – a network We do this by offering unique capabilities, and optimizing the network for data acquisition, data placement, data sharing, data mobility. 11/22/2018

Beyond the network core: The Science DMZ Friction-free network path Dedicated data transfer nodes (DTNs) Performance monitoring (perfSONAR) Over 120 labs and universities in the US have deployed this ESnet architecture NSF has invested >>$120M to accelerate adoption UK, Australia, Brazil, others following suit http://fasterdata.es.net/science-dmz/ I normally give entire talk on this idea, but will summarize in a single slide.

Data and HPC: The Petascale DTN Project Built on top of the Science DMZ Effort to improve data transfer performance between the DOE ASCR HPC facilities at ANL, LBNL, and ORNL, and also NCSA. Multiple current and future science projects need to transfer data between HPC facilities Performance goal of 15 gigabits per second (equivalent to 1PB/week) Realize performance goal for routine Globus transfers without special tuning Reference data set is 4.4TB of cosmology simulation data Use performant, easy-to-use tools with production options on Globus Transfer service (previously Globus Online) Use GUI just like a user would, with default options E.g. integrity checksums enabled, as they should be 11/22/2018

DTN Cluster Performance – HPC Facilities (2017) Final state at successful project conclusion 11/22/2018

Improvements in Multiple Places … DTN clusters expanded ALCF: 8 DTNs to 12 NERSC: 4 DTNs to 8 ORNL: 8 DTNs to 16 (for Globus pool) Network Upgrades ORNL: Nx10G to Nx40G NCSA: Nx10G to Nx100G NERSC: Nx10G to Nx40G Unrelated Globus change gave a big performance boost Change in behavior of transfers with large file counts This beneficially affects many production transfers at HPC facilities 11/22/2018

Petascale DTN Lifts All Boats Petascale DTN project benefits all projects which use the HPC facility DTNs Example: multi-facility users Many projects use multiple HPC facilities At-scale data movement between DOE HPC facilities is a huge benefit Example: Modern Research Data Portal architecture Data portals which use modern architecture benefit from DTN improvements DTN scaling/improvements benefit all data portals which use the same DTN cluster 11/22/2018

Legacy Portal Design Very difficult to improve performance without architectural change Software components all tangled together Difficult to put the whole portal in a Science DMZ because of security Even if you could put it in a DMZ, many components aren’t scalable How do we change the architecture? Legacy portal model: Web server with filesystem-level access to data store CGI, perl/python, database, etc. for portal logic all run on the web server Evolved in place over 15-ish years Need to work smarter not harder – re-think the architecture 11/22/2018

Next-Generation Portal Leverages Science DMZ Portal logic still runs in the enterprise network. However, instead of handing out pointers to data objects stored within the portal server, the portal hands out pointers to data objects stored on the big filesystem and accessible via the DTNs. https://peerj.com/articles/cs-144/ 11/22/2018

NSF’s Campus Cyberinfrastructure (CC NSF’s Campus Cyberinfrastructure (CC*) Plans With thanks to Kevin Thompson, NSF

Networking Programs in CISE/OAC Cognizant PO: Kevin Thompson kthompso@nsf.gov Networking as a fundamental layer and underpinning of CI CC* - Campus Cyberinfrastructure Campus networking upgrade (re-design to scienceDMZ at campus border, 10/100Gbps) and innovation program IRNC – International R&E Network Connections Scientific discovery as a global collaborative endeavor Provide network connections linking U.S. research with peer networks in other parts of the world – 100Gbps links, software defined exchanges Supports all R&E US data flows (not just NSF-funded) Includes performance flow measurement, monitoring, training November 15, 2017 Office of Advanced Cyberinfrastructure - Supercomputing 2017 Birds-of-a-Feather

CC* - Campus Cyberinfrastructure Cognizant PO: Kevin Thompson kthompso@nsf.gov NSF 18-508 – proposals due January 30, 2018 CC* awards will be supported in four program areas: (1) Data Driven Networking Infrastructure for the Campus and Researcher awards will be supported at up to $500,000 total for up to 2 years; (2) Network Design and Implementation for Small Institutions awards will be supported at up to $750,000 total for up to 2 years; (3) Network Integration and Applied Innovation awards will be supported at up to $1,000,000 total for up to 2 years; and (4) Network Performance Engineering and Outreach awards will be supported at up to $3,500,000 total for up to 4 years. [NEW for 2018] November 15, 2017 Office of Advanced Cyberinfrastructure - Supercomputing 2017 Birds-of-a-Feather

CC* Program Awards 2012-2017: ~225 awards Cognizant PO: Kevin Thompson kthompso@nsf.gov November 15, 2017 Office of Advanced Cyberinfrastructure - Supercomputing 2017 Birds-of-a-Feather

2017 OAC Networking Program Examples – Broadening Participation Guam – CC* award to U of Guam and IRNC-supported open exchange point #1659182: “CC* Network Design: Upgrading the University of Guam Network to Connect to Internet2 and Create a Science DMZ” Tribal Colleges and Universities CI Study - #1655185 – “TCU Cyberinfrastructure Initiative: A Study of Tribal College and University Cyberinfrastructure and Supported STEM Programs:, PI: Al Kuslikis, AIHEC 37 TCUs November 15, 2017 Office of Advanced Cyberinfrastructure - Supercomputing 2017 Birds-of-a-Feather