“The Pacific Research Platform”

Slides:



Advertisements
Similar presentations
Calit2-Living in the Future " Keynote Sharecase 2006 University of California, San Diego March 29, 2006 Dr. Larry Smarr Director, California Institute.
Advertisements

Bringing Mexico Into the Global LambdaGrid Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology Harry E. Gruber.
High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research Larry Smarr Prof. Computer Science and Engineering Director, Calit2 (UC.
Why Optical Networks Are Emerging as the 21 st Century Driver Scientific American, January 2001.
"The OptIPuter: an IP Over Lambda Testbed" Invited Talk NREN Workshop VII: Optical Network Testbeds (ONT) NASA Ames Research Center Mountain View, CA August.
The First Year of Cal-(IT) 2 Report to The University of California Regents UCSF San Francisco, CA March 13, 2002 Dr. Larry Smarr Director, California.
“A California-Wide Cyberinfrastructure for Data-Intensive Research” Invited Presentation CENIC Annual Retreat Santa Rosa, CA July 22, 2014 Dr. Larry Smarr.
AHM Overview OptIPuter Overview Third All Hands Meeting OptIPuter Project San Diego Supercomputer Center University of California, San Diego January 26,
“Building US/Mexico Collaborations Using Optical Networks” Opening Workshop Welcome Big Data Big Network 2 Calit2’s Qualcomm Institute February 10, 2014.
SAN DIEGO SUPERCOMPUTER CENTER Emerging HIPAA and Protected Data Requirements for Research Computing at SDSC Ron Hawkins Director of Industry Relations.
PRISM: High-Capacity Networks that Augment Campus’ General Utility Production Infrastructure Philip Papadopoulos, PhD. Calit2 and SDSC.
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
“An Integrated West Coast Science DMZ for Data-Intensive Research” Panel CENIC Annual Conference University of California, Irvine Irvine, CA March 9, 2015.
“Introduction to UC San Diego’s Integrated Digital Infrastructure” Opening Talk IDI Showcase 2015 University of California, San Diego May 6-7, 2015 Dr.
1 Building National Cyberinfrastructure Alan Blatecky Office of Cyberinfrastructure EPSCoR Meeting May 21,
“A UC-Wide Cyberinfrastructure for Data-Intensive Research” Invited Presentation UC IT Leadership Council Oakland, CA May 19, 2014 Dr. Larry Smarr Director,
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Why Optical Networks Will Become the 21 st Century Driver Scientific American, January 2001 Number of Years Performance per Dollar Spent Data Storage.
“An Integrated Science Cyberinfrastructure for Data-Intensive Research” Panel CISCO Executive Symposium San Diego, CA June 9, 2015 Dr. Larry Smarr Director,
“Creating a High Performance Cyberinfrastructure to Support Analysis of Illumina Metagenomic Data” DNA Day Department of Computer Science and Engineering.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
Developing a North American Global LambdaGrid Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology Harry E.
“Comparative Human Microbiome Analysis” Remote Video Talk to CICESE Big Data, Big Network Workshop Ensenada, Mexico October 10, 2013 Dr. Larry Smarr Director,
“The Pacific Research Platform: a Science-Driven Big-Data Freeway System.” Invited Presentation 2015 Campus Cyberinfrastructure PI Workshop Austin, TX.
What is Cyberinfrastructure? Russ Hobby, Internet2 Clemson University CI Days 20 May 2008.
Cal-(IT) 2 : A Public-Private Partnership in Southern California U.S. Business Council for Sustainable Development Year-End Meeting December 11, 2003 Institute.
Introduction to Calit2 Visit by NASA Ames February 29, 2008 Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology.
Chicago/National/International OptIPuter Infrastructure Tom DeFanti OptIPuter Co-PI Distinguished Professor of Computer Science Director, Electronic Visualization.
Innovative Research Alliances Invited Talk IUCRP Fellows Seminar UCSD La Jolla, CA July 10, 2006 Dr. Larry Smarr Director, California Institute for Telecommunications.
A Wide Range of Scientific Disciplines Will Require a Common Infrastructure Example--Two e-Science Grand Challenges –NSF’s EarthScope—US Array –NIH’s Biomedical.
Using Photonics to Prototype the Research Campus Infrastructure of the Future: The UCSD Quartzite Project Philip Papadopoulos Larry Smarr Joseph Ford Shaya.
SoCal Infrastructure OptIPuter Southern California Network Infrastructure Philip Papadopoulos OptIPuter Co-PI University of California, San Diego Program.
Russ Hobby Program Manager Internet2 Cyberinfrastructure Architect UC Davis.
A High-Performance Campus-Scale Cyberinfrastructure For Effectively Bridging End-User Laboratories to Data-Intensive Sources Presentation by Larry Smarr.
“ Calit2-Living in the Future " Briefing The Future in Review (FiRe) 2006 Conference University of California, San Diego May 15, 2006 Dr. Larry Smarr Director,
Project GreenLight Overview Thomas DeFanti Full Research Scientist and Distinguished Professor Emeritus California Institute for Telecommunications and.
The Interaction of UCSD Industrial Partners, the Jacobs School of Engineering, and Cal-(IT) 2 Dr. Larry Smarr Director, California Institute for Telecommunications.
“The Pacific Research Platform: a Science-Driven Big-Data Freeway System.” Opening Presentation Pacific Research Platform Workshop Calit2’s Qualcomm Institute.
Ocean Sciences Cyberinfrastructure Futures Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technologies Harry E.
Cyberinfrastructure: An investment worth making Joe Breen University of Utah Center for High Performance Computing.
The Future from the Perspective of The California Institute for Telecommunications and Information Technology Invited Paper to the Eighteenth IEEE Symposium.
“Building a Regional 100G Collaboration Infrastructure” Keynote Presentation CineGrid International Workshop 2015 Calit2’s Qualcomm Institute University.
The OptIPuter Project Tom DeFanti, Jason Leigh, Maxine Brown, Tom Moher, Oliver Yu, Bob Grossman, Luc Renambot Electronic Visualization Laboratory, Department.
The PRPv1 Architecture Model Panel Presentation Building the Pacific Research Platform Qualcomm Institute, Calit2 UC San Diego October 16, 2015.
“The Pacific Research Platform: a Science-Driven Big-Data Freeway System.” Big Data for Information and Communications Technologies Panel Presentation.
“ Collaborations Between Calit2, SIO, and the Venter Institute—a Beginning " Talk to the UCSD Representative Assembly La Jolla, CA November 29, 2005 Dr.
“ Ultra-Broadband and Peta-Scale Collaboration Opportunities Between UC and Canada Summary Talk Canada - California Strategic Innovation Partnership Summit.
Slide 1 UCSC 100 Gbps Science DMZ – 1 year 9 month Update Brad Smith & Mary Doyle.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
“The UCSD Big Data Freeway System” Invited Short Talk Workshop on “Enriching Human Life and Society” UC San Diego February 6, 2014 Dr. Larry Smarr Director,
“ OptIPuter Year Five: From Research to Adoption " OptIPuter All Hands Meeting La Jolla, CA January 22, 2007 Dr. Larry Smarr Director, California.
UCSD’s Distributed Science DMZ
Building the Pacific Research Platform: Science Engagement WORKSHOP, OCT. 15, 2015 OPENING REMARKS, CAMILLE CRITTENDEN, CITRIS.
Southern California Infrastructure Philip Papadopoulos Greg Hidley.
“Genomics: The CAMERA Project" Invited Talk 5 th Annual ON*VECTOR International Photonics Workshop UCSD February 28, 2006 Dr. Larry Smarr Director,
University of Illinois at Chicago Lambda Grids and The OptIPuter Tom DeFanti.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
“Pacific Research Platform Science Drivers” Opening Remarks PRP Science Driver PI Workshop UC Davis March 23, 2016 Dr. Larry Smarr Director, California.
“The Pacific Research Platform” Opening Keynote Lecture 15th Annual ON*VECTOR International Photonics Workshop Calit2’s Qualcomm Institute University of.
High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research Larry Smarr Prof. Computer Science and Engineering Director, Calit2 (UC.
“OptIPuter: From the End User Lab to Global Digital Assets" Panel UC Research Cyberinfrastructure Meeting October 10, 2005 Dr. Larry Smarr.
“ Building an Information Infrastructure to Support Microbial Metagenomic Sciences " Presentation to the NBCR Research Advisory Committee UCSD La Jolla,
“The Pacific Research Platform Two Years In”
Thomas Hutton – SDSC/Cailit2, University of California, San Diego
“A National Big Data Cyberinfrastructure Supporting Computational Biomedical Research” Invited Presentation Symposium on Computational Biology and Bioinformatics:
Optical SIG, SD Telecom Council
The OptIPortal, a Scalable Visualization, Storage, and Computing Termination Device for High Bandwidth Campus Bridging Presentation by Larry Smarr to.
Presentation transcript:

“The Pacific Research Platform” Briefing to The Quilt Visit to Calit2’s Qualcomm Institute University of California, San Diego February 10, 2016 Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology Harry E. Gruber Professor, Dept. of Computer Science and Engineering Jacobs School of Engineering, UCSD http://lsmarr.calit2.net

Use Lightpaths to Connect All Data Generators and Consumers, Vision: Creating a West Coast “Big Data Freeway” Connected by CENIC/Pacific Wave Use Lightpaths to Connect All Data Generators and Consumers, Creating a “Big Data” Freeway Integrated With High Performance Global Networks “The Bisection Bandwidth of a Cluster Interconnect, but Deployed on a 20-Campus Scale.” This Vision Has Been Building for Over a Decade

NSF’s OptIPuter Project: Demonstrating How SuperNetworks Can Meet the Needs of Data-Intensive Researchers LS Slide 2005 2003-2009 $13,500,000 OptIPortal– Termination Device for the OptIPuter Global Backplane In August 2003, Jason Leigh and his students used RBUDP to blast data from NCSA to SDSC over the TeraGrid DTFnet, achieving18Gbps file transfer out of the available 20Gbps Calit2 (UCSD, UCI), SDSC, and UIC Leads—Larry Smarr PI Univ. Partners: NCSA, USC, SDSU, NW, TA&M, UvA, SARA, KISTI, AIST Industry: IBM, Sun, Telcordia, Chiaro, Calient, Glimmerglass, Lucent

DOE ESnet’s Science DMZ: A Scalable Network Design Model for Optimizing Science Data Transfers A Science DMZ integrates 4 key concepts into a unified whole: A network architecture designed for high-performance applications, with the science network distinct from the general-purpose network The use of dedicated systems for data transfer Performance measurement and network testing systems that are regularly used to characterize and troubleshoot the network Security policies and enforcement mechanisms that are tailored for high performance science environments The DOE ESnet Science DMZ and the NSF “Campus Bridging” Taskforce Report Formed the Basis for the NSF Campus Cyberinfrastructure Network Infrastructure and Engineering (CC-NIE) Program Science DMZ Coined 2010 http://fasterdata.es.net/science-dmz/

Based on Community Input and on ESnet’s Science DMZ Concept, NSF Has Funded Over 100 Campuses to Build Local Big Data Freeways Red 2012 CC-NIE Awardees Yellow 2013 CC-NIE Awardees Green 2014 CC*IIE Awardees Blue 2015 CC*DNI Awardees Purple Multiple Time Awardees 2012-2015 CC-NIE / CC*IIE / CC*DNI Programs Source: NSF

The Pacific Research Platform: The Next Logical Step – Connect Multiple Campus Science DMZs with 10-100Gbps Lightpaths NSF CC*DNI Grant $5M 10/2015-10/2020 PI: Larry Smarr, UC San Diego Calit2 Co-Pis: Camille Crittenden, UC Berkeley CITRIS, Tom DeFanti, UC San Diego Calit2, Philip Papadopoulos, UC San Diego SDSC, Frank Wuerthwein, UC San Diego Physics and SDSC

FIONA – Flash I/O Network Appliance: Termination Device for 10-100Gbps Flows FIONAs Are Science DMZ Data Transfer Nodes & Optical Network Termination Devices UCSD CC-NIE Prism Award & UCOP Phil Papadopoulos & Tom DeFanti Joe Keefe & John Graham UCOP Rack-Mount Build: John Graham, Calit2’s QI Cost $8,000 $20,000 Intel Xeon Haswell Multicore E5-1650 v3 6-Core 2x E5-2697 v3 14-Core RAM 128 GB 256 GB SSD SATA 3.8 TB Network Interface 10/40GbE Mellanox 2x40GbE Chelsio+Mellanox GPU NVIDIA Tesla K80 RAID Drives 0 to 112TB (add ~$100/TB)

FIONAs as Uniform DTN End Points FIONA DTNs UC FIONAs Funded by UCOP “Momentum” Grant FIONAs as Uniform DTN End Points Existing DTNs As of October 2015

FIONA DTNs Now Deployed to All UC Campuses Ten Week Sprint to Demonstrate the West Coast Big Data Freeway System: PRPv0 FIONA DTNs Now Deployed to All UC Campuses And Most PRP Sites Presented at CENIC 2015 March 9, 2015

Pacific Research Platform Multi-Campus Science Driver Teams Jupyter Hub Biomedical Cancer Genomics Hub/Browser Microbiome and Integrative ‘Omics Integrative Structural Biology Earth Sciences Data Analysis and Simulation for Earthquakes and Natural Disasters Climate Modeling: NCAR/UCAR California/Nevada Regional Climate Data Analysis CO2 Subsurface Modeling Particle Physics Astronomy and Astrophysics Telescope Surveys Galaxy Evolution Gravitational Wave Astronomy Scalable Visualization, Virtual Reality, and Ultra-Resolution Video Campus Cyberinfrastructure – Network Infrastructure and Engineering (CC-NIE) Campus Cyberinfrastructure – Infrastructure, Innovation, and Engineering (CC-IIE) Campus Cyberinfrastructure – Data, Networking, and Innovation (CC-DNI) NSF 15-534 incorporates Data Infrastructure Building Blocks (CC-DNI-DIBBs) – Multi-Campus / Multi-Institution Model Implementation from Program Solicitation NSF 14-530

PRP First Application: Distributed IPython/Jupyter Notebooks: Cross-Platform, Browser-Based Application Interleaves Code, Text, & Images IScilab IMatlab ICSharp Bash Clojure Kernel Hy Kernel Redis Kernel jove, a kernel for io.js IJavascript Calysto Scheme Calysto Processing idl_kernel Mochi Kernel Lua (used in Splash) Spark Kernel Skulpt Python Kernel MetaKernel Bash MetaKernel Python Brython Kernel IVisual VPython Kernel IJulia IHaskell IFSharp IRuby IGo IScala IMathics Ialdor LuaJIT/Torch Lua Kernel IRKernel (for the R language) IErlang IOCaml IForth IPerl IPerl6 Ioctave Calico Project kernels implemented in Mono, including Java, IronPython, Boo, Logo, BASIC, and many others Source: John Graham, QI

PRP UC-JupyterHub Backbone Next Step: Deploy Across PRP Source: John Graham, Calit2 UC Berkeley UC San Diego GPU JupyterHub: 2 x 14-core CPUs 256GB RAM 1.2TB FLASH 3.8TB SSD Nvidia K80 GPU Dual 40GbE NICs And a Trusted Platform Module GPU JupyterHub: 1 x 18-core CPUs 128GB RAM 3.8TB SSD Nvidia K80 GPU Dual 40GbE NICs And a Trusted Platform Module 40Gbps

OSG Federates Clusters in 40/50 States: A Major XSEDE Resource Source: Miron Livny, Frank Wuerthwein, OSG

Open Science Grid Has Had a Huge Growth Over the Last Decade - Currently Federating Over 130 Clusters Source: Miron Livny, Frank Wuerthwein, OSG CMS Crossed 100 Million Core-Hours/Month In Dec 2015 Supported Over 200 Million Jobs In 2015 ATLAS Over 1 Billion Data Transfers Moved 200 Petabytes In 2015

PRP Prototype of LambdaGrid Aggregation of OSG Software & Services Across California Universities in a Regional DMZ ATLAS CMS other physics life sciences other sciences Aggregate Petabytes of Disk Space & PetaFLOPs of Compute, Connected at 10-100 Gbps Transparently Compute on Data at Their Home Institutions & Systems at SLAC, NERSC, Caltech, UCSD, & SDSC UCD SLAC OSG Hours 2015 by Science Domain UCSC CSU Fresno PRP Builds on SDSC’s LHC-UC Project UCSB Caltech UCR UCI Source: Frank Wuerthwein, UCSD Physics; SDSC; co-PI PRP UCSD & SDSC

Two Automated Telescope Surveys Creating Huge Datasets Will Drive PRP Precursors to LSST and NCSA PRP Allows Researchers to Bring Datasets from NERSC to Their Local Clusters for In-Depth Science Analysis 300 images per night. 100MB per raw image 30GB per night 120GB per night 250 images per night. 530MB per raw image 150 GB per night 800GB per night When processed at NERSC Increased by 4x Source: Peter Nugent, Division Deputy for Scientific Engagement, LBL Professor of Astronomy, UC Berkeley

Creates 10M Alerts/Night Within 1 Minute of Observing Global Scientific Instruments Will Produce Ultralarge Datasets Continuously Requiring Dedicated Optic Fiber and Supercomputers Square Kilometer Array Large Synoptic Survey Telescope Tracks ~40B Objects, Creates 10M Alerts/Night Within 1 Minute of Observing 2x40Gb/s https://tnc15.terena.org/getfile/1939 www.lsst.org/sites/default/files/documents/DM%20Introduction%20-%20Kantor.pdf https://tnc15.terena.org/getfile/1939

PRP Will Support the Computation and Data Analysis in the Search for Sources of Gravitational Radiation Augment the aLIGO Data and Computing Systems at Caltech, by connecting at 10Gb/s to SDSC Comet supercomputer, enabling LIGO computations to enter via the same PRP “job cache” as for LHC.

PRP CENIC 10G Link UCSD to SDSU HPWREN Users and Public Safety Clients Gain Redundancy and Resilience from PRP Upgrade PRP CENIC 10G Link UCSD to SDSU DTN FIONAs Endpoints Data Redundancy Disaster Recovery High Availability Network Redundancy San Diego Countywide Sensors and Camera Resources UCSD & SDSU Data & Compute UCSD UCR SDSU UCI 10X Increase During Wildfires Data From Hans-Werner Braun UCI & UCR Data Replication and PRP FIONA Anchors as HPWREN Expands Northward Source: Frank Vernon, Greg Hidley, UCSD

PRP Backbone Sets Stage for 2016 Expansion of HPWREN, Connected to CENIC, into Orange and Riverside Counties Anchor to CENIC at UCI PRP FIONA Connects to CalREN-HPR Network Data Replication Site Potential Future UCR CENIC Anchor Camera and Relay Sites at: Santiago Peak Sierra Peak Lake View Bolero Peak Modjeska Peak Elsinore Peak Sitton Peak Via Marconi UCR UCI UCSD Collaborations through COAST – County of Orange Safety Task Force SDSU Source: Frank Vernon, Greg Hidley, UCSD

PRP Links FIONA Clusters Creating Distributed Virtual Reality 20x40G PRP-connected 40G FIONAs WAVE@UC San Diego CAVE@UC Merced

PRP is NOT Just for Big Data Science and Engineering: Linking Cultural Heritage and Archaeology Datasets UCD UCSF Stanford NASA AMES/ NREN UCSC UCSB Caltech USC UCLA UCI UCSD SDSU UCR Esnet DoE Labs UW/ PNWGP Seattle Berkeley UCM Los Nettos Internet2 Building on CENIC’s Expansion To Libraries, Museums, and Cultural Sites We already have 11 major research universities in California poised to partner. “In an ideal world –Extremely high bandwidth to move large cultural heritage datasets around the PRP cloud for processing & viewing in CAVEs around PRP with Unlimited Storage for permanent archiving.” -Tom Levy, UCSD * Institutions with Active Archaeology Programs Note: This diagram represents a subset of sites and connections.

Current International Next Step: Global Research Platform Building on CENIC/Pacific Wave and GLIF Current International GRP Partners