A. Sim, CRD, L B N L 1 ANI and Magellan Launch, Nov. 18, 2009 Climate 100: Scaling the Earth System Grid to 100Gbps Networks Alex Sim, CRD, LBNL Dean N.

Slides:



Advertisements
Similar presentations
Extending the Information Power Grid Throughout the Solar System Al Globus, CSC.
Advertisements

1 Birmingham LIGO Gravitational Wave Observatory Replicate 1 TB/day of data to 10+ international sites Uses GridFTP, RFT, RLS, DRS Cardiff AEI/Golm
Fighting Malaria With The Grid. Computing on The Grid The Internet allows users to share information across vast geographical distances. Using similar.
YES New Mexico Enterprise Eligibility System
1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
DOE Global Modeling Strategic Goals Anjuli Bamzai Program Manager Climate Change Prediction Program DOE/OBER/Climate Change Res Div
Earth System Curator Spanning the Gap Between Models and Datasets.
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech.
U.S. Department of Energy’s Office of Science Basic Energy Sciences Advisory Committee Dr. Daniel A. Hitchcock October 21, 2003
Building on the BIRN Workshop BIRN Systems Architecture Overview Philip Papadopoulos – BIRN CC, Systems Architect.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
NGNS Program Managers Richard Carlson Thomas Ndousse ASCAC meeting 11/21/2014 Next Generation Networking for Science Program Update.
The Earth System Grid Discovery and Semantic Web Technologies Line Pouchard Oak Ridge National Laboratory Luca Cinquini, Gary Strand National Center for.
Data Deduplication in Virtualized Environments Marc Crespi, ExaGrid Systems
An Introduction to the Open Science Data Cloud Heidi Alvarez Florida International University Robert L. Grossman University of Chicago Open Cloud Consortium.
Commodity Grid (CoG) Kits Keith Jackson, Lawrence Berkeley National Laboratory Gregor von Laszewski, Argonne National Laboratory.
Presented by The Earth System Grid: Turning Climate Datasets into Community Resources David E. Bernholdt, ORNL on behalf of the Earth System Grid team.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
CoG Kit Overview Gregor von Laszewski Keith Jackson.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
High Performance GridFTP Transport of Earth System Grid (ESG) Data 1 Center for Enabling Distributed Petascale Science.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
1 4/23/2007 Introduction to Grid computing Sunil Avutu Graduate Student Dept.of Computer Science.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks G. Quigley, B. Coghlan, J. Ryan (TCD). A.
The New Zealand Institute for Plant & Food Research Limited Use of Cloud computing in impact assessment of climate change Kwang Soo Kim and Doug MacKenzie.
National Collaboratories Program Overview Mary Anne ScottFebruary 7, rd DOE/NSF Meeting on LHC and Global Computing “Infostructure”
Tools for collaboration How to share your duck tales…
- EGU 2010 ESSI May Building on the CMIP5 effort to prepare next steps : integrate community related effort in the every day workflow to.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
The Earth System Grid: A Visualisation Solution Gary Strand.
Web Portal Design Workshop, Boulder (CO), Jan 2003 Luca Cinquini (NCAR, ESG) The ESG and NCAR Web Portals Luca Cinquini NCAR, ESG Outline: 1.ESG Data Services.
The Earth System Grid (ESG) Computer Science and Technologies DOE SciDAC ESG Project Review Argonne National Laboratory, Illinois May 8-9, 2003.
- Vendredi 27 mars PRODIGUER un nœud de distribution des données CMIP5 GIEC/IPCC Sébastien Denvil Pôle de Modélisation, IPSL.
ESG Observational Data Integration Presented by Feiyi Wang Technology Integration Group National Center of Computational Sciences.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
7. Grid Computing Systems and Resource Management
1 Research and Development. 2 R&D Agenda  Security  Bulk Data Movement  Data Replication and Mirroring  Monitoring  Metrics  Versioning  Product.
1 Accomplishments. 2 Overview of Accomplishments  Sustaining the Production Earth System Grid Serving the current needs of the climate modeling community.
1 Overall Architectural Design of the Earth System Grid.
LLNL’s Data Center and Interoperable Services 5 th Annual ESGF Face-to-Face Conference ESGF 2015 Monterey, CA, USA Dean N. Williams, Tony Hoang, Cameron.
1 Earth System Grid Center For Enabling Technologies (ESG-CET) Introduction and Overview Dean N. Williams, Don E. Middleton, Ian T. Foster, and David E.
Earth System Curator and Model Metadata Discovery and Display for CMIP5 Sylvia Murphy and Cecelia Deluca (NOAA/CIRES) Hannah Wilcox (NCAR/CISL) Metafor.
1 Summary. 2 ESG-CET Purpose and Objectives Purpose  Provide climate researchers worldwide with access to data, information, models, analysis tools,
Evolving Scientific Data Workflow CAS 2011 Pamela Gillman
1 Earth System Grid Center for Enabling Technologies (ESG-CET) Overview ESG-CET Team Climate change is not only a scientific challenge of the first order.
2/22/2001Greenbook 2001/OASCR1 Greenbook/OASCR Activities Focus on technology to enable SCIENCE to be conducted, i.e. Software tools Software libraries.
Fire Emissions Network Sept. 4, 2002 A white paper for the development of a NSF Digital Government Program proposal Stefan Falke Washington University.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
SCD User Briefing The Community Data Portal and the Earth System Grid Don Middleton with presentation material developed by Luca Cinquini, Mary Haley,
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Data Requirements for Climate and Carbon Research John Drake, Climate Dynamics Group Computer.
Joseph Antony, Andrew Howard, Jason Andrade, Ben Evans, Claire Trenham, Jingbo Wang Production Petascale Climate Data Replication at.
NASA Earth Exchange (NEX) A collaborative supercomputing environment for global change science Earth Science Division/NASA Advanced Supercomputing (NAS)
A. Sim, CRD, L B N L 1 Production Data Management Workshop, Mar. 3, 2009 BeStMan and Xrootd Alex Sim Scientific Data Management Research Group Computational.
Supporting Advanced Scientific Computing Research Basic Energy Sciences Biological and Environmental Research Fusion Energy Sciences High Energy Physics.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
VisIt Project Overview
Clouds , Grids and Clusters
The Earth System Grid: A Visualisation Solution
DOE Facilities - Drivers for Science: Experimental and Simulation Data
SDM workshop Strawman report History and Progress and Goal.
“From a user standpoint, users have their own trial and error way of locating the fastest service. It seems when the network is slow, we simply let the.
Data Requirements for Climate and Carbon Research
Data Management Components for a Research Data Archive
Digital Object Management for ENES: Challenges and Opportunities
Presentation transcript:

A. Sim, CRD, L B N L 1 ANI and Magellan Launch, Nov. 18, 2009 Climate 100: Scaling the Earth System Grid to 100Gbps Networks Alex Sim, CRD, LBNL Dean N. Williams, PCMDI, LLNL

A. Sim, CRD, L B N L 2 ANI and Magellan Launch, Nov. 18, Gbps Network and Climate Community High performance network for Earth System GridHigh performance network for Earth System Grid Data replication around the world “Core dataset” – The Coupled Model Intercomparison Project, Phase 5 (CMIP-5) used for the Intergovernmental Panel on Climate Change (IPCC) Fifth Assessment Report (AR5) is estimated to 600TB-1.2PB in 2010 Climate model data is projected to exceed hundreds of Exabytes by 2020 (BES Science Network Requirements Workshop, 2007) Climate analysis requires dataset Large amount of data movement and management is needed Climate100Climate100 Research and integration effort for 100Gbps network from data intensive applications point of view Data analysis

A. Sim, CRD, L B N L 3 ANI and Magellan Launch, Nov. 18, 2009 Earth System Grid ESG-CET Scientific Production Portals: NCAR: LLNL: ORNL: ESG-CET Executive Committee Dean N. Williams, Ian Foster, Don Middleton

A. Sim, CRD, L B N L 4 ANI and Magellan Launch, Nov. 18, 2009 ESG current statistics NCAR ESG portalNCAR ESG portal 237 TB of data at four locations (NCAR, LBNL, ORNL, LANL) 965,551 files Includes the past 7 years of joint DOE/NSF climate modeling experiments LLNL CMIP-3 (IPCC AR4) ESG portalLLNL CMIP-3 (IPCC AR4) ESG portal 35 TB of data at one location 83,337 files, model data from 13 countries Generated by a modeling campaign coordinated by the Intergovernmental Panel on Climate Change (IPCC) Over 530 scientific peer-review publications Geographic distribution of the users that downloaded data from ESG web portalsGeographic distribution of the users that downloaded data from ESG web portals Over 2,700 sites 120 countries 16,000 users Over 850 TB downloaded Courtesy: Gary Strand - NCAR

A. Sim, CRD, L B N L 5 ANI and Magellan Launch, Nov. 18, 2009 Network Challenges in ESG Berkeley Storage Manager (BeStMan) and replication management to provide load balancing across a large number of Data NodesBerkeley Storage Manager (BeStMan) and replication management to provide load balancing across a large number of Data Nodes ESG plans to move vast amounts of data to and from sitesESG plans to move vast amounts of data to and from sites perfSONAR, NetLogger and SciDAC CEDPS troubleshooting technique to help monitor network performanceperfSONAR, NetLogger and SciDAC CEDPS troubleshooting technique to help monitor network performance ESG plans to utilize current ESnet’s 10 Gbps network for its federated architecture, and move onto 100Gbps when readyESG plans to utilize current ESnet’s 10 Gbps network for its federated architecture, and move onto 100Gbps when ready

A. Sim, CRD, L B N L 6 ANI and Magellan Launch, Nov. 18, 2009 As the results... Enable new capabilities for analysis of data and visual explorationEnable new capabilities for analysis of data and visual exploration Visualization of uncertainty and ensemble data Exploration of climate modeling data with ViSUS (LLNL VACET team) and CDAT (LLNL PCMDI) Help scientists understand long-term climate impactHelp scientists understand long-term climate impact

A. Sim, CRD, L B N L 7 ANI and Magellan Launch, Nov. 18, 2009 Timeline - Phase 1 10 Gbps Virtual Circuits, LLNL and LBNL connection10 Gbps Virtual Circuits, LLNL and LBNL connection ESG Data Nodes at LBNL and LLNLESG Data Nodes at LBNL and LLNL Distributed file systems for a large disk space accessibility Enhanced BeStMan and Bulk Data Mover components Multiple transfer servers with GridFTP Multiple ESnet 10 Gbps connections GoalGoal Move beyond the current machine hardware capability of 10 Gbps with multiple 10 Gbps connections Prepare extension to higher data transfer performance with the coming ESnet 100 Gbps network and multiple distributed storage systems

A. Sim, CRD, L B N L 8 ANI and Magellan Launch, Nov. 18, 2009 Timeline - Phase 2 Phase 1 + ORNL and ANL connectionPhase 1 + ORNL and ANL connection ESnet plan on deployment of 100 Gbps for NERSC, ANL, and ORNL Phase 1 case extends to ANL and ORNL GoalGoal Make use of the available 100 Gbps network capability between ESG Data Node sites in ANL, LBNL, and ORNL, with the designed data movement framework Continue work on data transfers including LLNL

A. Sim, CRD, L B N L 9 ANI and Magellan Launch, Nov. 18, 2009 Timeline - Phase 3 Phase 2 + NCAR connectionPhase 2 + NCAR connection 100 Gbps networking would be commercially available ESnet will be in the processes of integrating the 100 Gbps ANI network into ESnet’s production SDN network. We presume that NCAR will request funding at this time to upgrade their 10 Gbps connection. Phase 2 case extends to NCAR possibly GoalGoal Extend high data transfer performance with 100 Gbps to broader ESG communities Prepare these research activities with ESG production activities

A. Sim, CRD, L B N L 10 ANI and Magellan Launch, Nov. 18, 2009Summary Climate100Climate100 Integrating the latest networks, computer systems, datasets, and software technologies Enable at-scale experimentation and analysis of data transport and network use for peta- and exascale data Will improve understanding of network technologies Help climate community to transition to a 100 Gbps network for production and research Will prepare application communities ready to handle 100+ PB data and 100+ Gbps networks More informationMore information ContactContact Alex Sim Dean N. Williams