Community infrastructure for building and coupling high performance climate, weather, and coastal models Cecelia DeLuca NOAA / CIRES University of Colorado,

Slides:



Advertisements
Similar presentations
© GEO Secretariat Agenda Item 3. GEO UPDATE. © GEO Secretariat Membership 67 members and 43 Participating Organisations – New Members:Latvia, Moldova,
Advertisements

Expanding Regridding Capabilities of the Earth System Modeling Framework Andrew Scholbrock University of Colorado – Boulder Robert Oehmke NOAA/CIRES 1.
NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Emergence of the Earth System Modeling Framework NSIPP Seasonal Forecast.
Earth System Curator Spanning the Gap Between Models and Datasets.
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech.
State of Indiana Business One Stop (BOS) Program Roadmap Updated June 6, 2013 RFI ATTACHMENT D.
HP Quality Center Overview.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Integrated Frameworks for Earth and Space Weather Simulation Timothy Killeen and Cecelia DeLuca National Center for Atmospheric Research, Boulder, Colorado.
Building the LTER Network Information System. NIS History, Then and Now YearMilestone 1993 – 1996NIS vision formed by Information Managers (IMs) and LTER.
Annual SERC Research Review - Student Presentation, October 5-6, Extending Model Based System Engineering to Utilize 3D Virtual Environments Peter.
A Quick Tour of the NOAA Environmental Software Infrastructure and Interoperability Group Cecelia DeLuca and the ESMF team ESRL Directorate Seminar June.
Coupling Climate and Hydrological Models Interoperability Through Web Services.
Coupling Climate and Hydrological Models Interoperability Through Web Services Kathy Saint/SGI – NESII Jon Goodall/University of South Carolina Richard.
Chapter 3 – Agile Software Development 1Chapter 3 Agile software development.
The Earth System CoG Collaboration Environment Sylvia Murphy and Cecelia DeLuca (NOAA/CIRES), and Luca Cinquini (NASA/JPL) AGU Ocean Sciences February.
Metadata Creation with the Earth System Modeling Framework Ryan O’Kuinghttons – NESII/CIRES/NOAA Kathy Saint – NESII/CSG July 22, 2014.
CCSM Software Engineering Coordination Plan Tony Craig SEWG Meeting Feb 14-15, 2002 NCAR.
Fast Parallel Grid Remapping for Unstructured and Structured Grids Robert Oehmke NOAA Cooperative Institute for Research in Environmental Sciences University.
ESMF Town Hall Meeting AGU Fall Meeting 2010 San Francisco Gerhard Theurich, Fei Liu, Peggy Li, Cecelia DeLuca NOAA/CIRES December 15, 2010
An Introduction to Software Architecture
Metadata for the Coupled Ocean/Atmosphere Mesoscale Prediction System (COAMPS) using the Earth System Modeling Framework (ESMF) Peter Bosler University.
NE II NOAA Environmental Software Infrastructure and Interoperability Program Cecelia DeLuca Sylvia Murphy V. Balaji GO-ESSP August 13, 2009 Germany NE.
Cecelia DeLuca, Don Stark, Chris Hill Arctic System Model Workshop May 20, 2008 Earth System Modeling Framework.
ESMF Development Status and Plans ESMF 4 th Community Meeting Cecelia DeLuca July 21, 2005 Climate Data Assimilation Weather.
Update on ESMF, Earth System Curator, and Earth System CoG Cecelia DeLuca and the ESMF team CCSM Software Engineering Working Group June 23, 2011.
DOE BER Climate Modeling PI Meeting, Potomac, Maryland, May 12-14, 2014 Funding for this study was provided by the US Department of Energy, BER Program.
Computational Design of the CCSM Next Generation Coupler Tom Bettge Tony Craig Brian Kauffman National Center for Atmospheric Research Boulder, Colorado.
Initial Results from the Integration of Earth and Space Frameworks Cecelia DeLuca/NCAR, Alan Sussman/University of Maryland, Gabor Toth/University of Michigan.
High Level Architecture Overview and Rules Thanks to: Dr. Judith Dahmann, and others from: Defense Modeling and Simulation Office phone: (703)
Coupling Climate and Hydrological Models Interoperability Through Web Services.
The use of modeling frameworks to facilitate interoperability Cecelia DeLuca/NCAR (ESMF) Bill Putman/NASA GSFC (MAPL) David Neckels/NCAR.
CESM/ESMF Progress Report Mariana Vertenstein NCAR Earth System Laboratory CESM Software Engineering Group (CSEG) NCAR is sponsored by the National Science.
Earth System Modeling Framework Status Cecelia DeLuca NOAA Cooperative Institute for Research in Environmental Sciences University of Colorado, Boulder.
Components, Coupling and Concurrency in the Earth System Modeling Framework N. Collins/NCAR, C. DeLuca/NCAR, V. Balaji/GFDL, G. Theurich/SGI, A. da Silva/GSFC,
Opportunities for Research in the Dynamics of Water Processes in the Environment at NSF Pam Stephens Directorate of Geosciences, NSF Directorate of Geosciences,
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Earth System Modeling Framework Workshop on “Coupling Technologies for Earth System Modelling : Today and Tomorrow” CERFACS, Toulouse (France) – Dec 15.
Portable Infrastructure for the Metafor Metadata System Charlotte Pascoe 1, Gerry Devine 2 1 NCAS-BADC, 2 NCAS-CMS University of Reading PIMMS provides.
Earth System Modeling Framework Python Interface (ESMP) October 2011 Ryan O’Kuinghttons Robert Oehmke Cecelia DeLuca.
Strategic Plan Implementation Cecelia DeLuca/NCAR (ESMF) December 17, 2008 ESMF Board/Interagency Meeting.
Slides for NUOPC ESPC NAEFS ESMF. A NOAA, Navy, Air Force strategic partnership to improve the Nation’s weather forecast capability Vision – a national.
ESMF Strategic Discussion Cecelia DeLuca NOAA ESRL/University of Colorado ESMF Executive Board/Interagency Meeting June 12, 2014.
ESMF Regridding Update Robert Oehmke, Peggy Li, Ryan O’Kuinghttons, Mat Rothstein, Joseph Jacob NOAA Cooperative Institute for Research in Environmental.
Requirements Engineering Requirements Engineering in Agile Methods Lecture-28.
Foundations of Information Systems in Business. System ® System  A system is an interrelated set of business procedures used within one business unit.
Earth System Curator and Model Metadata Discovery and Display for CMIP5 Sylvia Murphy and Cecelia Deluca (NOAA/CIRES) Hannah Wilcox (NCAR/CISL) Metafor.
Community-Driven Development - The Change Review Board Robert Ferraro ESMF Community Meeting May 23, 2006 Climate Data Assimilation.
Welcome to the PRECIS training workshop
ESMF,WRF and ROMS. Purposes Not a tutorial Not a tutorial Educational and conceptual Educational and conceptual Relation to our work Relation to our work.
Building Community and Capability through Common Infrastructure: ESMF and the Earth System Curator Cecelia DeLuca MAP Meeting College.
Enhancements for Hydrological Modeling in ESMF Cecelia DeLuca/NCAR (ESMF) December 19, 2008 AGU Fall Meeting.
ESMF Change Review Board Robert Ferraro ESMF Board Meeting Dec 17, 2008 Climate Data Assimilation Weather.
Emergence of a Common Modeling Architecture for Earth System Science American Geophysical Union December 13, 2010 Cecelia DeLuca NOAA/CIRES.
ESMF and the future of end-to-end modeling Sylvia Murphy National Center for Atmospheric Research
State of ESMF: The NUOPC Layer Gerhard Theurich NRL/SAIC ESMF Executive Board / Interagency Working Group Meeting June 12, 2014.
The NOAA Environmental Modeling System at NCEP Mark Iredell and the NEMS group NOAA/NWS/NCEP Environmental Modeling Center June 12, 2014.
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech 5 th GO-ESSP Community Meeting.
Options for Framework Connectivity Cecelia DeLuca NOAA Environmental Software Infrastructure and Interoperability February.
A Quick Tour of the NOAA Environmental Software Infrastructure and Interoperability Group Cecelia DeLuca Dr. Robert Detrick visit March 28, 2012
Process 4 Hours.
GMAO Seasonal Forecast
ESMF Governance Cecelia DeLuca NOAA CIRES / NESII April 7, 2017
A Quick Tour of the NOAA Environmental Software Infrastructure and Interoperability Group Cecelia DeLuca and the ESMF team ESRL Directorate Seminar June.
Mariana Vertenstein (CGD)
Joint GEOS-Chem and NCAR Modeling Workshop:
Mariana Vertenstein CCSM Software Engineering Group NCAR
Metadata Development in the Earth System Curator
A brief introduction to NEMS
Presentation transcript:

Community infrastructure for building and coupling high performance climate, weather, and coastal models Cecelia DeLuca NOAA / CIRES University of Colorado, Boulder ESMF website: Building Community Codes for Effective Scientific Research on HPC Platforms September 6-7, 2012

Outline Vision and Context Technical Overview User Community Values, Processes and Governance Future

Vision Earth system models that can be built, assembled and reconfigured easily, using shared toolkits and standard interfaces. A growing pool of Earth system modeling components that, through their broad distribution and ability to interoperate, promotes the rapid transfer of knowledge. Earth system modelers who are able to work more productively, focusing on science rather than technical details. An Earth system modeling community with cost-effective, shared infrastructure development and many new opportunities for scientific collaboration. Accelerated scientific discovery and improved predictive capability through the social and technical influence of ESMF.

Computational Context Teams of specialists, often at different sites, contribute scientific or computational components to an overall modeling system Components may be at multiple levels: individual physical processes (e.g. atmospheric chemistry), physical realms (e.g. atmosphere, ocean), and members of same or multi-model ensembles (e.g. “MIP” experiments) Components contributed from multiple teams must be coupled together, often requiring transformations of data in the process(e.g. grid remapping and interpolation, merging, redistribution) Transformations are most frequently 2D data, but 3D is becoming more common There is an increasing need for cross-disciplinary and inter-framework coupling for climate impacts Running on tens of thousands of processors is fairly routine; utilizing hundreds of thousands of processors or GPUs is less common Modelers will tolerate virtually no framework overhead and seek fault tolerance and bit reproducibility Provenance collection is increasingly important for climate simulations

Architecture Low Level Utilities Fields and Grids Layer Model Layer Components Layer Gridded Components Coupler Components ESMF Infrastructure User Code ESMF Superstructure MPI, NetCDF, … External Libraries The Earth System Modeling Framework (ESMF) provides a component architecture or superstructure for assembling geophysical components into applications. ESMF provides an infrastructure that modelers use to – Generate and apply interpolation weights – Handle metadata, time management, I/O and communications, and other common functions The ESMF distribution does not include scientific models

Components ESMF is based on the idea of components – sections of code that are wrapped in standard interfaces Components can be arranged hierarchically, helping to organize the structure of complex models Different modeling groups may create different kinds or levels of components Some of the ESMF components in the GEOS-5 atmospheric GCM

Applications of information layer ESMF as an Information Layer Native model data structures Standard data structures Standard metadata Parallel generation and application of interpolation weights Run-time compliance checking of metadata and time behavior Fast parallel I/O Redistribution and other parallel communications Automated documentation of models and simulations Ability to run components in workflows and as web services Field Grid Clock Component Attributes: CF conventions, ISO standards, METAFOR Common Information Model Structured model information stored in ESMF wrappers User data is referenced or copied into ESMF structures modules fields grids timekeeping ESMF data structures

Standard Interfaces All ESMF components have the same three standard methods: – Initialize – Run – Finalize Each standard method has the same simple interface: call ESMF_GridCompRun (myComp, importState, exportState, clock, …) Where: myComp points to the component importState is a structure containing input fields exportState is a structure containing output fields clock contains timestepping information Steps to adopting ESMF Divide the application into components (without ESMF) Copy or reference component input and output data into ESMF data structures Register components with ESMF Set up ESMF couplers for data exchange Interfaces are wrappers and can often be set up in a non-intrusive way

Data Structures ESMF represents data as Fields that can be built on four discretization types: – Logically rectangular grids, which may be connected at the edges – Unstructured meshes – Observational data streams – Exchange grids – grids that are the union of the grids of components being coupled, and reference the data on the original grids ESMF can transform data among these representations, using an underlying finite element unstructured mesh engine

Component Overhead Representation of the overhead for ESMF wrapped native CCSM4 component For this example, ESMF wrapping required NO code changes to scientific modules No significant performance overhead (< 3% is typical) Few code changes for codes that are modular Platform: IBM Power 575, bluefire, at NCAR Model: Community Climate System Model (CCSM) Versions: CCSM_4_0_0_beta42 and ESMF_5_0_0_beta_snapshot_01 Resolution: 1.25 degree x 0.9 degree global grid with 17 vertical levels for both the atmospheric and land model, i.e. 288x192x17 grid. The data resolution for the ocean model is 320x384x60.

ESMF Regridding ESMF offers extremely fast parallel regridding with many options (regional, global; bilinear, higher order, first order conservative methods; logically rectangular grids and unstructured meshes; pole options; 2D or 3D; invocation as offline application or during model run) Summary of features: blic/ESMF_5_3_0/esmf_5_3_0_regridding_status.htm HOMME Cubed Sphere Grid with Pentagons Courtesy Mark Taylor of Sandia FIM Unstructured Grid Regional Grid ESMF supported grids IMPACT: “use of the parallel ESMF offline regridding capability has reduced the time it takes to create CLM surface datasets from hours to minutes” - Mariana Vertenstein, NCAR IMPACT: “use of the parallel ESMF offline regridding capability has reduced the time it takes to create CLM surface datasets from hours to minutes” - Mariana Vertenstein, NCAR

ESMF Offline RegridWeightGen Application: – Separate application that is built as part of ESMF, can be used independently – Application generates a netCDF weight file from two netCDF grid files – Supports SCRIP, GRIDSPEC, UGRID, and custom ESMF unstructured format mpirun –np 32 ESMF_RegridWeightGen –s src_grid.nc –d dst_grid.nc –m bilinear –w weights.nc Regridding during model execution: – ESMF library subroutine calls which do interpolation during model run – Can get weights or feed directly into ESMF parallel sparse matrix multiply – Can be used without ESMF components call ESMF_FieldRegridStore(srcField=src, dstField=dst, regridMethod=ESMF_REGRID_METHOD_BILINEAR, routehandle=rh) call ESMF_FieldRegrid(srcField=src, dstField=dst, routehandle=rh) Regridding Modes

9/7/2015 ESMP – Python Interface ● Flexible way to use ESMF parallel regridding ● Separate download: ● Requirements: – Python – Numpy – Ctypes ● Limited platform support: Linux/Darwin, GCC(g++/gfortran), OpenMPI ● Data type: ESMP_Field ● Grid types: – Single-tile 2D logically rectangular type: ESMP_Grid – Unstructured type: ESMP_Mesh ● Support for all ESMF interpolation options

ESMF Web Services for Climate Impacts GOAL: Develop a two-way coupled, distributed, service-based modeling system comprised of an atmospheric climate model and a hydrological model, utilizing standard component interfaces from each domain. IDEA: Bring the climate model to local applications Two way technical coupling completed during 2012 using CAM (with active land) and SWAT (Soil Water Assessment Tool) Utilized switch in CAM’s ESMF component wrapper that enables web service interface Coupled system is run using the OpenMI configuration editor, a web service driver developed in the hydrological community Next steps: using WRF within the CESM framework and an updated hydrological model

Summary of Features Components with multiple coupling and execution modes for flexibility, including a web service execution mode Fast parallel remapping with many features Core methods are scalable to tens of thousands of processors Supports hybrid (threaded/distributed) programming for optimal performance on many computer architectures; works with codes that use OpenMP and OpenACC Time management utility with many calendars, forward/reverse time operations, alarms, and other features Metadata utility that enables comprehensive, standard metadata to be written out in standard formats Runs on 30+ platform/compiler combinations, exhaustive nightly regression test suite (4500+ tests) and documentation Couples Fortran or C-based model components Open source license

Major Users ESMF Components: NOAA National Weather Service operational weather models (GFS, Global Ensemble, NEMS) NASA atmospheric general circulation model GEOS-5 Navy and related atmospheric, ocean and coastal research and operational models – COAMPS, NOGAPS, HYCOM, WaveWatch, others Hydrological modelers at Delft hydraulics, space weather modelers at NCAR and NOAA ESMF Regridding and Python Libraries NCAR/DOE Community Earth System Model (CESM) Analysis and visualization packages: NCAR Command Language, Ultrascale Visualization - Climate Data Analysis Tools (UV-CDAT), PyFerret users Community Surface Dynamics Modeling System

Usage Metrics Updated ESMF component listing at: Includes 85 components with ESMF interfaces, 12 coupled cross- agency modeling systems in space weather, climate, weather, hydrology, and coastal prediction, for operational and research use About 4500 registered downloads

Values and Principles Community driven development and community ownership Openness of project processes, management, code and information Correctness Commitment to a globally distributed and diverse development and customer base Simplicity Efficiency User engagement Environmental stewardship Web link for detail:

Agile Development vs Community Values In general, Agile processes promote to community development. However, there are areas of potential conflict: Agile emphasis on co-location: “most agile methodologies or approaches assume that the team is located in a single team room” (from Miller, Dist. Agile Development at Microsoft, 2008) vs ESMF emphasis on distributed co-development Agile emphasis on single product owner vs ESMF emphasis on multi-party ownership

Making Distributed Co-Development Work Hinges on asynchronous, all-to-all communication patterns: everybody must have information Archived list where all development correspondence gets cc’d Minutes for all telecons Web browsable repositories (main and contributions), mail summary on check-ins Daily, publicly archived test results Monthly archived metrics Public archived trackers (bugs, feature requests, support requests, etc.) Discouraged: IMing, one-to-one correspondence or calls – the medium matters

Change Review Board CRB established as a vehicle for shared ownership through user task prioritization and release content decisions Consists of technical leads from key user communities Not led by the development team! Sets the schedule and expectations for future functionality enhancements in ESMF internal and public distributions – Based on broad user community and stakeholder input – Constrained by available developer resources – Updated quarterly to reflect current realities CRB reviews releases after delivery for adherence to release plan

Governance Highlights Management of ESMF required governance that recognized social and cultural factors as well as technical factors Main practical objectives of governance: Enabling stakeholders to fight and criticize in a civilized, contained, constructive way Enabling people to make priority decisions based on resource realities Observations: Sometimes just getting everyone equally dissatisfied and ready to move on is a victory Thorough, informed criticism is the most useful input a project can get Governance changes and evolves over the life span of a project

Governance Functions Prioritize development tasks in a manner acceptable to major stakeholders and the broader community, and define development schedules based on realistic assessments of resource constraints (CRB) Deliver a product that meets the needs of critical applications, including adequate and correct functionality, satisfactory performance and memory use,... (Core) Support users via prompt responses to questions, training classes, minimal code changes for adoption, thorough documentation,... (Core) Encourage community participation in design and implementation decisions frequently throughout the development cycle (JST) Leverage contributions of software from the community when possible (JST) Create frank and constructive mechanisms for feedback (Adv. Board) Enable stakeholders to modify the organizational structure as required (Exec. Board) Coordinate and communicate at many levels in order to create a knowledgeable and supportive network that includes developers, technical management, institutional management, and program management (IAWG and other bodies)

Governance Executive Board Strategic Direction Organizational Changes Board Appointments Interagency Working Group Stakeholder Liaison Programmatic Assessment & Feedback Advisory Board External Projects Coordination General Guidance & Evaluation Functionality Change Requests Joint Specification Team Requirements Definition Design and Code Reviews External Code Contributions Implementation Schedule Resource Constraints Collaborative Design Beta Testing Working Project Executive Management Reporting weekly Core Development Team Project Management Software Development Testing & Maintenance Distribution & User Support daily annually Change Review Board Development Priorities Release Review & Approval quarterly

Evolution Phase 1: NASA’s Earth Science Technology Office ran a solicitation to develop an Earth System Modeling Framework (ESMF). A multi-agency collaboration (NASA/NSF/DOE/NOAA) won the award. The core development team was located at NCAR. A prototype ESMF software package (version 2r) demonstrated feasibility. Phase 2: New sponsors included Department of Defense and NOAA. A multi-agency governance plan including the CRB was created: Many new applications and requirements were brought into the project, motivating a complete redesign of framework data structures (version 3r). Phase 3: (and beyond) The core development team moved to NOAA/CIRES for closer alignment with federal models. Basic framework development completed with version 5r (ports, bugs, feature requests, user support etc. still require resources). The focus is on increasing adoption and creating a community of interoperable codes.

Technical Evolution: National Unified Operational Prediction Capability (NUOPC) ESMF allows for many levels of components, types of components, and types of connections In order to achieve greater interoperability, usage and content conventions and component templates are needed A tri-agency collaboration (NOAA, Navy, Air Force) is building a “NUOPC Layer” that constrains how ESMF is used, and introduces metadata and other content standards, along with inheritable templates for different usage scenarios A production version of the NUOPC Layer is scheduled for delivery at the end of 2012

CoG Collaboration Environment Right side: Services Central navigation: Templated content Central Section: Freeform content Left side: Auto generated navigation of project freeform content The CoG environment exposes and collates the information needed for distributed, multi-project development, including project repositories, trackers, and governance processes. It does this in an environment that’s linked to data search, metadata, and visualization services and set up to enable component comparisons.

DCMIP on CoG Atmospheric Dynamical Core Model Intercomparison Project Planned MIPs: 2013 Downscaling 2014 Atm-surface hydrology 2014 Frameworks

9/7/2015 Future Governance and process largely follow established patterns Active development ● Python API redesign and addition of features ● Preparation for NUOPC Layer major release ● Web services for climate impacts ● Regridding options – higher order conservative, filling in gaps ● Advanced fault-tolerance ● Performance optimizations and GPUs ● CoG collaboration environment ESMF and related development is supported by NASA, NOAA, NSF, and the Department of Defense.