NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH www.esmf.ucar.edu Architecture of the Earth System Modeling Framework GMAO Seasonal.

Slides:



Advertisements
Similar presentations
DOE Global Modeling Strategic Goals Anjuli Bamzai Program Manager Climate Change Prediction Program DOE/OBER/Climate Change Res Div
Advertisements

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Arlindo da Silva, NASA/GSFC/GMAO The Earth System Modeling Framework.
NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Emergence of the Earth System Modeling Framework NSIPP Seasonal Forecast.
Earth System Curator Spanning the Gap Between Models and Datasets.
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech.
1 Earth System Modeling Framework ESMF and the Transformation of Earth System Modeling Sylvia Murphy
Integrated Frameworks for Earth and Space Weather Simulation Timothy Killeen and Cecelia DeLuca National Center for Atmospheric Research, Boulder, Colorado.
Introduction to the Earth System Modeling Framework Don Stark Gerhard Theurich Shujia Zhou.
1 ESMF in Production at NCEP Mark Iredell Chief NCEP/EMC Global Climate and Weather Modeling Branch May 23, 2006 NCEP: “where America’s climate, weather,
Introduction to the Earth System Modeling Framework Nancy Collins July 22, 2005 Climate Data Assimilation Weather.
Mesoscale & Microscale Meteorological Division / NCAR ESMF and the Weather Research and Forecast Model John Michalakes, Thomas Henderson Mesoscale and.
1 NCEP Mark Iredell Chief NCEP/EMC Global Climate and Weather Modeling Branch May 23, 2006 NCEP: “where America’s climate, weather, and ocean services.
The NASA Modeling, Analysis and Prediction (MAP) Modeling Environment Don Anderson NASA HQ Sience Mission Directorate Earth-Sun Division Manager, Modeling,
The Earth System Modeling Framework and the Earth System Curator Cecelia DeLuca and the ESMF Joint Specification Team.
Earth System Modeling Framework Capabilities Cecelia DeLuca SCD Users Forum May 18, GMAO.
Components and Concurrency in ESMF Nancy Collins Community Meeting July 21, GMAO Seasonal.
NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT Adoption and field tests of M.I.T General Circulation Model (MITgcm) with ESMF Chris Hill ESMF.
NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH First Field Tests of ESMF GMAO Seasonal Forecast NCAR/LANL CCSM NCEP.
1 NOAA’s Environmental Modeling Plan Stephen Lord Ants Leetmaa November 2004.
Metadata Creation with the Earth System Modeling Framework Ryan O’Kuinghttons – NESII/CIRES/NOAA Kathy Saint – NESII/CSG July 22, 2014.
Project Overview GMAO Seasonal Forecast NCAR/LANL CCSM NCEP Forecast GFDL FMS Suite MITgcm NASA GMAO Analysis Climate Data Assimilation.
Science Computing BranchGoddard Space Flight Center Code 930 Support for GMI Tom Clune (NASA) Bigyani Das (CSC) Jae-Hoon Kim (CSC)
What is ESMF and what does it mean to adopt it? 3 rd ESMF Community Meeting Cecelia DeLuca Nancy Collins
NE II NOAA Environmental Software Infrastructure and Interoperability Program Cecelia DeLuca Sylvia Murphy V. Balaji GO-ESSP August 13, 2009 Germany NE.
Coordination of Common Modeling Infrastructure Cecelia DeLuca WGCM/WMP Meeting, Exeter, UK Oct 6, 2005 Climate Data Assimilation Weather.
ESMF Development Status and Plans ESMF 4 th Community Meeting Cecelia DeLuca July 21, 2005 Climate Data Assimilation Weather.
Computational Design of the CCSM Next Generation Coupler Tom Bettge Tony Craig Brian Kauffman National Center for Atmospheric Research Boulder, Colorado.
Initial Results from the Integration of Earth and Space Frameworks Cecelia DeLuca/NCAR, Alan Sussman/University of Maryland, Gabor Toth/University of Michigan.
ESMF Application Status GMAO Seasonal Forecast NCAR/LANL CCSM NCEP Forecast GFDL FMS Suite MITgcm NCEP/GMAO Analysis Climate Data Assimilation.
Introduction to the Earth System Modeling Framework V. Balaji, GFDL Cecelia DeLuca, Chris Hill, MIT
The use of modeling frameworks to facilitate interoperability Cecelia DeLuca/NCAR (ESMF) Bill Putman/NASA GSFC (MAPL) David Neckels/NCAR.
CESM/ESMF Progress Report Mariana Vertenstein NCAR Earth System Laboratory CESM Software Engineering Group (CSEG) NCAR is sponsored by the National Science.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Earth System Modeling Framework Status Cecelia DeLuca NOAA Cooperative Institute for Research in Environmental Sciences University of Colorado, Boulder.
Introduction to the Earth System Modeling Framework International Workshop on Next Generation Climate Models for Advanced High Performance Computing Facilities.
ESMF/Curator Status Cecelia DeLuca CCSM Software Engineering Working Group Boulder, CO March 16, 2007 Climate Data Assimilaton Weather.
Components, Coupling and Concurrency in the Earth System Modeling Framework N. Collins/NCAR, C. DeLuca/NCAR, V. Balaji/GFDL, G. Theurich/SGI, A. da Silva/GSFC,
V. Balaji SGI/GFDL Princeton University First PRISM Project Meeting Toulouse, 22 May 2002 NASA/GSFC A High-Performance Framework for Earth.
NSF NCAR / NASA GSFC / DOE LANL ANL / NOAA NCEP GFDL / MIT / U MICH 8 April 2003 V. Balaji / GFDL, Cecelia DeLuca / NCAR, Chris Hill / MIT, Max Suarez,
A Portable Regional Weather and Climate Downscaling System Using GEOS-5, LIS-6, WRF, and the NASA Workflow Tool Eric M. Kemp 1,2 and W. M. Putman 1, J.
Earth System Modeling Framework Workshop on “Coupling Technologies for Earth System Modelling : Today and Tomorrow” CERFACS, Toulouse (France) – Dec 15.
CCSM Portability and Performance, Software Engineering Challenges, and Future Targets Tony Craig National Center for Atmospheric Research Boulder, Colorado,
Experimental El Niño/Southern Oscillation Predictions by the UCLA Atmospheric General Circulation Model (GCM) Coupled to the MIT and POP Oceanic GCMs using.
Strategic Plan Implementation Cecelia DeLuca/NCAR (ESMF) December 17, 2008 ESMF Board/Interagency Meeting.
Slides for NUOPC ESPC NAEFS ESMF. A NOAA, Navy, Air Force strategic partnership to improve the Nation’s weather forecast capability Vision – a national.
NSF NCAR / NASA GSFC / DOE LANL ANL / NOAA NCEP GFDL / MIT / U MICH C. DeLuca/NCAR, J. Anderson/NCAR, V. Balaji/GFDL, B. Boville/NCAR, N. Collins/NCAR,
NSF NCAR / NASA GSFC / DOE LANL ANL / NOAA NCEP GFDL / MIT / U MICH May 15, 2003 Nancy Collins, NCAR 2nd Community Meeting, Princeton, NJ Earth System.
NCEP ESMF GFS Global Spectral Forecast Model Weiyu Yang, Mike Young and Joe Sela ESMF Community Meeting MIT, Cambridge, MA July 21, 2005.
NSF NCAR / NASA GSFC / DOE LANL ANL / NOAA NCEP GFDL / MIT / U MICH 15 May 2003 Cecelia DeLuca / NCAR 2 nd ESMF Community Meeting Princeton, NJ NSIPP Seasonal.
NSF NCAR / NASA GSFC / DOE LANL ANL / NOAA NCEP GFDL / MIT / U MICH May 14, 2003 Nancy Collins, NCAR Components Workshop, Princeton, NJ Components in the.
ESMF Strategic Discussion Cecelia DeLuca NOAA ESRL/University of Colorado ESMF Executive Board/Interagency Meeting June 12, 2014.
CCSM Performance, Successes and Challenges Tony Craig NCAR RIST Meeting March 12-14, 2002 Boulder, Colorado, USA.
Extension of the ESMF for Space Weather Cecelia DeLuca SWW April 7, NSIPP Seasonal Forecast.
1 Accomplishments. 2 Overview of Accomplishments  Sustaining the Production Earth System Grid Serving the current needs of the climate modeling community.
Earth System Curator and Model Metadata Discovery and Display for CMIP5 Sylvia Murphy and Cecelia Deluca (NOAA/CIRES) Hannah Wilcox (NCAR/CISL) Metafor.
ESMF,WRF and ROMS. Purposes Not a tutorial Not a tutorial Educational and conceptual Educational and conceptual Relation to our work Relation to our work.
Building Community and Capability through Common Infrastructure: ESMF and the Earth System Curator Cecelia DeLuca MAP Meeting College.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Data Requirements for Climate and Carbon Research John Drake, Climate Dynamics Group Computer.
Climate Data Assimilation Weather Earth System Modeling Framework (ESMF) Community Meeting, DODNSFDOENASANOAA.
NSF NCAR / NASA GSFC / DOE LANL ANL / NOAA NCEP GFDL / MIT / U MICH 4 February 2003 Cecelia DeLuca, NCAR CCSM Software Engineering Working Group Mtg Boulder,
ESMF and the future of end-to-end modeling Sylvia Murphy National Center for Atmospheric Research
NSF NCAR / NASA GSFC / DOE LANL ANL / NOAA NCEP GFDL / MIT / U MICH C. DeLuca/NCAR, J. Anderson/NCAR, V. Balaji/GFDL, B. Boville/NCAR, N. Collins/NCAR,
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech 5 th GO-ESSP Community Meeting.
A Quick Tour of the NOAA Environmental Software Infrastructure and Interoperability Group Cecelia DeLuca Dr. Robert Detrick visit March 28, 2012
GMAO Seasonal Forecast
GMAO Seasonal Forecast
Software Practices for a Performance Portable Climate System Model
ESMF A High-Performance Framework for Earth Science Modeling and Data Assimilation V. Balaji SGI/GFDL Princeton University First PRISM.
Mariana Vertenstein CCSM Software Engineering Group NCAR
Metadata Development in the Earth System Curator
Presentation transcript:

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Architecture of the Earth System Modeling Framework GMAO Seasonal Forecast NCAR/LANL CCSM NCEP Forecast GFDL FMS Suite MITgcm NASA GMAO Analysis Climate Data Assimilation Weather Cecelia DeLuca GEM Snowmass, CO

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Outline Background and Motivation Applications Architecture Implementation Status Future Plans Conclusions

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Motivation for ESMF In climate research and NWP... increased emphasis on detailed representation of individual physical processes; requires many teams of specialists to contribute components to an overall modeling system In computing technology... increase in hardware and software complexity in high-performance computing, shift toward the use of scalable computing architectures In software … development of frameworks, such as the GFDL Flexible Modeling System (FMS) and Goddard Earth Modeling System (GEMS) that encourage software reuse and interoperability The ESMF is a focused community effort to tame the complexity of models and the computing environment. It leverages, unifies and extends existing software frameworks, creating new opportunities for scientific contribution and collaboration.

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Project Description GOALS: To increase software reuse, interoperability, ease of use and performance portability in climate, weather, and data assimilation applications PRODUCTS: Core framework: Software for coupling geophysical components and utilities for building components Applications: Deployment of the ESMF in 15 of the nation’s leading climate and weather models, assembly of 8 new science-motivated applications METRICS: RESOURCES and TIMELINE: $9.8M over 3 years, starting February 2002 ReuseInteroperabilityEase of AdoptionPerformance 15 applications use ESMF component coupling services and 3+ utilities 8 new applications comprised of never- before coupled components 2 codes adopt ESMF with < 2% lines of code changed, or within 120 FTE-hours No more than 10% overhead in time to solution, no degradation in scaling

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Outline Background and Motivation Applications Architecture Implementation Status Future Plans Conclusions

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Modeling Applications SOURCEAPPLICATION GFDLFMS B-grid atmosphere at N45L18 FMS spectral atmosphere at T63L18 FMS MOM4 ocean model at 2°x2°xL40 FMS HIM isopycnal C-language ocean model at 1/6°x1/6°L22 MITMITgcm coupled atmosphere/ocean at 2.8°x2.8°, atmosphere L5, ocean L15 MITgcm regional and global ocean at 15kmL30 GMAO/NSIPPNSIPP atmospheric GCM at 2°x2.5°xL34 coupled with NSIPP ocean GCM at 2/3°x1.25°L20 NCAR/LANLCCSM2 including CAM with Eulerian spectral dynamics and CLM at T42L26 coupled with POP ocean and data ice model at 1°x1°L40

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Data Assimilation Applications SOURCEAPPLICATION GMAO/DAOPSAS interface layer with 2O0K observations/day CAM with finite volume dynamics at 2°x2.5°L55, including CLM NCEPGlobal atmospheric spectral model at T170L42 SSI analysis system with 250K observations/day, 2 tracers WRF regional atmospheric model at 22km resolution CONUS forecast 345x569L50 GMAO/NSIPPODAS with OI analysis system at 1.25°x1.25°L20 resolution with ~10K observations/day MITMITgcm 2.8° century / millennium adjoint sensitivity

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Interoperability Demonstrations COUPLED CONFIGURATIONNEW SCIENCE ENABLED GFDL B-grid atm / MITgcm ocnIntroduction of global biogeochemistry into seasonal forecasts. GFDL MOM4 / NCEP forecastNew seasonal forecasting system. NSIPP ocean / LANL CICEExtension of seasonal prediction system to centennial timescales. NSIPP atm / DAO analysisAssimilated initial state for seasonal prediction system. DAO analysis / NCEP modelIntercomparison of systems for NASA/NOAA joint center for satellite data assimilation. NCAR fvCAM/ NCEP analysisIntercomparison of systems for NASA/NOAA joint center for satellite data assimilation. NCAR CAM Eul / MITgcm ocnComponent exchange for improved climate prediction capability. NCEP WRF / GFDL MOM4Development of hurricane prediction capability.

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Interoperability Experiments Completed NCAR Community Atmospheric Model (CAM) coupled to NCEP Spectral Statistical Interpolation (SSI) System, both set up as ESMF components Experiment utilizes same observational stream used operationally at NCEP 3 NCAR Community Atmospheric Model (CAM) coupled to MITgcm ocean Atmosphere, ocean, and coupler are set up as ESMF components Uses ESMF regridding tools 1 GFDL B-grid atmosphere coupled to MITgcm ocean Atmosphere, ocean, and coupler are set up as ESMF components Uses ESMF regridding tools 2 Temperature SSI import Temperature SSI export Temperature difference

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Outline Background and Motivation Applications Architecture Implementation Status Future Plans Conclusions

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Characteristics of Weather and Climate Simulation Mix of global transforms and local communications Load balancing for diurnal cycle, event (e.g. storm) tracking Applications typically require 10s of GFLOPS, 100s of PEs – but can go to 10s of TFLOPS, 1000s of PEs Required Unix/Linux platforms span laptop to Earth Simulator Multi-component applications: component hierarchies, ensembles, and exchanges Data and grid transformations between components Applications may be MPMD/SPMD, concurrent/sequential, combinations Parallelization via MPI, OpenMP, shmem, combinations Large applications (typically 100,000+ lines of source code) Platforms assim sea ice ocean land atm physicsdycore assim_atm atmland Seasonal Forecast coupler

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Architecture 1.ESMF provides an environment for assembling geophysical components into applications, with support for ensembles and hierarchies. 2.ESMF provides a toolkit that components use to i.increase interoperability ii.improve performance portability iii.abstract common services

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Hierarchies and Ensembles assim sea ice ocean land atm physicsdycore assim_atm atmland Seasonal Forecast coupler ESMF encourages applications to be assembled hierarchically and intuitively Coupling interfaces are standard at each layer Components can be used in different contexts assim_atm Ensemble Forecast assim_atm ESMF supports ensembles with multiple instances of components running sequentially (and soon, concurrently)

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Class Structure DELayout Communications State Data imported or exported Bundle Collection of fields GridComp Land, ocean, atm, … model F90 Superstructure Infrastructure Field Physical field, e.g. pressure Grid LogRect, Unstruct, etc. Data Communications C++ Regrid Computes interp weights CplComp Xfers between GridComps Utilities Machine, TimeMgr, LogErr, I/O, Config, Base etc. Array Hybrid F90/C++ arrays Route Stores comm paths DistGrid Grid decomposition PhysGrid Math description

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Data Classes Model data is contained in a hierarchy of multi-use classes. The user can reference a Fortran array to an Array or Field, or retrieve a Fortran array out of an Array or Field. Array – holds a cross-language Fortran / C++ array Field – holds an Array, an associated Grid, and metadata Bundle – collection of Fields on the same Grid State – contains States, Bundles, Fields, and/or Arrays Component – associated with an Import and Export State

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF DataMap Classes These classes give the user a systematic way of expressing interleaving and memory layout, also hierarchically (partially implemented) ArrayDataMap – relation of array to decomposition and grid, row / column major order, complex type interleave FieldDataMap – interleave of vector components BundleDataMap – interleave of Fields in a Bundle

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Standard Methods ESMF uses consistent names and behavior throughout the framework, for example Create / Destroy – create a new object, e.g. FieldCreate Set / Get – set or get a value, e.g. ArrayGetDataPtr Add / Get / Remove – add to, retrieve from, or remove from a list, e.g. StateAddField Print – to print debugging info, e.g. BundlePrint And so on

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Superstructure ESMF is a standard component architecture, similar to CCA but designed for the Earth modeling domain and for ease of use with Fortran codes Components and States are superstructure classes All couplers are the same derived type (ESMF_CplComp) and have a standard set of methods with prescribed interfaces All component models (atm, ocean, etc.) are the same derived type (ESMF_GridComp) and have a standard set of methods with prescribed interfaces Data is transferred between components using States. ESMF components can interoperate with CCA components – demonstrated at SC03

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Becoming an ESMF GridComp ESMF GridComps have 2 parts: one part user code, one part ESMF code The ESMF part is a GridComp derived type with standard methods including Initialize, Run, Finalize User code must also be divided into Initialize, Run, and Finalize methods – these can be multi-phase (e.g. Run phase 1, Run phase 2) User code interfaces must follow a standard form – that means copying or referencing data to ESMF State structures Users write a public SetServices method that contains ESMF SetEntryPoint calls - these associate a user method (“POPinit”) with a framework method (the Initialize call for a GridComp named “POP”) Now you’re an ESMF GridComp

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Infrastructure Data classes are Bundles, Fields, and Arrays Tools for expressing interleaved data stuctures Tools for resource allocation, decomposition, load balancing Toolkits for communications, time management, logging, IO

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Virtual Machine (VM) VM handles resource allocation Elements are Persistent Execution Threads or PETs PETs reflect the physical computer, and are one-to-one with Posix threads or MPI processes Parent Components assign PETs to child Components PETs will soon have option for computational and latency / bandwidth weights The VM communications layer does simpleMPI-like communications between PETs (alternative communication mechanisms are layered underneath)

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH DELayout Handles decomposition Elements are Decomposition Elements, or DEs (decomposition that’s 2 pieces in x by 4 pieces in y is a 2 by 4 DELayout) DELayout maps DEs to PETs, can have more than one DE per PET (for cache blocking, user-managed OpenMP threading) A DELayout can have a simple connectivity or more complex connectivity, with weights between DEs - users specify dimensions where greater connection speed is needed DEs will also have computation weights Array, Field, and Bundle methods perform inter-DE communications

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Communications Communication methods include Regrid, Redist, Halo, Gather, Scatter, etc. Communications methods are implemented at multiple levels, e.g. FieldHalo, ArrayHalo Communications hide underlying ability to switch between shared and distributed memory parallelism

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Load Balancing Three levels of graphs: Virtual Machine – machine-level PETs will have computational and connectivity weights DELayout – DE chunks have connectivity weights, will have computational weights Grid – grid cells will have computational and connectivity weights Intended to support standard load balancing packages (e.g. Parmetis) and user-developed load balancing schemes

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Outline Background and Motivation Applications Architecture Implementation Status Future Plans Conclusions

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Open Development Open source Currently ~800 unit tests, ~15 system tests are bundled with the ESMF distribution, can be run in non-exhaustive or exhaustive modes Results of nightly tests on many platforms are accessible on a Test and Validation webpage Test coverage, lines of code, requirements status are available on a Metrics webpage Exhaustive Reference Manual, including design and implementation notes, is available on a Downloads and Documentation webpage Development is designed to allow users clear visibility into the workings and status of the system, to allow users to perform their own diagnostics, and to encourage community ownership

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Port Status SGI IBM Compaq Linux (Intel, PGI, NAG, Absoft, Lahey)

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Outline Background and Motivation Applications Architecture Implementation Status Future Plans Conclusions

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Key Accomplishments Public delivery of prototype ESMF v1.0 in May 2003 Monthly ESMF internal releases with steadily increasing functionality Completion of first 3 coupling demonstrations using ESMF in March 2004 – NCAR CAM with NCEP SSI – NCAR CAM with MITgcm ocean – GFDL B-grid atmosphere with MITgcm ocean – All codes above running as ESMF components and coupled using the framework, codes available from Applications link on website – Other codes running as ESMF components: MOM4, GEOS-5 – Less than 2% lines of source code change Delivered ESMF v2.0 in June rd Community Meeting to be held on 15 July 2004 at NCAR

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Priorities Status Components, States, Bundles, Fields mature On-line parallel regridding (bilinear, 1 st order conservative) completed Other parallel methods, e.g. halo, redist, low-level comms implemented Comm methods overloaded for r4 and r8 Communications layer with uniform interface to shared / distributed memory, hooks for load balancing Near-term priorities Concurrent components – currently ESMF only runs in sequential mode More optimized grids (tripolar, spectral, cubed sphere) and more regridding methods (bicubic, 2 nd order conservative) from SCRIP Comms optimization and load balancing capability IO (based on WRF IO) Development schedule on-line, see Development link, SourceForge site tasks

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Outline Background and Motivation Applications Architecture Implementation Status Future Plans Conclusions

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Next Steps – Integration with data archives and metadata standardization efforts, anticipate collaboration with Earth System Grid (ESG) and European infrastructure project PRISM – Integration with scientific model intercomparison projects (MIPs), anticipate collaboration with the Program for Climate Model Diagnosis and Intercomparison (PCMDI), other community efforts – Integration with visualization and diagnostic tools for end-to-end modeling support, anticipate collaboration with the Earth Science Portal (ESP) – ESMF “vision” for the future articulated in multi-agency white paper on the Publications and Talks webpage

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Multi-Agency Follow-on 3 ESMF FTEs at NCAR slated to have ongoing funding through core NCAR funds NASA commitment to follow-on support, level TBD DoD and NSF proposals outstanding Working with other agencies to secure additional funds

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH Outline Background and Motivation Applications Architecture Implementation Status Future Plans Conclusions

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH ESMF Overall Clear, simple hierarchy of data classes Multi-use objects mean that the same object can carry information about decomposition, communications, IO, coupling Tools for multithreading, cache blocking, and load balancing are being integrated into the architecture Objects have consistent naming and behavior across the framework

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH The Benefits Standard interfaces to modeling components promote increased interoperability between centers, faster movement of modeling components from research to operations The ability to construct models hierarchically enables developers to add new modeling components more systematically and easily, facilitates development of complex coupled systems Multi-use objects mean that the same data structure can carry information about decomposition, communications, IO, coupling – this makes code smaller and simpler, and therefore less bug-prone and easier to maintain Shared utilities encourage efficient code development, higher quality tools, more robust codes

NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT | U MICH More Information ESMF website: Acknowledgements The ESMF is sponsored by the NASA Goddard Earth Science Technology Office.