Earth System Modeling Infrastructure Cecelia DeLuca and the ESMF Team NCAR/CISL CCSM Software Engineering Working Group June 17, 2009.

Slides:



Advertisements
Similar presentations
Expanding Regridding Capabilities of the Earth System Modeling Framework Andrew Scholbrock University of Colorado – Boulder Robert Oehmke NOAA/CIRES 1.
Advertisements

Earth System Curator Spanning the Gap Between Models and Datasets.
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech.
Development of an Integrated Earth System Climate Modeling Environment
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Earth System Modeling Framework ESMF and the Transformation of Earth System Modeling Sylvia Murphy
ARCS Data Analysis Software An overview of the ARCS software management plan Michael Aivazis California Institute of Technology ARCS Baseline Review March.
DCS Architecture Bob Krzaczek. Key Design Requirement Distilled from the DCS Mission statement and the results of the Conceptual Design Review (June 1999):
© , Michael Aivazis DANSE Software Issues Michael Aivazis California Institute of Technology DANSE Software Workshop September 3-8, 2003.
1 Geophysical Fluid Dynamics Laboratory Review June 30 - July 2, 2009.
January, 23, 2006 Ilkay Altintas
Coupling Climate and Hydrological Models Interoperability Through Web Services.
CCSM Portal/ESG/ESGC Integration (a PY5 GIG project) Lan Zhao, Carol X. Song Rosen Center for Advanced Computing Purdue University With contributions by:
Coupling Climate and Hydrological Models Interoperability Through Web Services Kathy Saint/SGI – NESII Jon Goodall/University of South Carolina Richard.
Metadata Creation with the Earth System Modeling Framework Ryan O’Kuinghttons – NESII/CIRES/NOAA Kathy Saint – NESII/CSG July 22, 2014.
Fast Parallel Grid Remapping for Unstructured and Structured Grids Robert Oehmke NOAA Cooperative Institute for Research in Environmental Sciences University.
Earth System Modeling Infrastructure Cecelia DeLuca/ESMF-NCAR March 31-April 1, 2009 CHyMP Meeting.
Metadata for the Coupled Ocean/Atmosphere Mesoscale Prediction System (COAMPS) using the Earth System Modeling Framework (ESMF) Peter Bosler University.
NE II NOAA Environmental Software Infrastructure and Interoperability Program Cecelia DeLuca Sylvia Murphy V. Balaji GO-ESSP August 13, 2009 Germany NE.
Cecelia DeLuca, Don Stark, Chris Hill Arctic System Model Workshop May 20, 2008 Earth System Modeling Framework.
ESMF Development Status and Plans ESMF 4 th Community Meeting Cecelia DeLuca July 21, 2005 Climate Data Assimilation Weather.
Computational Design of the CCSM Next Generation Coupler Tom Bettge Tony Craig Brian Kauffman National Center for Atmospheric Research Boulder, Colorado.
Initial Results from the Integration of Earth and Space Frameworks Cecelia DeLuca/NCAR, Alan Sussman/University of Maryland, Gabor Toth/University of Michigan.
Introduction to Apache OODT Yang Li Mar 9, What is OODT Object Oriented Data Technology Science data management Archiving Systems that span scientific.
Coupling Climate and Hydrological Models Interoperability Through Web Services.
A framework to support collaborative Velo: Knowledge Management for Collaborative (Science | Biology) Projects A framework to support collaborative 1.
The use of modeling frameworks to facilitate interoperability Cecelia DeLuca/NCAR (ESMF) Bill Putman/NASA GSFC (MAPL) David Neckels/NCAR.
Integrated Model Data Management S.Hankin ESMF July ‘04 Integrated data management in the ESMF (ESME) Steve Hankin (NOAA/PMEL & IOOS/DMAC) ESMF Team meeting.
IBM Software Group ® Overview of SA and RSA Integration John Jessup June 1, 2012 Slides from Kevin Cornell December 2008 Have been reused in this presentation.
Shannon Hastings Multiscale Computing Laboratory Department of Biomedical Informatics.
Earth System Modeling Framework Status Cecelia DeLuca NOAA Cooperative Institute for Research in Environmental Sciences University of Colorado, Boulder.
ESMF/Curator Status Cecelia DeLuca CCSM Software Engineering Working Group Boulder, CO March 16, 2007 Climate Data Assimilaton Weather.
Large Scale Nuclear Physics Calculations in a Workflow Environment and Data Provenance Capturing Fang Liu and Masha Sosonkina Scalable Computing Lab, USDOE.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
1 Earth System Modeling Framework Documenting and comparing models using Earth System Curator Sylvia Murphy: Julien Chastang:
ESMF Status and Future Plans Cecelia DeLuca BEI Technical Review Boulder, CO March 13-14, 2007 Climate Data Assimilaton Weather.
GEON2 and OpenEarth Framework (OEF) Bradley Wallet School of Geology and Geophysics, University of Oklahoma
Portable Infrastructure for the Metafor Metadata System Charlotte Pascoe 1, Gerry Devine 2 1 NCAS-BADC, 2 NCAS-CMS University of Reading PIMMS provides.
Earth System Modeling Framework Python Interface (ESMP) October 2011 Ryan O’Kuinghttons Robert Oehmke Cecelia DeLuca.
Strategic Plan Implementation Cecelia DeLuca/NCAR (ESMF) December 17, 2008 ESMF Board/Interagency Meeting.
ESMF Strategic Discussion Cecelia DeLuca NOAA ESRL/University of Colorado ESMF Executive Board/Interagency Meeting June 12, 2014.
ESMF Regridding Update Robert Oehmke, Peggy Li, Ryan O’Kuinghttons, Mat Rothstein, Joseph Jacob NOAA Cooperative Institute for Research in Environmental.
ESMF Regridding Update Robert Oehmke Ryan O’Kuinghttons Amik St. Cyr.
Architecture View Models A model is a complete, simplified description of a system from a particular perspective or viewpoint. There is no single view.
August 2003 At A Glance The IRC is a platform independent, extensible, and adaptive framework that provides robust, interactive, and distributed control.
1 Accomplishments. 2 Overview of Accomplishments  Sustaining the Production Earth System Grid Serving the current needs of the climate modeling community.
Towards Self-Describing Workflows for Climate Models Kathy Saint – UCAR Ufuk Utku Turuncoglu – ITU Sylvia Murphy – NCAR Cecelia DeLuca – NCAR.
Earth System Curator and Model Metadata Discovery and Display for CMIP5 Sylvia Murphy and Cecelia Deluca (NOAA/CIRES) Hannah Wilcox (NCAR/CISL) Metafor.
Fire Emissions Network Sept. 4, 2002 A white paper for the development of a NSF Digital Government Program proposal Stefan Falke Washington University.
Curator: Gap Analysis (from a schema perspective) Rocky Dunlap Spencer Rugaber Georgia Tech.
Mike Hildreth DASPOS Update Mike Hildreth representing the DASPOS project 1.
Building Community and Capability through Common Infrastructure: ESMF and the Earth System Curator Cecelia DeLuca MAP Meeting College.
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
Enhancements for Hydrological Modeling in ESMF Cecelia DeLuca/NCAR (ESMF) December 19, 2008 AGU Fall Meeting.
ESMF and the future of end-to-end modeling Sylvia Murphy National Center for Atmospheric Research
Application of RDF-OWL in the ESG Ontology Sylvia Murphy: Julien Chastang: Luca Cinquini:
Climate-SDM (1) Climate analysis use case –Described by: Marcia Branstetter Use case description –Data obtained from ESG –Using a sequence steps in analysis,
A TIME-GCM CAM Multi-executable Coupled Model Using ESMF and InterComm Robert Oehmke, Michael Wiltberger, Alan Sussman, Wenbin Wang, and Norman Lo.
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech 5 th GO-ESSP Community Meeting.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
The Earth System Curator Metadata Infrastructure for Climate Modeling Rocky Dunlap Georgia Tech.
A Semi-Automated Digital Preservation System based on Semantic Web Services Jane Hunter Sharmin Choudhury DSTC PTY LTD, Brisbane, Australia Slides by Ananta.
Metadata Support for Model Intercomparison Projects Sylvia Murphy: Cecelia DeLuca: Julien.
A Quick Tour of the NOAA Environmental Software Infrastructure and Interoperability Group Cecelia DeLuca Dr. Robert Detrick visit March 28, 2012
VisIt Project Overview
INTAROS WP5 Data integration and management
University of Technology
Metadata Development in the Earth System Curator
Overview of Workflows: Why Use Them?
Scientific Workflows Lecture 15
Presentation transcript:

Earth System Modeling Infrastructure Cecelia DeLuca and the ESMF Team NCAR/CISL CCSM Software Engineering Working Group June 17, 2009

Evolution Extending beyond historical emphasis on high performance frameworks Integration across model, data, and organizational elements Multi-agency and international contributions and partnerships essential

Elements of Modeling Infrastructure 1.Tight coupling tools and interfaces Emphasis on performance Frequent, high volume transfers on high performance computers Examples: ESMF, Flexible Modeling System (FMS) 2.Loose coupling protocols and interfaces Emphasis on ease of use Lower volume and infrequent transfers on desktop and distributed systems Examples: Open Modeling Interface (OpenMI, hydrology), web services 3.Science gateways browse, search, and distribution of model components, models, and datasets workflows, visualization and analysis services workspaces and management tools for collaboration Examples: Earth System Grid (ESG), Hydrologic Information System (HIS)

Elements of Modeling Infrastructure 4.Metadata conventions and ontologies Structured information describing modeling artifacts Ideally, with automated production of metadata from models Examples: CF conventions, CIM (from EU Metafor project) in CMIP5 5.Governance Coordinated and controlled evolution of systems Example: ESMF Change Review Board – quarterly multi-agency meeting that sets priorities and tasks and creates release schedules

Sample Applications Multi-model ensembles for weather and climate Automated modeling workflows Meeting stakeholder requests for model output outside the scope of the original research Coordinating standards development across the international community

Programs National Unified Operational Prediction Capability (NOAA and DoD operational NWP centers) Global Interoperability Program (NOAA climate and weather) Modeling Analysis and Prediction Program (NASA) Battlespace Environments Institute (DoD regional codes) Common Metadata for Climate Modelling Digital Repositories (METAFOR, IS-ENES)

ESMF Update Cecelia DeLuca, Kathy Saint

Release Plan ESMF v1 Prototype ESMF v3 Index Space Operations ESMF_ArraySparseMatMul() ESMF v4 Grid Operations ESMF_GridCreate() ESMF_FieldRegrid() ESMFv5 Standardization Build, init, data types, error handling, … Beta release May 2009 ESMF v4.0.0 ESMF v2 Components, VM and Utils ESMF_GridCompRun()

Current Release Beta Release May 2009 Basis for next public release On-line and off-line (file to file) parallel generation of interpolation weights, bilinear and higher order methods Two higher-level data representations: connected logically rectangular blocks or fully unstructured mesh Attributes for Metadata handling, read and write in XML or plain text formats Optimization of sparse matrix multiply for large processor counts (10,000+) The Flow-Following Finite Volume Icosahedral Model (FIM) from NOAA GSD is converting to ESMF to couple to National Weather Service models ESMF parallel regrid: Grid-Grid, Grid-Mesh, Mesh-Mesh Bilinear and higher order

Performance Portablity 30+ platform/compiler combinations regression tested nightly (3000+ system and unit tests) Newer ports include Solaris and native Windows Performance at the petascale… The chart at right shows scaling of the ESMF sparse matrix multiply, used in regridding transformations, out to 16K processors. (ESMF v3.1.0rp2, MCT 2.5.1) Plot from Peggy Li, NASA/JPL Tested on ORNL XT4 in a variety of configurations. -N1 means 1 core per node. ASMM Run-Time Comparison msec

Loose coupling and integration with data services Prototype completed that translates ESMF initialize/run/finalize component interfaces into web services Supports loose coupling with hydrological, human impacts, and other services, and integration of components into service oriented architectures Will enable invocation of ESMF components and applications from a portal in a Grid-based (e.g. Teragrid) gateway Data and metadata generated by ESMF Attributes during runs can be stored back to portals for search, browse, comparison, and decision support Portal developed jointly with the Earth System Grid displays detailed scientific metadata … work funded by NSF.

Web services: design goals ESMF components can be turned into web services with minimal impact to component developers – Fortran and C++ Use standard web technologies – SOAP/XML – NetCDF – OpenDAP/Thredds Simple deployment and user interface 2 for 1 interface – with ESMF can get both a tight coupling interface and a service interface

Client Application Service Architecture Internet Tomcat/ Axis2 SOAP Service Component Connector Application libnetesmf Grid/Coupler Component Grid/Coupler Component WSDL SOAP Client WSDL OpenDAP NetCDF Component List Registrar

Current Status Initial prototype released – Grid Component wrapper w/ SOAP interface – Initialize, Run, Finalize component operations Current version – Support for Coupler components – Single SOAP service to support multiple components – Data movement supported using OpenDAP access to NetCDF files Next steps – New project to integrate with hydrology data and applications, including Hydrologic Information System (HIS) Desktop

ESMF in CCSM4 Fei Liu, Bob Oehmke, Ufuk Turuncoglu

ESMF in CCSM4 Motivation: create an ESMF based CCSM4 platform to explore: – Self-describing models – Interoperability with other modeling systems – Component as web service – Automated model workflow – Online regridding – Improved scalability General Approach – Introduce a wrapper layer between CCSM4 driver and ESMF component – Convert MCT-based model component to ESMF-based component – Verify against baseline global integrals to ensure they are bit- wise reproducing. – Completed all of CCSM4 dead, data and active components

Plan for CCSM4 MCT based CCSM Driver Shared code MCT based model component ESMF Library ESMFshr code MCT -> ESMF_Array Domain -> ESMF_ArrayBundle Infodata -> ESMF_State Attributes ESMF compliant model component ESMF_Array -> MCT ESMF_ArrayBundle -> Domain ESMF_State Attributes -> infodata Data structure independent model physics

Regridding Bob Oehmke

Higher-Order Interpolation CCSM now uses ESMF higher-order interpolation as its standard non-conservative remapping Replaces bilinear interpolation of atmosphere to ocean states Results in a large reduction in noise in ocean surface transports (33% in a measured case) Higher-order interpolation: – A patch is a 2nd order n-D polynomial representing source data – Patches generated for each corner of source cell – Each patch created by least-square fit through source data in cells surrounding corner – Destination value is weighted average of patch values – Full algorithm presented here in 2007 by David Neckels – Longer description in ESMF v4.0.0 Reference Manual

Weight Generation Higher-order weights generated by ESMF off-line application – Takes netCDF grid files and generates netCDF weight file – Format same as SCRIP, can be used as an alternative – Runs in parallel Multiple improvements over last year – Improved accuracy – More robust – Pole options are now: no pole, full-circle average, and n-point stencil pole Future work – More pole options (improved vector quantity handling) – Weight generation for conservative regridding

Noise reduction in CCSM transport Interp. noise Interpolation noise in a derivative of the zonal wind stress grid index in latitudinal direction ESMF higher order interpolation weights were used to map from a 2- degree Community Atmospheric Model (CAM) grid to an irregularly spaced POP ocean grid (384x320) dTAUx/dy was computed using interpolated fields – this is closely related to the curl of the wind stress, which drives the upper ocean circulation Noise is calculated as deviation of a point from the sum of itself plus four neighbors 33% reduction in noise globally compared to original bilinear interpolation Black = bilinear Red = higher-order ESMF v3.1.1 Green = higher order ESMF v4.0.0

Attributes in CCSM4 Ufuk Turuncoglu

CCSM4 and ESMF Attributes ESMF Attributes hold metadata about model components, fields, etc. Using ESMF Attributes, metadata about CCSM4 driver fields was exported in XML format Metadata includes: name long name (or description) standard name units whether it is an import or export field All the field metadata uses the CF conventions and Metafor Common Information Model (CIM) The standard names of CCSM driver fields were collected and corrected by working with the Curator project and component liaisons.

CCSM and ESMF Attributes Example: Contents of output XML file for CCSM atmosphere component;

Towards CCSM self-describing workflows Ufuk Utku Turuncoglu 1 Sylvia Murphy 2 Cecelia DeLuca 2 1 Istanbul Technical University 2 National Center for Atmospheric Research * please come and see our poster for more information

Outline Motivation Basic Concepts Implementation of CCSM Workflow Preliminary Results Future Plans

Motivation Create easy to use work environments for running complex models like CCSM4 Automatically document parameter changes and other aspects of model runs Hide interactions with complicated computing technologies Archive, reproduce and share experiments

Workflows A modeling workflow separates modeling tasks into smaller chunks, and combines them using dataflow pipelines. KEPLER was chosen because it is open source and supports different models of computation. A KEPLER director controls the execution of a workflow and sends execution instructions to the actors that are specialized to do sub-tasks. In other words, actors specify “what processing occurs” while the director specifies “when it occurs”.

Provenance Provenance is defined as structured information that keeps track of the origin and derivation of the workflow. The basic types of provenance information: System (system environment, OS, CPU architecture, compiler versions etc.) Data (history or lineage of data, data flows, input and outputs, data transformations) Process (statistics about workflow run, transferred data size, elapsed time etc.) Workflow (version, modifications etc.) Seek to collect provenance in a format that is easily displayed in a portal and linked to data outputs

Components of Workflow Environment Typical workflow

Hierarchical Collection of Provenance Information Multi-component structure of CCSM makes it complicated to collect provenance information.

Conceptual CCSM Workflow Split into two parts: run preparation and postprocessing Postprocessing is automically triggered when run workflow completes

Kepler Graphical Interface main workflow composite actor CCSM workflow

Development Tasks CCSM modifications for ESMF Attributes (explained earlier) A Perl script was written to set up the third-party environment on remote machines. ORNL/NCSU scripts that gather system provenance information were modified to work with CCSM. Added support for multiple source directories, multi- component models, XML output Collects extra system provenance information (compiler etc.) Created new Kepler actors: Grid; CCSM: build, modify, and run; post-processing Listener to trigger other workflows automatically Provenance recorder with modified XML option

CCSM Workflow Actors Created KEPLER actors: Grid actors; Job submission to TeraGrid CCSM specific actors; create, modify (env_*.xml) Post-processing actors Utility actors

Results & Future Plans A fully active CCSM4 workflow with BTR1 component set and 1.9x2.5_gx1v6 resolution was run on the TeraGrid (IBM BG/L Frost) for five days Provenance information was automatically collected Near future plans: Longer simulation will be done with two different resolutions and two different computers (Frost and Kraken) simultaneously Export collected provenance information into ESG

Questions? Contact: Project Repository: ewvc/esmfcontrib/workflow/ Demo Movie: w.htm

CCSM Metadata for IPCC Assessment Report 5 Sylvia Murphy/NCAR

Curator and CMIP5 Curator is coordinating metadata generation, standardization, and implementation in science gateways We are partners with Metafor who are responsible for the collection of metadata for AR5 We are collaborating closely with the Earth System Grid (ESG) to implement metadata technologies to search, explore, and compare model information Curator Partners National Center for Atmospheric Research Geophysical Fluid Dynamics Laboratory Georgia Institute of Technology EU Partners (BADC, BODC, etc.)

ESG data collections

CCSM AR4 Metadata

End-to-end modeling Linking datasets and simulations within a science gateway is just one part of an end-to-end modeling system that includes: – Self-describing models – Broader range of coupling interactions – Web services – Workflows – Community-developed ontologies – Science gateways Future work within Curator will combine these pieces into a variety of modeling systems

Questions? For more information please us at