Porting MM5 and BOLAM codes to the GRID

Slides:



Advertisements
Similar presentations
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Advertisements

A tool to enable CMS Distributed Analysis
DIRAC API DIRAC Project. Overview  DIRAC API  Why APIs are important?  Why advanced users prefer APIs?  How it is done?  What is local mode what.
SEE-GRID-SCI Applications of the Meteorology VO in the frame of SEE-GRID-SCI project The SEE-GRID-SCI initiative is co-funded by the.
Tools and Utilities for parallel and serial codes in ENEA-GRID environment CRESCO Project: Salvatore Raia SubProject I.2 C.R. ENEA-Portici. 11/12/2007.
RISICO on the GRID architecture First implementation Mirko D'Andrea, Stefano Dal Pra.
K.Harrison CERN, 6th March 2003 GANGA: GAUDI/ATHENA AND GRID ALLIANCE - Aims and design - Progress with low-level software - Progress with Graphical User.
:: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: :: GridKA School 2009 MPI on Grids 1 MPI On Grids September 3 rd, GridKA School 2009.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) GISELA Additional Services Diego Scardaci
CE Operating Systems Lecture 3 Overview of OS functions and structure.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
E-science grid facility for Europe and Latin America E2GRIS1 Gustavo Miranda Teixeira Ricardo Silva Campos Laboratório de Fisiologia Computacional.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
The EDGeS project receives Community research funding 1 Porting Applications to the EDGeS Infrastructure A comparison of the available methods, APIs, and.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Status report on Application porting at SZTAKI.
SEE-GRID-SCI WRF – ARW application: Overview The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
INFSO-RI Enabling Grids for E-sciencE Charon Extension Layer. Modular environment for Grid jobs and applications management Jan.
G.Govi CERN/IT-DB 1 September 26, 2003 POOL Integration, Testing and Release Procedure Integration  Packages structure  External dependencies  Configuration.
Distributed Logging Facility Castor External Operation Workshop, CERN, November 14th 2006 Dennis Waldron CERN / IT.
OPTIMIZATION OF DIESEL INJECTION USING GRID COMPUTING Miguel Caballer Universidad Politécnica de Valencia.
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
SEE-GRID-SCI REFS application: NOA The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures.
Grid Remote Execution of Large Climate Models (NERC Cluster Grid) Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre
SEE-GRID-SCI Vangelis Floros, Vasso Kotroni, Kostas Lagouvardos – NOA, Athens, GREECE Goran Pejanovic, Luka Ilic, Momcilo Zivkovic.
EGEE is a project funded by the European Union under contract IST Enabling bioinformatics applications to.
SEE-GRID-SCI Meteo-VO: Overview The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures.
David Adams ATLAS ATLAS Distributed Analysis and proposal for ATLAS-LHCb system David Adams BNL March 22, 2004 ATLAS-LHCb-GANGA Meeting.
INFSO-RI Enabling Grids for E-sciencE Ganga 4 Technical Overview Jakub T. Moscicki, CERN.
Ganga/Dirac Data Management meeting October 2003 Gennady Kuznetsov Production Manager Tools and Ganga (New Architecture)
EGEE is a project funded by the European Union under contract IST Generic Applications Requirements Roberto Barbera NA4 Generic Applications.
GDB Meeting CERN 09/11/05 EGEE is a project funded by the European Union under contract IST A new LCG VO for GEANT4 Patricia Méndez Lorenzo.
Seven things you should know about Ganga K. Harrison (University of Cambridge) Distributed Analysis Tutorial ATLAS Software & Computing Workshop, CERN,
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Implementation of GLUE 2.0 support in the EMI Data Area Elisabetta Ronchieri on behalf of JRA1’s GLUE 2.0 Working Group INFN-CNAF 13 April 2011, EGI User.
INFSO-RI Enabling Grids for E-sciencE EGEE is a project funded by the European Union under contract IST Report from.
SEE-GRID-SCI WRF-ARW model: Grid usage The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures.
UNICORE and Argus integration Krzysztof Benedyczak ICM / UNICORE Security PT.
Databases and DBMSs Todd S. Bacastow January 2005.
Practical using C++ WMProxy API advanced job submission
Licenses and Interpreted Languages for DHTC Thursday morning, 10:45 am
OGF PGI – EDGI Security Use Case and Requirements
Dag Toppe Larsen UiB/CERN CERN,
GridBench: A Tool for Benchmarking Grids
A comparison between a Computational Grid and a High-end Multicore Server in an academic environment David Risinamhodzi – North-west University- South.
Dag Toppe Larsen UiB/CERN CERN,
Design rationale and status of the org.glite.overlay component
GWE Core Grid Wizard Enterprise (
POW MND section.
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
Grid Application Support Group Case study Schrodinger equations on the Grid Status report 16. January, Created by Akos Balasko
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
BOSS: the CMS interface for job summission, monitoring and bookkeeping
CRESCO Project: Salvatore Raia
CompChem VO: User experience using MPI
Copyright © 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 2 Database System Concepts and Architecture.
Monitoring of the infrastructure from the VO perspective
Chapter 2: Database System Concepts and Architecture
Compiler Construction
Data, Databases, and DBMSs
Introduction to Apache
Module 01 ETICS Overview ETICS Online Tutorials
Wide Area Workload Management Work Package DATAGRID project
Overview of Workflows: Why Use Them?
Database System Concepts and Architecture
Site availability Dec. 19 th 2006
Production Manager Tools (New Architecture)
Introduction to the SHIWA Simulation Platform EGI User Forum,
Presentation transcript:

Porting MM5 and BOLAM codes to the GRID Earth Science Workshop January 30, 2009 – Paris, France The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures contract no. 211338 Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Goal Run MM5 and BOLAM models on the grid to perform Ensemble weather forecasting Develop a generic weather model execution framework Support for deterministic forecasting Easily adopt to various other forecast models (e.g. WRF, RAMS etc). Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Target workflow Weather models follow a specific workflow of execution Retrieval of Initial Conditions Pre- Processing Model Run Post Processing HTTP Framework should be able to incorporate different codes for pre/post-processing and model execution. Parametric configuration of initial data retrieval N.O.M.A.D.S NCEP-GFS (USA) Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Requirements Adopt existing NOA procedures for model execution Hide the Grid as much as possible Give the feeling of local execution Simplify existing procedures and improve execution times Utilize high-level tools that facilitate better quality code and overcome low-level interactions with the Grid Satisfy specific model requirements Usage of commercial compiler not available in the Grid Time restrictions for completing application execution Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Design Approach Keep existing “command-line” look’n’feel Re-use and improve existing code base (shell scripts) Utilize Python language to replace various parts of the existing workflow Exploit the GANGA framework for job management and monitoring Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Utilized Grid Services gLite WMS – Job management LFC – Data management MPICH 1.2.7 on gLite sites Ganga Developed in CERN. Endorsed by EGEE RESPECT program Provides a Python programming library and interpreter for object-oriented job management Facilitates high-level programming abstractions for job management More information: http://ganga.web.cern.ch/ganga/ Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Implementation Details MM5 and BOLAM codes compiled locally in UI with PGI Fortran 3 different binaries produced for MM5, for 2, 6 and 12 CPUs respectively MPICH also compiled with PGI. MPICH libraries used for MM5 binaries generation Binaries were packed and stored on LFC Downloaded in WNs before execution Include Terrain data Models are running daily as cronjobs. Notifications are send to users by email Log files and statistics are kept for post-mortem analysis Ganga also useful for identifying problems after execution Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Implemented Architecture UI UI N jobs UI WMS UI Ganga CE/WN LJM (Python) Lead-In/Out (shell script) ModelConfigfile LFC Binaries and Results SE Results Workflow Orchestrator (Python) Decode (Shell script) Pre-process (shell script) Model Run (shell script) Post-Process (shell script) http N.O.M.A.D.S NCEP-GFS (USA) mpiexec WN WN WN … WN GRID Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Ensemble Forecasting Each member is executed as a separate job 10 members in total, both for MM5 and BOLAM models Each member separately downloads its initial data from NCEP servers Whole ensemble execution is handled by a single compound job Compound job definition, execution and management handled by Ganga constructs (job splitters) Final stage of forecast production and graphics preparation performed locally on UI Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Initial Performance Results MM5: Typical execution time: ~2hrs (including scheduling overheads) Different completion times per member depending on total processors used. 12 process version takes ~40mins per member but takes longer time to get scheduled in a grid site. BOLAM: Typical execution time for 10 member ensemble forecast: 60-90mins (including scheduling overheads) One member takes ~25 minutes to complete in a local cluster with optimized binary. Ensemble would take ~4 hrs locally Overall, non uniformity of completion times due to Grid resources (un)availability Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Adopting the framework for different models Models that implement similar workflows should be easy to adopt Ultimately the user should only provide: The four workflow hooks Decode Pre-process Model run Post-process Model configuration file(s) Definition of different initial data sources Forecast region Terrain data Model binaries stored on the LFC Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Problems/Pending Issues Problems with initial data NCEP servers sometimes down or cannot generate requested files Grid resources availability imbed timely execution Not all members manage to complete on time Some may still be in scheduled state when time expires Grid robustness and predictability Jobs may be rescheduled while running in different sites for no apparent reason Central grid services might be unavailable (WMS, LFC) MM5 sensitive to execution environment Dying processes while model in parallel section MPI notoriously not well supported by grid sites (some sites “better” than others) Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Future Work Application is still in pilot phase Planned to execute “super-ensemble” runs by April Multi-model multi-analysis ensemble forecasting combining results from MM5, BOLAM, NCEP/ETA, NCEP/NMM Presentation to be delivered in UF4 in Catania. Anticipating more resources and better support from existing once. Earth Science Workshop, Paris FRANCE, 30 Jan 2009

Questions Earth Science Workshop, Paris FRANCE, 30 Jan 2009