Www.see-grid-sci.eu SEE-GRID-SCI Vangelis Floros, Vasso Kotroni, Kostas Lagouvardos – NOA, Athens, GREECE Goran Pejanovic, Luka Ilic, Momcilo Zivkovic.

Slides:



Advertisements
Similar presentations
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Advertisements

SEE-GRID-SCI Hands-On Session: Workload Management System (WMS) Installation and Configuration Dusan Vudragovic Institute of Physics.
© , Michael Aivazis DANSE Software Issues Michael Aivazis California Institute of Technology DANSE Software Workshop September 3-8, 2003.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Supporting MPI applications on the EGEE Grid.
SEE-GRID-SCI Applications of the Meteorology VO in the frame of SEE-GRID-SCI project The SEE-GRID-SCI initiative is co-funded by the.
Christopher Jeffers August 2012
SRNWP workshop - Bologne Short range ensemble forecasting at Météo-France status and plans J. Nicolau, Météo-France.
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
Tools and Utilities for parallel and serial codes in ENEA-GRID environment CRESCO Project: Salvatore Raia SubProject I.2 C.R. ENEA-Portici. 11/12/2007.
1 portal.p-grade.hu Further information on P-GRADE Gergely Sipos MTA SZTAKI Hungarian Academy of Sciences.
Flexibility and user-friendliness of grid portals: the PROGRESS approach Michal Kosiedowski
Nicholas LoulloudesMarch 3 rd, 2009 g-Eclipse Testing and Benchmarking Grid Infrastructures using the g-Eclipse Framework Nicholas Loulloudes On behalf.
RISICO on the GRID architecture First implementation Mirko D'Andrea, Stefano Dal Pra.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
SEE-GRID-SCI Regional Grid Infrastructure: Resource for e-Science Regional eInfrastructure development and results IT’10, Zabljak,
:: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: :: GridKA School 2009 MPI on Grids 1 MPI On Grids September 3 rd, GridKA School 2009.
SEE-GRID-SCI SEE-GRID-SCI Operations Procedures and Tools Antun Balaz Institute of Physics Belgrade, Serbia The SEE-GRID-SCI.
DORII Joint Research Activities DORII Joint Research Activities Status and Progress 6 th All-Hands-Meeting (AHM) Alexey Cheptsov on.
LCG Middleware Testing in 2005 and Future Plans E.Slabospitskaya, IHEP, Russia CERN-Russia Joint Working Group on LHC Computing March, 6, 2006.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
CE Operating Systems Lecture 3 Overview of OS functions and structure.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks WMSMonitor: a tool to monitor gLite WMS/LB.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
Enabling Grids for E- sciencE EGEE and gLite are registered trademarks EGEE-III INFSO-RI Analysis of Overhead and waiting times.
SEE-GRID-SCI The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures contract no.
The EDGeS project receives Community research funding 1 Porting Applications to the EDGeS Infrastructure A comparison of the available methods, APIs, and.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Status report on Application porting at SZTAKI.
SEE-GRID-SCI SEE-GRID-SCI ES VOs The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Services for advanced workflow programming.
SEE-GRID-SCI Overview of YAIM and SEE-GRID-SCI YAIM templates Dusan Vudragovic Institute of Physics Belgrade Serbia The.
SEE-GRID-SCI WRF – ARW application: Overview The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
INFSO-RI Enabling Grids for E-sciencE Charon Extension Layer. Modular environment for Grid jobs and applications management Jan.
SAM Sensors & Tests Judit Novak CERN IT/GD SAM Review I. 21. May 2007, CERN.
The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures contract no Workflow repository, user.
Overview Background: the user’s skills and knowledge Purpose: what the user wanted to do Work: what the user did Impression: what the user think of Ganga.
On the D4Science Approach Toward AquaMaps Richness Maps Generation Pasquale Pagano - CNR-ISTI Pedro Andrade.
SPI NIGHTLIES Alex Hodgkins. SPI nightlies  Build and test various software projects each night  Provide a nightlies summary page that displays all.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI How to integrate portals with the EGI monitoring system Dusan Vudragovic.
OPTIMIZATION OF DIESEL INJECTION USING GRID COMPUTING Miguel Caballer Universidad Politécnica de Valencia.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
SEE-GRID-SCI REFS application: NOA The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures.
INFSO-RI Enabling Grids for E-sciencE Using of GANGA interface for Athena applications A. Zalite / PNPI.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Practical using WMProxy advanced job submission.
SEE-GRID-SCI Meteo-VO: Overview The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Ganga/Dirac Data Management meeting October 2003 Gennady Kuznetsov Production Manager Tools and Ganga (New Architecture)
Scenario use cases Szymon Mueller PSNC. Agenda 1.General description of experiment use case. 2.Detailed description of use cases: 1.Preparation for observation.
M.-E. Bégin¹, S. Da Ronco², G. Diez-Andino Sancho¹, M. Gentilini³, E. Ronchieri ², and M. Selmi² ¹CERN, Switzerland, ² INFN-Padova, Italy, ³INFN-CNAF,
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Bob Jones EGEE project director.
GDB Meeting CERN 09/11/05 EGEE is a project funded by the European Union under contract IST A new LCG VO for GEANT4 Patricia Méndez Lorenzo.
Seven things you should know about Ganga K. Harrison (University of Cambridge) Distributed Analysis Tutorial ATLAS Software & Computing Workshop, CERN,
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks GOCDB4 Gilles Mathieu, RAL-STFC, UK An introduction.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
SEE-GRID-SCI WRF-ARW model: Grid usage The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures.
The SCEC CSEP TESTING Center Operations Review
Porting MM5 and BOLAM codes to the GRID
Grid Application Support Group Case study Schrodinger equations on the Grid Status report 16. January, Created by Akos Balasko
Robert Szuman – Poznań Supercomputing and Networking Center, Poland
A Web-enabled Approach for generating data processors
CRESCO Project: Salvatore Raia
CompChem VO: User experience using MPI
Monitoring of the infrastructure from the VO perspective
Leigh Grundhoefer Indiana University
Overview of Workflows: Why Use Them?
Presentation transcript:

SEE-GRID-SCI Vangelis Floros, Vasso Kotroni, Kostas Lagouvardos – NOA, Athens, GREECE Goran Pejanovic, Luka Ilic, Momcilo Zivkovic – SEWA, Belgrade, SERBIA Weather multi-model and multi-analysis ensemble forecasting on the Grid 4 th EGEE User Forum March 4, Catania, ITALY 4th EGEE User Forum, Catania ITALY, 4 March 2009 The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures contract no

Overview Scientific context – Problem definition Application gridification  Requirements  Architecture  Current implementation Sample results Issues and problems Future Work 4th EGEE User Forum, Catania ITALY, 4 March 2009

Application Problem Description 4th EGEE User Forum, Catania ITALY, 4 March 2009 Lorenz in 1963 discovered that the atmosphere, like any unstable dynamic system has : “a finite limit of predictability even if the model is perfect and even if the initial conditions are known almost perfectly”. In addition it is known that neither the models nor the initial conditions are perfect Problem: deterministic forecasts have limited predictability that relates to the chaotic behaviour of the atmosphere Solution: base the final forecast not only on the predictions of one model (deterministic forecast) but on an ensemble of weather model outputs

4th EGEE User Forum, Catania ITALY, 4 March 2009 MULTI-ANALYSIS ENSEMBLE based on perturbing the initial conditions provided to individual models, in order to generate inter-forecast variability depending on a realistic spectrum of initial errors One forecast model “driven” by various initial conditions MULTI-ANALYSIS ENSEMBLE based on perturbing the initial conditions provided to individual models, in order to generate inter-forecast variability depending on a realistic spectrum of initial errors One forecast model “driven” by various initial conditions MULTI-MODEL ENSEMBLE based on the use of multiple models that run with the same initial conditions, sampling thus the uncertainty in the models Many forecast models “driven” by the same initial conditions MULTI-MODEL ENSEMBLE based on the use of multiple models that run with the same initial conditions, sampling thus the uncertainty in the models Many forecast models “driven” by the same initial conditions Mult-model/analysis Ensemble

In the context of SEE-GRID-SCI project we are developing a REgional scale Multi-model, Multi- analysis Ensemble Forecasting system (REFS) The application exploits the grid infrastructure in South Eastern Europe region. This system comprises of four different weather prediction models (multi-model system).  BOLAM, MM5, NCEP/Eta and NCEP/NMM  The models run for the same region (South-East Europe) many times, each initialized with different initial conditions (multi-analysis) Production of a multitude of forecasts 4th EGEE User Forum, Catania ITALY, 4 March 2009 Regional scale ensemble forecasting system

REFS Goals Run BOLAM, MM5, NCEP/NMM and NCEP/Eta models on the Grid to perform Ensemble weather forecasting  Combine final results to generate a “super- ensemble” forecast Develop a generic weather model execution framework  Use the same code-base for all four models  Support for deterministic forecasting  Easily adopt to various other forecast models for the same workflow 4th EGEE User Forum, Catania ITALY, 4 March 2009 MPICH enabled Serial code

Generic workflow 4th EGEE User Forum, Catania ITALY, 4 March 2009 Retrieval of Initial Conditions Pre- Processing Model Run Post Processing FTP N.O.M.A.D.S NCEP-GFS (USA) Weather models follow a specific workflow of execution A generic grid weather forecast framework should be able to incorporate different codes for pre/post- processing and model execution. Parametric configuration of initial data preparation Parametric execution based on Grid infrastructure capavbilities Customisation of execution steps

REFS Requirements Hide the Grid from end-users  Apply a regular “command-line” look’n’feel  Give the impression of local execution Re-use existing code base  Simplify existing procedures and improve execution times Utilize high-level tools that facilitate better quality code and overcome low-level interactions with the Grid  Use Python to replace various parts of the existing scripting code- base  Exploit the GANGA framework for job management and monitoring Satisfy specific technical restrictions  Usage of commercial compiler not available in grid sites  Time-bounded job execution 4th EGEE User Forum, Catania ITALY, 4 March 2009

Utilised Grid Services gLite  WMS, CE – Job management  LFC, SE – Data management and storage  MPICH on gLite sites Ganga  Developed in CERN. Endorsed by EGEE RESPECT program  Provides a Python programming library and interpreter for object-oriented job management  Facilitates high-level programming abstractions for job management  More information: 4th EGEE User Forum, Catania ITALY, 4 March 2009

Implementation Details Model sources compiled locally in UI with PGI Fortran  Multiple binaries produced for MM5, optimised for different number of processors (e.g. 2, 6, 12 CPUs) Binaries packed and stored on LFC  Downloaded in WNs before execution  Includes Terrain data Models are running daily as cronjobs. Notifications are send to users by  Log files and statistics are kept for post-mortem analysis  Ganga also useful for debugging and results archiving 4th EGEE User Forum, Catania ITALY, 4 March 2009

LFC Software Architecture 4th EGEE User Forum, Catania ITALY, 4 March 2009 UI Ganga LJM (Python) UI CE/WN Workflow Execution Decode (Shell script) Pre- process (shell script) Model Run (shell script) Post- Process (shell script) N.O.M.A.D.S NCEP-GFS (USA) ftp SE Binaries and Results WMS N jobs Results GRID MPICH WNWN WNWN WNWN WNWN WNWN WNWN WNWN WNWN … Model Config file NCEPget (Python) NCEPget (Python) Job Config File

Ensemble Forecast Execution Each member is executed as a separate job  10 members in total, both for MM5 (2-12 CPUs per member) and BOLAM models  10 members + 1 control job for NMM (8 CPUs per member)  Each member separately downloads its initial data from NCEP servers Whole ensemble execution is handled by a single compound job Compound job definition, execution and management handled by Ganga constructs (job splitters) Final stage of forecast production and graphics preparation performed locally on UI 4th EGEE User Forum, Catania ITALY, 4 March 2009

PRECIPITATION FORECASTS: BOLAM 10 members

Example of probabilistic forecasts 4th EGEE User Forum, Catania ITALY, 4 March 2009 From the multitude of model outputs probabilistic forecasts can be issued like: Probability of more than: -1mm of rain - 5 mm of rain -10 mm of rain Probability of exceedance of a temperature threshold Probability of exceedance of a wind speed threshold

Problems/Pending Issues Problems with initial data  NCEP servers sometimes down or cannot generate requested files Grid resources availability imbed timely execution  Not all members manage to complete on time  Some may still be in scheduled state when time expires Grid robustness and predictability  Jobs may be rescheduled while running in different sites for no apparent reason  Unavailability of central grid services (WMS, LFC) MM5 sensitive to execution environment  Dying processes while model in parallel section  MPI notoriously not well supported by grid sites (some sites “better” than others) 4th EGEE User Forum, Catania ITALY, 4 March 2009

Initial Performance Results MM5: Expected completion time: ~2hrs (including scheduling overheads) but large failure rate.  Different completion times per member depending on total processors used.  12 process version takes ~40mins per member but exposes larger scheduling overhead BOLAM: Expected completion time for 10 members: 2,5 hrs (including scheduling overheads).  One member takes ~25 minutes to complete in a local cluster with optimized binary. Ensemble would take ~4 hrs locally NMM: Expected completion time per member ~9-10 mins minutes. 4th EGEE User Forum, Catania ITALY, 4 March 2009

Current Status and Future Work Application running in pilot phase  Already ported MM5, BOLAM  NCEP/NMM on-going  NCEP/Eta under way Planned to start “super-ensemble” runs by April Anticipating more resources and better support from existing once.  Support from EGEE Earth Science VO 4th EGEE User Forum, Catania ITALY, 4 March 2009 More information, documentation and source code available from

Thank you 4th EGEE User Forum, Catania ITALY, 4 March 2009