Grid Remote Execution of Large Climate Models (NERC Cluster Grid) Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre www.resc.rdg.ac.uk.

Slides:



Advertisements
Similar presentations
Building Portals to access Grid Middleware National Technical University of Athens Konstantinos Dolkas, On behalf of Andreas Menychtas.
Advertisements

3rd Campus Grid SIG Meeting. Agenda Welcome OMII Requirements document Grid Data Group HTC Workshop Research Computing SIG? AOB Next meeting (AG)
The Reading e-Science Centre Jon Blower Reading e-Science Centre Environmental Systems Science Centre University of Reading United Kingdom.
South-east Regional e- Research Consortium (SeReRC) Keith Haines, Jon Blower, Dan Bretherton, Alastair Gemmell Reading e-Science Centre University of Reading.
EHarmony in Cloud Subtitle Brian Ko. eHarmony Online subscription-based matchmaking service Available in United States, Canada, Australia and United Kingdom.
UK Campus Grid Special Interest Group Dr. David Wallom University of Oxford.
1 OBJECTIVES To generate a web-based system enables to assemble model configurations. to submit these configurations on different.
1. Topics Is Cloud Computing the way to go? ARC ABM Review Configuration Basics Setting up the ARC Cloud-Based ABM Hardware Configuration Software Configuration.
Publishing applications on the web via the Easa Portal and integrating the Sun Grid Engine Publishing applications on the web via the Easa Portal and integrating.
Data streaming, collaborative visualization and computational steering using Styx Grid Services Jon Blower 1 Keith Haines 1 Ed Llewellin 2 1 Reading e-Science.
The NERC Cluster Grid Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre Environmental Systems Science Centre.
CGAM Running the Met Office Unified Model on HPCx Paul Burton CGAM, University of Reading
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
Cyberaide Virtual Appliance: On-demand Deploying Middleware for Cyberinfrastructure Tobias Kurze, Lizhe Wang, Gregor von Laszewski, Jie Tao, Marcel Kunze,
Grid for Coupled Ensemble Prediction (GCEP) Keith Haines, William Connolley, Rowan Sutton, Alan Iwi University of Reading, British Antarctic Survey, CCLRC.
CCSM Portal/ESG/ESGC Integration (a PY5 GIG project) Lan Zhao, Carol X. Song Rosen Center for Advanced Computing Purdue University With contributions by:
DIRAC Web User Interface A.Casajus (Universitat de Barcelona) M.Sapunov (CPPM Marseille) On behalf of the LHCb DIRAC Team.
SICSA student induction day, 2009Slide 1 Social Simulation Tutorial Session 6: Introduction to grids and cloud computing International Symposium on Grid.
GADS: A Web Service for accessing large environmental data sets Jon Blower, Keith Haines, Adit Santokhee Reading e-Science Centre University of Reading.
Tools and Utilities for parallel and serial codes in ENEA-GRID environment CRESCO Project: Salvatore Raia SubProject I.2 C.R. ENEA-Portici. 11/12/2007.
Open Web App. Purpose To explain Open Web Apps To explain Open Web Apps To demonstrate some opportunities for a small business with this technology To.
Running Climate Models On The NERC Cluster Grid Using G-Rex Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre Environmental.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
March 3rd, 2006 Chen Peng, Lilly System Biology1 Cluster and SGE.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
TRASC Globus Application Launcher VPAC Development Team Sudarshan Ramachandran.
Building simple, easy-to-use grids with Styx Grid Services and SSH Jon Blower, Keith Haines Reading e-Science Centre Environmental Systems Science Centre.
London e-Science Centre GridSAM Job Submission and Monitoring Web Service William Lee, Stephen McGough.
INFSO-RI Enabling Grids for E-sciencE Supporting legacy code applications on EGEE VOs by GEMLCA and the P-GRADE portal P. Kacsuk*,
A Web-based Distributed Simulation System Christopher Taewan Ryu Computer Science Department California State University, Fullerton.
BARRODALE COMPUTING SERVICES LTD. Spatial Data Activities at the Reading e-Science Centre Adit Santokhee, Jon Blower, Keith Haines Reading.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Styx Grid Services: Lightweight, easy-to-use middleware for e-Science Jon Blower Keith Haines Reading e-Science Centre, ESSC, University of Reading, RG6.
Composing workflows in the environmental sciences using Web Services and Inferno Jon Blower, Adit Santokhee, Keith Haines Reading e-Science Centre Roger.
DAME: A Distributed Diagnostics Environment for Maintenance Duncan Russell University of Leeds.
Wrapping Scientific Applications As Web Services Using The Opal Toolkit Wrapping Scientific Applications As Web Services Using The Opal Toolkit Sriram.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
Holding slide prior to starting show. A Portlet Interface for Computational Electromagnetics on the Grid Maria Lin and David Walker Cardiff University.
© Geodise Project, University of Southampton, Geodise Middleware & Optimisation Graeme Pound, Hakki Eres, Gang Xue & Matthew Fairman Summer 2003.
SEE-GRID-SCI The SEE-GRID-SCI initiative is co-funded by the European Commission under the FP7 Research Infrastructures contract no.
Building the e-Minerals Minigrid Rik Tyer, Lisa Blanshard, Kerstin Kleese (Data Management Group) Rob Allan, Andrew Richards (Grid Technology Group)
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
The EDGeS project receives Community research funding 1 Porting Applications to the EDGeS Infrastructure A comparison of the available methods, APIs, and.
Interactive Workflows Branislav Šimo, Ondrej Habala, Ladislav Hluchý Institute of Informatics, Slovak Academy of Sciences.
GIS in the cloud: implementing a Web Map Service on Google App Engine Jon Blower Reading e-Science Centre University of Reading United Kingdom
The impacts of climate change on global hydrology and water resources Simon Gosling and Nigel Arnell, Walker Institute for Climate System Research, University.
Some comments on Portals and Grid Computing Environments PTLIU Laboratory for Community Grids Geoffrey Fox, Marlon Pierce Computer Science, Informatics,
Status Report on the Validation Framework S. Banerjee, D. Elvira, H. Wenzel, J. Yarba Fermilab 15th Geant4 Collaboration Workshop 10/06/
Grid Interoperability Update on GridFTP tests Gregor von Laszewski
John Kewley e-Science Centre All Hands Meeting st September, Nottingham GROWL: A Lightweight Grid Services Toolkit and Applications John Kewley.
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Spanish National Research Council- CSIC Isabel.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Practical using WMProxy advanced job submission.
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
OGSA-DAI.
Reading e-Science Centre Technical Director Jon Blower ESSC Director Rachel Harrison CS Director Keith Haines ESSC Associated Personnel External Collaborations.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
Holding slide prior to starting show. GECEM: Grid-Enabled Computational Electromagnetics David W. Walker School of Computer Science Cardiff University.
Reading e-Science Centre
Working With Azure Batch AI
UK GridPP Tier-1/A Centre at CLRC
CRESCO Project: Salvatore Raia
Grid Canada Testbed using HEP applications
Initial job submission and monitoring efforts with JClarens
Laura Bright David Maier Portland State University
Presentation transcript:

Grid Remote Execution of Large Climate Models (NERC Cluster Grid) Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre Environmental Systems Science Centre University of Reading, UK

Main themes of presentation Sharing HPC clusters used for running climate models  Why share clusters  Grid approach to cluster sharing (NERC Cluster Grid: UK Environmental Res. Council)‏ G-Rex Grid middleware Large climate models as grid services Please also see demonstration and poster

Background Many NERC institutes now have HPC clusters  Beowulf clusters with commodity hardware  Common applications are ocean, atmosphere and climate models Pressure to justify spending and increase utilisation  Sharing clusters helps increase utilisation  Sharing clusters facilitates collaborations Running climate models on remote clusters in traditional way is not easy

Using remote clusters the traditional way Input data Output data Local Remote 100 GB SCP SSH Model input and output Model setup, including source code, work-flow scripts, model input and output

Computational challenges of Climate models Typical requirements  Parallel processing (MPI) with large number of processors (usually )‏  Each cluster needs high speed interconnection (e.g. Myrinet or Infiniband)‏  Long runs lasting several days  Large volumes of output  Large number of separate output files

NEMO Ocean Model (eg. European Operational Oceanogr.) Main parameters of a typical 1° Global Assimilation run for a one year:  Run with 40 processors  2-3 hours per year on Cluster Outputs 300 MB in 700 separate files as diagnostics every 5-10 minutes Output for a one year is roughly 20 GB, a total of separate files 50-year `Reanalysis` = 1Tb. Model automatically re-submitted as a new job each year

NERC Cluster Grid Includes 3 clusters so far... (plans for 11 clusters)  Reading (64 procs.), Proudman, (360 pr.), British Antarctic Survey (160 pr.)‏ Main aim  Make it easier to use remote clusters for running large models Key features  Minimal data footprint on remote clusters  Easy job submission and control  Light-weight grid middleware (G-Rex)  Load and performance monitoring (Ganglia)  Security

Grid Remote EXecution G-Rex is light-weight grid middleware Implemented in Java using Spring framework G-Rex server is a Web application  Allows applications to be exposed as services  Runs inside a servlet container G-Rex client program, grexrun, behaves as if the remote service were actually running on the user's own computer Security based on HTTP digest authentication

NEMO G-Rex service: Deployment scenario 1 Client Server NEMO launch scripts and forcing data (same every run)‏ Input and output via HTTP Port 9092 G-Rex server Tomcat port open to client Apache Tomcat G- Rex client NEMO model setup, including source code, work-flow scripts, input data and output from all runs

NEMO G-Rex service: Deployment scenario 2 Client Server NEMO launch scripts and forcing data (same every run)‏ Input and output via HTTP Port 9092 G-Rex server Apache Tomcat G- Rex client NEMO model setup, including source code, work-flow scripts, input data and output from all runs

Advantages of G-Rex Output continuously transferred back to user  Job can be monitored easily  No data transfer delay at end of run Files deleted from server when no longer needed  Prevents unnecessary accumulation of data  Reduces data footprint of services Work-flows can be created using shell scripts Very easy to install and use See Poster; Demonstration also available