The eMinerals minigrid and the national grid service: A user’s perspective NGS169 (A. Marmier)

Slides:



Advertisements
Similar presentations
E-Science Update Steve Gough, ITS 19 Feb e-Science large scale science increasingly carried out through distributed global collaborations enabled.
Advertisements

John Kewley CCLRC Daresbury Laboratory NW-GRID Training Event 26 th January 2007 GROWL Scripts and Web Services John Kewley Grid Technology Group E-Science.
Cluster Computing at IQSS Alex Storer, Research Technology Consultant.
The UCL Condor Pool Experience John Brodholt 1, Paul Wilson 3, Wolfgang Emmerich 2 and Clovis Chapman Department of Earth Sciences, University College.
A Grid approach to Environmental Molecular Simulations: Deployment and use of Condor within the eMinerals Mini Grid. Paul Wilson 1, Mark Calleja 2, John.
Job Submission Using PBSPro and Globus Job Commands.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Jon Wakelin Condor, Globus and SRB: Tools for Constructing a Campus Grid.
Dr. David Wallom Experience of Setting up and Running a Production Grid on a University Campus July 2004.
Dr. David Wallom Use of Condor in our Campus Grid and the University September 2004.
GRID workload management system and CMS fall production Massimo Sgaravatto INFN Padova.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Workload Management Massimo Sgaravatto INFN Padova.
First steps implementing a High Throughput workload management system Massimo Sgaravatto INFN Padova
Evaluation of the Globus GRAM Service Massimo Sgaravatto INFN Padova.
EMinerals and the Condor Pool John Brodholt UCL Arnaud Marmier Zhimei Du Maria Alfredsson Clovis Chapman Marc Blanchard Presenting the work of :
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Track 1: Cluster and Grid Computing NBCR Summer Institute Session 2.2: Cluster and Grid Computing: Case studies Condor introduction August 9, 2006 Nadya.
Building a Real Workflow Thursday morning, 9:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Matthew Palmer, Cambridge University01/10/2015 First Use of the UK e-Science Grid Overview The Physics Experiences Looking forward Conclusions Matthew.
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
ATLAS DC2 seen from Prague Tier2 center - some remarks Atlas sw workshop September 2004.
Grid tool integration within the eMinerals project Mark Calleja.
Ian C. Smith ULGrid – Experiments in providing a campus grid.
The National Grid Service Guy Warner.
Neil Geddes GridPP-10, June 2004 UK e-Science Grid Dr Neil Geddes CCLRC Head of e-Science Director of the UK Grid Operations Support Centre.
Dr Neil Geddes CCLRC Head of e-Science Director of the UK Grid Operations Support Centre.
Building a Real Workflow Thursday morning, 9:00 am Greg Thain University of Wisconsin - Madison.
NGS Innovation Forum, Manchester4 th November 2008 Condor and the NGS John Kewley NGS Support Centre Manager.
Grid NERSC demo Shreyas Cholia Open Software and Programming NERSC User Group Meeting September 19, 2007.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
© 2007 UC Regents1 Track 1: Cluster and Grid Computing NBCR Summer Institute Session 1.1: Introduction to Cluster and Grid Computing July 31, 2007 Wilfred.
Report from USA Massimo Sgaravatto INFN Padova. Introduction Workload management system for productions Monte Carlo productions, data reconstructions.
“Grids and eScience” Mark Hayes Technical Director - Cambridge eScience Centre GEFD Summer School 2003.
TeraGrid Advanced Scheduling Tools Warren Smith Texas Advanced Computing Center wsmith at tacc.utexas.edu.
The UK eScience Grid (and other real Grids) Mark Hayes NIEeS Summer School 2003.
Building a Real Workflow Thursday morning, 9:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
Applications & a Reality Check Mark Hayes. Applications on the UK Grid Ion diffusion through radiation damaged crystal structures (Mark Calleja, Mark.
© Geodise Project, University of Southampton, Geodise Middleware & Optimisation Graeme Pound, Hakki Eres, Gang Xue & Matthew Fairman Summer 2003.
July 11-15, 2005Lecture3: Grid Job Management1 Grid Compute Resources and Job Management.
Building the e-Minerals Minigrid Rik Tyer, Lisa Blanshard, Kerstin Kleese (Data Management Group) Rob Allan, Andrew Richards (Grid Technology Group)
Review of Condor,SGE,LSF,PBS
12th September 2007UK e-Science All Hands Meeting1 John Kewley Grid Technology Group e-Science Centre STFC Daresbury Laboratory GROWL.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
Creating and running an application.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
HEP SYSMAN 23 May 2007 National Grid Service Steven Young National Grid Service Manager Oxford e-Research Centre University of Oxford.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
GRID activities in Wuppertal D0RACE Workshop Fermilab 02/14/2002 Christian Schmitt Wuppertal University Taking advantage of GRID software now.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
John Kewley e-Science Centre All Hands Meeting st September, Nottingham GROWL: A Lightweight Grid Services Toolkit and Applications John Kewley.
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
Performance guided scheduling in GENIE through ICENI
Grid Compute Resources and Job Management. 2 Grid middleware - “glues” all pieces together Offers services that couple users with remote resources through.
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
The National Grid Service Mike Mineter.
The National Grid Service Mike Mineter
Collaborative Tools for the Grid V.N Alexandrov S. Mehmood Hasan.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 22 February 2006.
First evaluation of the Globus GRAM service Massimo Sgaravatto INFN Padova.
The UK National Grid Service Andrew Richards – CCLRC, RAL.
1 eScience Grid Environments th May 2004 NESC - Edinburgh Deployment of Storage Resource Broker at CCLRC for E-science Projects Ananta Manandhar.
UK Grid: Moving from Research to Production
US CMS Testbed.
The National Grid Service
The National Grid Service Mike Mineter NeSC-TOE
Presentation transcript:

The eMinerals minigrid and the national grid service: A user’s perspective NGS169 (A. Marmier)

Objectives 1. User Profile 2. Two real resources: eMinerals Minigrid National Grid Service 3. Practical Difficulties 4. Amateurish rambling (discussion/suggestions)

User Profile 1 Atomistic modelling community Chemistry/physics/material science Potentially big users of eScience (CPU intensive, NOT data) VASP, SIESTA, DL_POLY, CASTEP … Want to run parallel codes

User Profile 2 Relative proficiency with Unix, mainframes, etc … Scripting  parallel programming Note of caution: Speaker might be biased Want to run parallel codes

eMinerals Virtual Organisation, NERC The eMinerals project brings together simulation scientists, applications developers and computer scientists to develop UK eScience/grid capabilities for molecular simulations of environmental issues Grid prototype: the minigrid

eMinerals: Minigrid 3 clusters of 16 pentiums UCL condor pool Earth Science Cambridge condor pool SRB vaults SRB manager at Daresbury

eMinerals: Minigrid philosophy Globus 2 No Login possible (except one debug/compile cluster) No easy Files transfer (have to use SRB, see later) Feels very ‘gridy’, but not painless Promotes condorG and home wrappers

eMinerals: Minigrid example Universe = globus Globusscheduler = lake.bath.ac.uk/jobmanager-pbs Executable = /home/arnaud/bin/vasp-lam-intel Notification = NEVER transfer_executable = true Environment = LAMRSH=ssh -x GlobusRSL = (job_type=mpi)(queue=workq)(count=4)(mpi_type=lam-intel) Sdir = /home/amr.eminerals/run/TST.VASP3 Sget = INCAR,POTCAR,POSCAR,KPOINTS Sget = OUTCAR,CONTCAR SRBHome = /home/srbusr/SRB3_3_1/utilities/bin log = vasp.log error = vasp.err output = vasp.out Queue My_condor_submit script example

NGS: What ? VERY NICE PEOPLE who offer access to LOVELY clusters Real GRID approximation

NGS: Resources “Data” Clusters: 20 compute nodes with dual Intel Xeon 3.06 GHz CPUs, 4 GB RAM grid-data.rl.ac.uk - RAL grid-data.man.ac.uk – Manchester “Compute” Clusters: 64 compute nodes with dual Intel Xeon 3.06 GHz CPUs, 2 GB RAM grid-compute.leeds.ac.uk - WRG Leeds grid-compute.oesc.ox.ac.uk – Oxford Plus Other nodes : HPCx, Cardiff, Bristol …

NGS: Setup Grid-proxy-init Gsi-ssh … Then, a “normal” machine  Permanent fixed account (NGS169)  unix  queuing system With gsi-ftp for file transfer

NGS: example globus-job-run grid-compute.oesc.ox.ac.uk/jobmanager-fork /bin/ls globusrun -b grid-compute.oesc.ox.ac.uk/jobmanager-pbs example1.rsl [EXAMPLE1.RSL: & (executable=DLPOLY.Y) (jobType=mpi) (count=4) (environment=(NGSMODULES intel- math:gm:dl_poly))

Interlude

Difficulty 1: access Well known problem - Certificate - Globus enabled machine - SRB account (2.0)

Difficulty 2: Usability How do I submit a job ? Directly (gsi-ssh…) Remotely (globus,condorG) Direct Login, checkq, submit, (kill), logout Different Batch Queuing Systems (PBS, condor,LoadLeveler …)

Usability 2 Usually requires a “script” Almost nobody writes their own scripts Works by inheritance and adaptation At the moment eScience forces the user to learn the syntax of the B.Q.S.

Usability 3 Remote [EXAMPLE1.RSL: & (executable=DLPOLY.Y) (jobType=mpi) (count=4) (environment=(NGSMODULES intel-math:gm:dl_poly)) Ignores file transfer Ignores more complex submit structures

Usability 4 Ignores more complex submit structures  abinit <inp.txt  Cpmd.x MgO.inp => User has to learn globus syntax :o/ (environment and RSL)

Finally At the moment no real incentives to submit remotely Mechanism to reward the early adopters Access to special queues Longer walltime ? More cpus ?

CONCLUSION Submission scripts are very important and useful pieces of information Easily accessible examples would save a lot of time Mechanism to encourage remote submission (access to better queues)