EGEE-II INFSO-RI-031688 Enabling Grids for E-sciencE www.eu-egee.org EGEE and gLite are registered trademarks INFSO-RI-508833 Enabling Grids for E-sciencE.

Slides:



Advertisements
Similar presentations
INFSO-RI Enabling Grids for E-sciencE Workload Management System and Job Description Language.
Advertisements

FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Development of test suites for the certification of EGEE-II Grid middleware Task 2: The development of testing procedures focused on special details of.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks World-wide in silico drug discovery against.
INFSO-RI Enabling Grids for E-sciencE WISDOM mini-workshop Vincent Breton (CNRS-IN2P3, LPC Clermont-Ferrand) ISGC 2007 March 28th,
INFSO-RI Enabling Grids for E-sciencE EGEE Middleware The Resource Broker EGEE project members.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
Makrand Siddhabhatti Tata Institute of Fundamental Research Mumbai 17 Aug
The ATLAS Production System. The Architecture ATLAS Production Database Eowyn Lexor Lexor-CondorG Oracle SQL queries Dulcinea NorduGrid Panda OSGLCG The.
KISTI’s Activities on the NA4 Biomed Cluster Soonwook Hwang, Sunil Ahn, Jincheol Kim, Namgyu Kim and Sehoon Lee KISTI e-Science Division.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management Services - Overview Mike Mineter National e-Science Centre, Edinburgh.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Configuring and Maintaining EGEE Production.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Building Grid-enabled Virtual Screening Service.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Application Case Study: Distributed.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Grid application development with gLite and P-GRADE Portal Miklos Kozlovszky MTA SZTAKI.
Enabling Grids for E-sciencE ENEA and the EGEE project gLite and interoperability Andrea Santoro, Carlo Sciò Enea Frascati, 22 November.
L ABORATÓRIO DE INSTRUMENTAÇÃO EM FÍSICA EXPERIMENTAL DE PARTÍCULAS Enabling Grids for E-sciencE Grid Computing: Running your Jobs around the World.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
INFSO-RI Enabling Grids for E-sciencE Workload Management System Mike Mineter
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Information System on gLite middleware Vincent.
INFSO-RI Enabling Grids for E-sciencE V. Breton, 30/08/05, seminar at SERONO Grid added value to fight malaria Vincent Breton EGEE.
Grid Enabled High Throughput Virtual Screening Against Four Different Targets Implicated in Malaria Presented by Vinod.
- Distributed Analysis (07may02 - USA Grid SW BNL) Distributed Processing Craig E. Tull HCG/NERSC/LBNL (US) ATLAS Grid Software.
The EDGeS project receives Community research funding 1 SG-DG Bridges Zoltán Farkas, MTA SZTAKI.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
INFSO-RI Enabling Grids for E-sciencE The gLite Workload Management System Elisabetta Molinari (INFN-Milan) on behalf of the JRA1.
INFSO-RI Enabling Grids for E-sciencE Biomedical applications V. Breton, CNRS-IN2P3.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
INFSO-RI Enabling Grids for E-sciencE In silico docking on EGEE infrastructure, the case of WISDOM Nicolas Jacq LPC of Clermont-Ferrand,
EGEE-II INFSO-RI Enabling Grids for E-sciencE WISDOM in EGEE-2, biomed meeting, 2006/04/28 WISDOM : Grid-enabled Virtual High Throughput.
INFSO-RI Enabling Grids for E-sciencE EGEE Review WISDOM demonstration Vincent Bloch, Vincent Breton, Matteo Diarena, Jean Salzemann.
1 Grid2Win: porting of gLite middleware to Windows Dario Russo INFN Catania
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Activités biomédicales dans EGEE-II Nicolas.
Glite. Architecture Applications have access both to Higher-level Grid Services and to Foundation Grid Middleware Higher-Level Grid Services are supposed.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
INFSO-RI Enabling Grids for E-sciencE Αthanasia Asiki Computing Systems Laboratory, National Technical.
Enabling Grids for E-sciencE Workload Management System on gLite middleware - commands Matthieu Reichstadt CNRS/IN2P3 ACGRID School, Hanoi.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
SAM Sensors & Tests Judit Novak CERN IT/GD SAM Review I. 21. May 2007, CERN.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Grid2Win : gLite for Microsoft Windows Roberto.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Command Line Grid Programming Spiros Spirou Greek Application Support Team NCSR “Demokritos”
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks The LCG interface Stefano BAGNASCO INFN Torino.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Practical using WMProxy advanced job submission.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks gLite – UNICORE interoperability Daniel Mallmann.
User Interface UI TP: UI User Interface installation & configuration.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
SAM Status Update Piotr Nyczyk LCG Management Board CERN, 5 June 2007.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Overview of gLite, the EGEE middleware Mike Mineter Training Outreach Education National.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Job Management Claudio Grandi.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Enabling Grids for E-sciencE Work Load Management & Simple Job Submission Practical Shu-Ting Liao APROC, ASGC EGEE Tutorial.
2 nd EGEE/OSG Workshop Data Management in Production Grids 2 nd of series of EGEE/OSG workshops – 1 st on security at HPDC 2006 (Paris) Goal: open discussion.
Enabling Grids for E-sciencE Claudio Cherubino INFN DGAS (Distributed Grid Accounting System)
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Grid Computing: Running your Jobs around the World
Introduction to Grid Technology
Nicolas Jacq LPC, IN2P3/CNRS, France
WISDOM-II, status of preparation
In silico docking on grid infrastructures
Presentation transcript:

EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks INFSO-RI Enabling Grids for E-sciencE The Wisdom Environment Vincent Bloch CNRS-IN2P3 ACGRID School Hanoi (Vietnam) November 8th, 2007 Credits: Jean Salzemann

Enabling Grids for E-sciencE EGEE-II INFSO-RI WISDOM initiative WISDOM initiative aims to demonstrate the relevance and the impact of the grid approach to address drug discovery for neglected and emerging diseases. First achieved experiences: –Summer 2005: Wide In Silico Docking On Malaria (WISDOM) –Spring 2006: Accelerate drug design against H5N1 neuraminidase –Winter 2006: Second data challenge on Malaria Partners: –Grid infrastructures: EGEE, Auvergrid, TWGrid, EELA, EuChinaGrid, EuMedGrid –European projects: Embrace, BioinfoGrid, EGEE –Institutes and association: Fraunhofer SCAI, Academia Sinica of Taiwan, ITB, Unimo University, LPC, CMBA, CERN-ARDA, HealthGrid

Enabling Grids for E-sciencE EGEE-II INFSO-RI Challenges for high throughput virtual docking Example: data challenge against H5N1 NA 300,000 Chemical compounds: ZINC & Chemical combinatorial library Target (PDB) : Neuraminidase (8 structures) Millions of chemical compounds available in laboratories In vitro high Throughput Screening 1$/compound, nearly impossible Molecular docking (Autodock) ~100 CPU years, 600 GB data Data challenge on EGEE, Auvergrid, TWGrid ~6 weeks on ~2000 computers In vitro screening of 100 hits Hits sorting and refining

Enabling Grids for E-sciencE EGEE-II INFSO-RI Issues for the grid-enabled high throughput virtual docking Computer-based in-silico screening can help to identify the most promising leads for biological tests –Involve whole databases (ZINC) –reduces the cost of trail-and-error approach In silico docking is well-fitted for grid deployment –CPU intensive application –Huge amount of output –Embarrassingly Parallel Issues of a large scale grid deployment –The rate of submitted jobs must be carefully monitored –The amount of transferred data impacts on grid performance –Grid process introduces significant delays –Licensed software requires licenses distribution strategy on grid

Enabling Grids for E-sciencE EGEE-II INFSO-RI Grid tools used during the data challenges WISDOM –a workflow of grid job handling: automated job submission, status check and report, error recovery –push model job scheduling –batch mode job handling – DIANE –a framework for applications with master-worker model –pull mode job scheduling –interactive mode job handling with flexible failure recovery feature –

Enabling Grids for E-sciencE EGEE-II INFSO-RI Grid components interacting with WISDOM The WMS: –The user submits jobs via the Workload Management System –The Goal of WMS is the distributed scheduling and resource management in a Grid environment. –What does it allow Grid users to do?  To submit their jobs  To get information about their status  To cancel them  To retrieve their output –The WMS tries to optimize the usage of resources as well as execute user jobs as fast as possible

Enabling Grids for E-sciencE EGEE-II INFSO-RI WMS Components WMS is currently composed of the following parts: User Interface (UI) : access point for the user to the WMS Resource Broker (RB) : the broker of GRID resources, responsible to find the “best” resources where to submit jobs Job Submission Service (JSS) : provides a reliable submission system Information Index (BDII) : a server (based on LDAP) which collects information about Grid resources – used by the Resource Broker to rank and select resources Logging and Bookkeeping services (LB) : store Job Info available for users to query

Enabling Grids for E-sciencE EGEE-II INFSO-RI Grid components interacting with WISDOM DMS: Data Management system –The user can store files on the grid through the DMS. –The goal of the DMS is to virtualize data on the grid and guarantee security integrity, and reliability of the data –What it allows Grid users to do:  Copy Files on the Grid  Register files on the Grid with a logical name  Store and manage metadata related to a file  Replicate files on the Grid  Delete files on the Grid  Retrieve files from the Grid

Enabling Grids for E-sciencE EGEE-II INFSO-RI DMS Components LFC (LCG File Catalogue): –It is used to register files on the grid –LFC provides functionalities to give logical names to files and organize them in directories GridFTP: –Low level file transfer protocol –Secured and reliable AMGA: –It is an grid interface for relational databases –Can be used to store medata –Can be used as a file catalogue

Enabling Grids for E-sciencE EGEE-II INFSO-RI other components interacting with WISDOM VOMS (Virtual Organisation Membership Service) –Store information concerning VO and roles FlexLm floating licenses server Web Portals –Can be used to visualize statistics or results Remote Database Servers –Can be used to store some information remotely (results, metadata etc..)

Enabling Grids for E-sciencE EGEE-II INFSO-RI WISDOM technology WISDOM has been specifically developed around EGEE middlewares (LCG-2.7, Glite). It uses a Java Multithreaded submission Engine Main scripts are written in perl Job-related scripts in written shell script (bash) Future environment will include – Web Services technology (WS-I profile) –Java and Python AMGA clients –All the code written in Java –Security and fine-grained ACLs

Enabling Grids for E-sciencE EGEE-II INFSO-RI main scripts: –wisdom_submit:  submits the jobs with a java multithreaded submission engine  stores the job ID and command lines and store them in a database. –wisdom_status:  checks the status of jobs regularly  handle the resubmissions of failed, aborted and cancelled jobs.  reads the IDs from wisdom_submit database  stores the job IDs in a table to prevent crushing wisdom_submit files, along with other parameters: job number a submission job status (unsubmitted, submitted, done) job submission count.  The process will loop until all the jobs of the instance are not finished. WISDOM ENVIRONMENT (1/2)

Enabling Grids for E-sciencE EGEE-II INFSO-RI WISDOM ENVIRONMENT (2/2) Several Features: –No input and output sandboxes in jobs.  All the target files, ligands and software are copied dynamically from the SE to the WN to unload the RB i/o.  FlexX outputs and Grid outputs are saved on several SEs through LFC and GridFTP –Jobs JDL and scripts are generated just before any submission  to take the wisdom.conf modifications into account (CE and RB black lists, job submission frequency)  are deleted afterward to save disk space. –Dynamic insertions of docking results and statistics in databases which allow real-time visualisation of the DC status. –wisdom_status can be stopped at any given time and restarted: it saves its own memory environment, so it can be restarted after a crash.

Enabling Grids for E-sciencE EGEE-II INFSO-RI Instance Definition The instance is a set of jobs regrouped accordingly to different criteria. The instance is unique, and has its own name The instance is submitted entirely on the grid, then it is followed up The instance name is by default: Instance’s jobs are called after the instance name: J

Enabling Grids for E-sciencE EGEE-II INFSO-RI WISDOM deployment GRID Grid services (RB, RLS…) Grid resources (CE, SE) Application components (Software, database) installation InstallerTester Test the grid wisdom_execution Workload definition Job submission Job monitoring Job bookkeeping Fault tracking Fault fixing Job resubmission Set of jobs User Collection Accounting data Superviser Web site database License server

Enabling Grids for E-sciencE EGEE-II INFSO-RI WISDOM Integration example DMS/GFTPDMS/GFTP User Interface HealthGrid Server Web Site WMS SEsCEs &WNs User Interface Wisdom_submit Wisdom_status WMS Submits the jobs Checks job status resubmits CEs &WNs FlexX job SEs Structure file Compounds file inputs outputs Output file Web Site WISDOM DB Output DB Docking information Statistics FlexX Statistics Flexlm server Flexlm server Flexlm server

Enabling Grids for E-sciencE EGEE-II INFSO-RI Environment architecture (1/2) wisdom.conf –the file that define the configuration of the instance wisdom_submit.sh –the execution script that launch the instance submission wisdom_submit.pl –the perl script of the execution process which submits the instance wisdom_status.sh –the execution script that launches the instance status checking wisdom_status.pl –the perl script of the status checking

Enabling Grids for E-sciencE EGEE-II INFSO-RI Environment architecture (2/2) bin/flexx.sh (the flexx script that is run by the jobs) bin/mt-job-submit (the execution script of the multi-threaded submission engine) bin/MTJobSubmitter.jar (the jar file of the multi-threaded java submission engine) bin/checkit.sh (a script used at the end of jobs to check the status of the job and store everything on the grid) bin/lfc_env.sh (a script to set up the environment variables for LFC) input/ /db_urls* ( there are several files, 1000, 2000, 3000, 4000… each of these files has the sfn of the database subsets replica. It is used in case of failure of the LFC server). edg_wl_ui_config/* (this directory hold all the configuration files of the resource brokers) Files need to be edited accordingly to the application and the VO!

Enabling Grids for E-sciencE EGEE-II INFSO-RI Simplified grid workflow for WISDOM StorageElement ComputingElement Site1 Site2 StorageElement User interface ComputingElement Compounds database Parameter settings Target structures Results Statistics ResourceBroker Software WISDOM production system Jobs Subsets

Enabling Grids for E-sciencE EGEE-II INFSO-RI WISDOM and Security Instances are submitted by a given user –All the jobs of the instance are belonging to the same user –Resources are dependant on the user’s VO Outputs files –Stored on the VO storage elements and register with the VO LFC. –If LFC is failing, files are stored on the VO writable directory on a given storage element This implies that: Users must follow-up the execution, and need to renew their proxy if necessary Files stored, are, a priori, available to all the VO members

Enabling Grids for E-sciencE EGEE-II INFSO-RI The new environment Web Services Interface – better interoperability –Everything is controlled through a few set of operations (no more modification of the files are required) Dynamic storing/querying of the results and jobs information on the Grid using AMGA metadata management system Improved fault tolerance Improved flexibility –New applications can be deployed more easily –As well as corresponding data Secured and multi-user

Enabling Grids for E-sciencE EGEE-II INFSO-RI The new environment Entirely developed using Java –No need to use text files to send information between the submission and the monitoring of the jobs –Improved fault tolerance –Improved flexibility –New application easier to deploy Improved monitoring of the grid resources –Uses its own ranking based on BDII information –Takes into account the number of jobs submitted to the sites to avoid overloads (the jobs are sent where the free CPUs are) –Takes into account the jobs failures and failing reasons (the “bad” sites are penalized)

Enabling Grids for E-sciencE EGEE-II INFSO-RI The new environment Uses AMGA for jobs and data monitoring –Improved monitoring and statistics –Dynamic storage and query of the data and results –Allows “Pull Model” Web Services Interface –Better interoperability –Ease the access to the environment: everything is controlled through a few set of operations

Enabling Grids for E-sciencE EGEE-II INFSO-RI New environment process Retrieve BDII information concerning the CE (number of CPUs, free CPUs,…) Define a workload according to the CE information Initialize the voms proxy Generate the jobs JDL Submit the jobs using multithreaded submission Until all the jobs are successful: –Check the status of the jobs using multithreaded check –Resubmit jobs if needed –Re-initialize voms proxy if needed –Update instance information in AMGA

Enabling Grids for E-sciencE EGEE-II INFSO-RI New Environment Architecture

Enabling Grids for E-sciencE EGEE-II INFSO-RI hits hits_id rank simulation_id energy_level run mean_energy Cluster_count simulation dlg_file simulation_id Target_id Ligand_id Histogram_file file coordinates_blob hits_id ligands Ligand_id Library_id name pdbq_file mass psa logp donor acceptor Logd7_4 ring rb far atoms refra library Library_name library_id target name Target_id pdbqs maps 1,1 1,n 1,1 1,n 1,11,n 1,1 1,n project description Project_id Program_name Program_version Program_options Project_id 1,n 1,1 coordinates_file Agent_id job Result Database Schema (autodock)

Enabling Grids for E-sciencE EGEE-II INFSO-RI Monitoring schema

Enabling Grids for E-sciencE EGEE-II INFSO-RI Pull Model Instead of sending a task with the job, the job retrieve a task from the task database while running The job performs tasks as long as it is running Pros: –No need to define a workload before the job submission –No need to have all the jobs running –When a job fails, only the last task need to be recomputed Cons: –Need to store the results on the fly –No access to the output sandbox –Retrieving a task can increase the job overhead

Enabling Grids for E-sciencE EGEE-II INFSO-RI Questions?