The WeNMR HADDOCK portal goes crowd computing Alexandre M.J.J. Bonvin Project coordinator Bijvoet Center for Biomolecular Research Faculty of Science,

Slides:



Advertisements
Similar presentations
DataTAG WP4 Meeting CNAF Jan 14, 2003 Interfacing AliEn and EDG 1/13 Stefano Bagnasco, INFN Torino Interfacing AliEn to EDG Stefano Bagnasco, INFN Torino.
Advertisements

EU 2nd Year Review – Jan – Title – n° 1 WP1 Speaker name (Speaker function and WP ) Presentation address e.g.
INFSO-RI Enabling Grids for E-sciencE EGEE Middleware The Resource Broker EGEE project members.
Catania Science Gateway Framework Motivations, architecture, features Catania, 09/06/2014Riccardo Rotondo
Portals and Credentials David Groep Physics Data Processing group NIKHEF.
Volunteer Computing and Hubs David P. Anderson Space Sciences Lab University of California, Berkeley HUBbub September 26, 2013.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
:: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: :: GridKA School 2009 MPI on Grids 1 MPI On Grids September 3 rd, GridKA School 2009.
4th EGEE user forum / OGF 25 Catania, TheoSSA on AstroGrid-D Iliya Nickelt (AIP / GAVO), Thomas Rauch (IAAT / GAVO), Harry Enke (AIP.
INFSO-RI Enabling Grids for E-sciencE Workload Management System Mike Mineter
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Belle MC Production on Grid 2 nd Open Meeting of the SuperKEKB Collaboration Soft/Comp session 17 March, 2009 Hideyuki Nakazawa National Central University.
Group 1 : Grid Computing Laboratory of Information Technology Supervisors: Alexander Ujhinsky Nikolay Kutovskiy.
The EDGeS project receives Community research funding 1 SG-DG Bridges Zoltán Farkas, MTA SZTAKI.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
The ILC And the Grid Andreas Gellrich DESY LCWS2007 DESY, Hamburg, Germany
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
 Our mission Deploying and unifying the NMR e-Infrastructure in System Biology is to make bio-NMR available to the scientific community in.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Status report on Application porting at SZTAKI.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Evolution of a High Performance Computing and Monitoring system onto the GRID for High Energy Experiments T.L. Hsieh, S. Hou, P.K. Teng Academia Sinica,
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
A worldwide e-Infrastructure for NMR and structural biology A worldwide e-Infrastructure for NMR and structural biology Introduction Structural biology.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
Structural Biology on the GRID Dr. Tsjerk A. Wassenaar Biomolecular NMR - Utrecht University (NL)
© Geodise Project, University of Southampton, Geodise Middleware Graeme Pound, Gang Xue & Matthew Fairman Summer 2003.
SHIWA and Coarse-grained Workflow Interoperability Gabor Terstyanszky, University of Westminster Summer School Budapest July 2012 SHIWA is supported.
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI VO auger experience with large scale simulations on the grid Jiří Chudoba.
1 P-GRADE Portal hands-on Gergely Sipos MTA SZTAKI Hungarian Academy of Sciences.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks gLite – UNICORE interoperability Daniel Mallmann.
OpenNebula: Experience at SZTAKI Peter Kacsuk, Sandor Acs, Mark Gergely, Jozsef Kovacs MTA SZTAKI EGI CF Helsinki.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
SHIWA Simulation Platform (SSP) Gabor Terstyanszky, University of Westminster EGI Community Forum Munnich March 2012 SHIWA is supported by the FP7.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Update report Marco Verlato INFN (National Institute of Nuclear Physics) Division of Padova Italy CHAIN Workshop, Taipei, Taiwan,
RI EGI-TF 2010, Tutorial Managing an EGEE/EGI Virtual Organisation (VO) with EDGES bridged Desktop Resources Tutorial Robert Lovas, MTA SZTAKI.
11 Extending EMI middleware with DGs Peter Kacsuk, MTA SZTAKI Start date: Duration:
1 Globe adapted from wikipedia/commons/f/fa/ Globe.svg IDGF-SP International Desktop Grid Federation - Support Project SZTAKI.
INFN/IGI contributions Federated Clouds Task Force F2F meeting November 24, 2011, Amsterdam.
A worldwide e-Infrastructure and Virtual Research Community for NMR and structural biology Alexandre M.J.J. Bonvin Project coordinator Bijvoet Center for.
InSilicoLab – Grid Environment for Supporting Numerical Experiments in Chemistry Joanna Kocot, Daniel Harężlak, Klemens Noga, Mariusz Sterzel, Tomasz Szepieniec.
Enabling Grids for E-sciencE LRMN ThIS on the Grid Sorina CAMARASU.
Structural Biology on the Grid Christophe Schmitz Bijvoet Center for Biomolecular Research Faculty of Science, Utrecht University The Netherlands
Antonio Fuentes RedIRIS Barcelona, 15 Abril 2008 The GENIUS Grid portal.
SCI-BUS project Pre-kick-off meeting University of Westminster Centre for Parallel Computing Tamas Kiss, Stephen Winter, Gabor.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Requirements from the WeNMR VRC User Forum 2011, Vilnius Workshop.
EGI-InSPIRE RI EGI Compute and Data Services for Open Access in H2020 Tiziana Ferrari Technical Director, EGI.eu
Toward an open-data infrastructure for WeNMR and structural biology-related services Alexandre M.J.J. Bonvin WeNMR Project coordinator Bijvoet Center for.
An short overview of INSTRUCT and its computational requirements Alexandre M.J.J. Bonvin Project coordinator Bijvoet Center for Biomolecular Research Faculty.
The LGI Pilot job portal EGI Technical Forum 20 September 2011 Jan Just Keijser Willem van Engen Mark Somers.
EGI-InSPIRE RI An Introduction to European Grid Infrastructure (EGI) March An Introduction to the European Grid Infrastructure.
A Competence Center to Serve Translational Research from Molecule to Brain Alexandre M.J.J. Bonvin MoBrain CC coordinator Bijvoet Center for Biomolecular.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Accessing the VI-SEEM infrastructure
GPGPU use cases from the MoBrain community
BaBar-Grid Status and Prospects
Dynamic Deployment of VO Specific Condor Scheduler using GT4
Belle II Physics Analysis Center at TIFR
Marc van Dijk | WeNMR | StickyBits
Alexandre M.J.J. Bonvin MoBrain CC coordinator
Roberto Barbera (a nome di Livia Torterolo)
DIRAC 2014, Helsinki, May 21th, 2014 Alexandre M.J.J. Bonvin
CRC exercises Not happy with the way the document for testbed architecture is progressing More a collection of contributions from the mware groups rather.
Leveraging GPGPU Computing in Grid and Cloud Environments: First Results from the Mobrain CC Alexandre Bonvin Utrecht University
Promoting Desktop Grids Virtual Team (Q1/2014 updates)
Presentation transcript:

The WeNMR HADDOCK portal goes crowd computing Alexandre M.J.J. Bonvin Project coordinator Bijvoet Center for Biomolecular Research Faculty of Science, Utrecht University the Netherlands IDGF 2014, Helsinki, May 19 th, 2014

The WeNMR VRC

WeNMR VRC (May 2014) One of the largest (#users) VO in life sciences > 620 VO registered users (36% outside EU) ~ 1200 VRC members (>60% outside EU) ~ CPU cores > 2000 SI2K CPU years > 5M jobs User-friendly access to Grid via web portals NMR SAXS A worldwide e-Infrastructure for NMR and structural biology A worldwide e-Infrastructure for NMR and structural biology

A glimpse of the WeNMR services portfolio

Distribution of resources Currently (May 2014) ~ 100’000 CPU cores

WeNMR VRC users distribution Over 1200 VRC and 620 VO members (64% / 36% outside Europe)! (May 2014)

The HADDOCK portal

Haddock web portal

Molecular Docking

Over 4000 registered (now all have access to the grid-server via robot cert) HADDOCK server users distribution

What is happening behind the scene?

Grid daemons beyond a portal Data transfer / jobs User level – data transfer via web interface: Data input: ~10 kB to 10 MB Data output: ~10MB to a few GB Grid level – data/jobs transfer (UI WMS CE WN) : Job submission: ~250 to 2500 single jobs Data transfer: ~1 to 10MB Output: ~1-20MB per job Local cluster– data/jobs transfer: Job submission: ~10 to 50 single jobs Data: a few GB Some portals make use of storage elements for the data

HADDOCK Grid submission Each job in the pool directory consists of: – Shell script to be executed on WN – Compressed tar archive containing the data/directory structure for the job – File specifying the location where results should be copied – JDL script with requirements/ranks for selecting suitable sites, e.g. : Requirements = (Member("OSG_VERSION",other.GlueHostApplicationSoftwareRunTimeEnvironment) || (other.GlueCEPolicyMaxCPUTime 110 && Member("VO-enmr.eu-CNS1.2",other.GlueHostApplicationSoftwareRunTimeEnvironment)) ); Rank = ( other.GlueCEStateFreeJobSlots > 0 ? other.GlueCEStateFreeJobSlots : (-other.GlueCEStateWaitingJobs * 4 / ( other.GlueCEStateRunningJobs + 1 )) - 1 ); Software is remotely deployed and CE tagged with a software tag: For HADDOCK: VO-enmr.eu-CNS1.2 Submission via robot proxie – fully automated – no user/operator intervention.

HADDOCK goes crowd computing IDGF resources enabled at minimum cost! – IDGF CE is not (yet) discoverable in normal EGI resources – Requires direct submission to CE – Simple change to JDL script: SubmitTo = "cr2.edgi-grid.eu:8443/cream-pbs-homeboinc” Pre-deployed mini Linux VMs on BOINC resources with the software required for running HADDOCK jobs (thanks Jozsef!) Currently every 10 th HADDOCK grid job is submitted to IDGF resources and duplicated to ensure a high throughput of the HADDOCK server

Some statistics HADDOCK server grid jobs (total 2014 ~2.5M / IDGF ~38k)

Conclusions  Successful and smooth porting of the HADDOCK portal to IDGF resources  Current implementation combines EGI/OSG and IDGF resources in a transparent manner  A pre-defined percentage is targeted to IDGF  Full transparency once IDGF resources will be discoverable in the EGI brokering system  Performance / reliability similar to regular grid sites

Acknowledgments Arne Visscher DDSG VICI NCF (BigGrid) BioNMR WeNMR HADDOCK Inc. Gydo van Zundert, Charleen Don, Adrien Melquiond, Ezgi Karaca, Marc van Dijk, Joao Rodrigues, Mikael Trellet,..., Koen Visscher, Manisha Anandbahadoer, Christophe Schmitz, Panos Kastritis, Jeff Grinstead IDGF & EGI VT Joszef Kovacs, Robert Lovas, Ad Emmen, Tamas Kiss,...

The End Thank you for your attention!