Workflows in Computational Chemistry Prof

Slides:



Advertisements
Similar presentations
Running Workflows on Clouds and Grids Gabor Terstyanszky, University of Westminster T. Fahringer, P. Kacsuk, J. Montagnat, I. Taylor e-Science Workshop,
Advertisements

© 2006 Open Grid Forum The Astro Community and DCIs in Europe and the role of Astro-CG C. Vuerli - INAF.
ER-flow Purpose of the meeting C. Vuerli Contributions by G. Terstyanszky.
P-GRADE and WS-PGRADE portals supporting desktop grids and clouds Peter Kacsuk MTA SZTAKI
Porto, January Grid Computing Course Summary of day 2.
A Definition and Analysis of the Role of Meta-workflows in Workflow Interoperability Junaid Arshad, Gabor Terstyanszky, Tamas Kiss, Noam Weingarten Center.
EXTENDING SCIENTIFIC WORKFLOW SYSTEMS TO SUPPORT MAPREDUCE BASED APPLICATIONS IN THE CLOUD Shashank Gugnani Tamas Kiss.
Workflows Information Flows Prof. Silvia Olabarriaga Dr. Gabriele Pierantoni.
EUROPEAN UNION Polish Infrastructure for Supporting Computational Science in the European Research Space Cracow Grid Workshop’10 Kraków, October 11-13,
SCI-BUS is supported by the FP7 Capacities Programme under contract nr RI WS-PGRADE/gUSE Supporting e-Science communities in Europe Zoltan Farkas.
ER-flow C. Vuerli Contributions by G. Terstyanszky, K. Varga.
Workflow sharing and integration services by the ER-flow project on behalf of the ER-flow consortium EGI Community Forum, Manchester,
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
Sharing, integrating and executing different workflows in heterogeneous multi-cloud systems Peter Kacsuk MTA SZTAKI SCI-BUS is supported.
G. Terstyanszky, T. Kukla, T. Kiss, S. Winter, J.: Centre for Parallel Computing School of Electronics and Computer Science, University of.
From P-GRADE to SCI-BUS Peter Kacsuk, Zoltan Farkas and Miklos Kozlovszky MTA SZTAKI - Computer and Automation Research Institute of the Hungarian Academy.
Sharing Workflows through Coarse-Grained Workflow Interoperability : Sharing Workflows through Coarse-Grained Workflow Interoperability G. Terstyanszky,
Introduction to SHIWA Technology Peter Kacsuk MTA SZTAKI and Univ.of Westminster
Introduction to WS-PGRADE and gUSE Tutorial Akos Balasko 04/17/
WS-PGRADE portal and its usage in the CancerGrid project M. Kozlovszky, P. Kacsuk Computer and Automation Research Institute of the Hungarian Academy of.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Convert generic gUSE Portal into a science gateway Akos Balasko 02/07/
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Services for advanced workflow programming.
A scalable and flexible platform to run various types of resource intensive applications on clouds ISWG June 2015 Budapest, Hungary Tamas Kiss,
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
Convert generic gUSE Portal into a science gateway Akos Balasko.
SHIWA and Coarse-grained Workflow Interoperability Gabor Terstyanszky, University of Westminster Summer School Budapest July 2012 SHIWA is supported.
Università di Perugia Enabling Grids for E-sciencE Status of and requirements for Computational Chemistry NA4 – SA1 Meeting – 6 th April.
Building an European Research Community through Interoperable Workflows and Data ER-flow project Gabor Terstyanszky, University of Westminster, UK EGI.
SHIWA: Is the Workflow Interoperability a Myth or Reality PUCOWO, June 2011, London Gabor Terstyanszky, Tamas Kiss, Tamas Kukla University of Westminster.
EGI Technical Forum Amsterdam, 16 September 2010 Sylvain Reynaud.
Application Specific Module Tutorial Zoltán Farkas, Ákos Balaskó 03/27/
1 SCI-BUS: building e-Science gateways in Europe: building e-Science gateways in Europe Peter Kacsuk and Zoltan Farkas MTA SZTAKI.
1 WS-PGRADE/gUSE generic DCI gateway framework for EGI user communities Zoltan Farkas and Peter Kacsuk MTA SZTAKI SCI-BUS is supported.
Introduction to the program of the summer school Peter Kacsuk MTA SZTAKI SCI-BUS is supported by the FP7 Capacities Programme under contract.
SCI-BUS is supported by the FP7 Capacities Programme under contract nr RI MTA SZTAKI background for the DARIAH CC Zoltan Farkas MTA SZTAKI LPDS,
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
The EDGeS project receives Community research funding 1 Support services for desktop grids and service grids by the EDGeS project Tamas Kiss – University.
Introduction to SHIWA project EGI User Forum, Vilnius Peter Kacsuk MTA SZTAKI
Supporting Big Data Processing via Science Gateways EGI CF 2015, November, Bari, Italy Dr Tamas Kiss, CloudSME Project Director University of Westminster,
Porting workflows for the Heliophysics Community Dr. Gabriele Pierantoni Trinity College Dublin.
SHIWA Simulation Platform (SSP) Gabor Terstyanszky, University of Westminster EGI Community Forum Munnich March 2012 SHIWA is supported by the FP7.
Usage of WS-PGRADE and gUSE in European and national projects Peter Kacsuk 03/27/
1 Globe adapted from wikipedia/commons/f/fa/ Globe.svg IDGF-SP International Desktop Grid Federation - Support Project SZTAKI.
New developments of gUSE & WS-PGRADE to support e-science gateways Peter Kacsuk and Miklos Kozlovszky MTA SZTAKI - Computer and Automation Research Institute.
SHIWA project presentation Project 1 st Review Meeting, Brussels 09/11/2011 Peter Kacsuk MTA SZTAKI
SCI-BUS project Pre-kick-off meeting University of Westminster Centre for Parallel Computing Tamas Kiss, Stephen Winter, Gabor.
Convert generic gUSE Portal into a science gateway Akos Balasko.
Exposing WS-PGRADE/gUSE for large user communities Peter Kacsuk, Zoltan Farkas, Krisztian Karoczkai, Istvan Marton, Akos Hajnal,
Building an European Research Community through Interoperable Workflows and Data ER-flow Prof. Gabor Terstyanszky, University of Westminster, UK Heildelberg,
HELIOGATE – A USER EXPERIENCE PERSPECTIVE G.Pierantoni, E. Carley, B. Bentley S. Olabarriaga E. Sciacca, G. Taffoni, G. Castelli, U. Becciani.
Building an European Research Community through Interoperable Workflow and Data Gabor Terstyanszky University of Westminster.
SHIWA SIMULATION PLATFORM = SSP Gabor Terstyanszky, University of Westminster e-Science Workflows Workshop Budapest 09 nd February 2012 SHIWA is supported.
Enabling Grids for E-sciencE University of Perugia Computational Chemistry status report EGAAP Meeting – 21 rst April 2005 Athens, Greece.
Accessing the VI-SEEM infrastructure
Coarse Grained Interoperability scenarios
Integrating Scientific Tools and Web Portals
Peter Kacsuk, Zoltan Farkas MTA SZTAKI
AAI for a Collaborative Data Infrastructure
INTAROS WP5 Data integration and management
Defining and tracking requirements for New Communities
WS-PGRADE for Molecular Sciences and XSEDE
Peter Kacsuk MTA SZTAKI
Recap: introduction to e-science
MIK 2.1 DBNS - introduction to WS-PGRADE, 2013
Introduction to the SHIWA Simulation Platform EGI User Forum,
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Workflows in Computational Chemistry Prof Workflows in Computational Chemistry Prof. Gabor Terstyanszky Centre for Parallel Computing University of Westminster School of Open Science Cloud Perugia, Italy 05 – 09 June 2017

Challenges and Motivation Learn how to create and use workflows Many different workflow systems exist that are not interoperable Technological choice (data and computer resources) isolates users and user communities Benefits of workflow Share your own workflow, re-use workflows of others Create and execute ‘meta-workflows’: built from smaller workflows that use different workflow languages/technologies Combine workflows and DCIs

Key Players and Challenges Domian researchers Researchers of one particular research field, for example Astrophysicists, Computational Chemists, Heliophysicists, Bio Scientists, etc. with basic computing knowledge 10 or 100 thousands or even millions Challenges: They are not familiar with the technology to run experiments on computing infrastructures and probably they will never learn it. Workflow developers They are familiar with both Computer Science and a particular research field up to a few thousands Workflow system developers Computer Scientists with knowledge about data and compute technologies up to a few hundreds 3

Workflow Developer’s Scenario They want - to develop & publish WFs They need - to develop WFs somewhere - to publish WFs somewhere - to manage WFs somewhere - to execute WFs somewhere Workflow Repository Science Gateway Clouds Local clusters Supercomputers Desktop grids (DGs) (BOINC, Condor, etc.) Cluster based service grids (SGs) (EGEE, OSG, etc.) Supercomputer based SGs (DEISA, TeraGrid) Grid systems E-science infrastructure

Domain Researcher’s Scenario They need - to run experiments seamlessly i.e. executing workflows which access to data and compute resources hiding all technical details They want - to run experiments Workflow Repository Science Gateway Clouds Local clusters Supercomputers Desktop grids (DGs) (BOINC, Condor, etc.) Cluster based service grids (SGs) (EGEE, OSG, etc.) Supercomputer based SGs (DEISA, TeraGrid) Grid systems 5 5 E-science infrastructure

Co-operation with Research Communities Phase 1 – introduction to the workflow technology Target group: communities without any or basic experience in the workflow technology Phase 2 – creating and running workflows Target group: communities those use workflows to run experiments Phase 3 – combining workflows of different workflow systems Target group: communities those use workflows to run experiments and are interested in using workflows of other workflow systems 6

Workflow Levels and Users Infrastructure DCI Gateway Conceptual Level Abstract Concrete Domain Researchers Experts Developers Workflow execution and monitoring API, Libraries, Development and Debugging tools Configuration, Statistics, Monitoring, Troubleshooting Workflow Development and Execution Workflow Execution

Atomic Workflow Concept meta-meta-workflow Execution of embedded meta-workflow Execution of meta-workflows meta-workflow Execution of the atomic workflow Execution on the TAVERNA line command tool atomic worfklows 8

Computational Chemistry: Meta-Workflow Concept (1) 9

Computational Chemistry: Meta-Workflow Concept (2)

Computational Chemistry: Meta-Workflow Concept (3)

Workflows versus Scientic Experiments Type of Workflow Technology Driven by Usage Atomic Workflow Basic workflows Use Case Execution of Atomic Workflows Meta-workflows Composite workflows of atomic workflows Science Case Orchestration of Atomic Workflows Iterative meta- workflows Parameter study of workflows or meta-workflows Iterative Science Case Parameter-Sweep Execution of Science Cases

Quantum Chemistry Workflows Name Description Engine Middle-ware Type ID Spectroscopic analysis Explore the spectroscopic characteristics of a molecule WS-PGRADE UNICORE meta-workflow 5739 Spectroscopic benchmarking Explore the spectroscopic characteristics of a molecule with more functionals/basis sets meta-meta-workflow 5745 Parameter sweep Benchmark a molecule using larger arrays of functionals and basis sets -- Transition state analysis Find a reaction transition state and analyse its frequencies together with the spectroscopic properties 5751 TD-UNI Calculate the optical response of a molecule Workflow   SpecWSUNI PopulationUNI Apply various population schemes to better electronic understanding of the molecules. Galaxy QM Optimise a molecule quantum chemically in Galaxy Galaxy 5754 QM-MD Optimise the protein scaffold by molecular dynamics and optimise then the metallo-active center by quantum mechanics together with a spectroscopic analysis 5752

Quantum Chemistry Science Cases Name Description Workflow Links Spectroscopic Analysis Explore the spectroscopic characteristics of a molecule http://www.erflow.eu/spectrosco pic-analysis-science-case Spectroscopic Benchmarking Calculation of optimized geometries, molecular orbitals, population analyses, frequencies, or optical absorptions. http://www.erflow.eu/spectrosco pic-benchmarking-science-case Population UNI Apply various population schemes to better electronic understanding of the molecules. http://www.erflow.eu/population- uni-science-case

Spectroscopic Analysis Science Cases Highly useful for Quantum Chemists in everyday work Full simulation of all spectroscopic features of a molecule Combination of 5 atomic workflows

Spectroscopic Benchmarking Science Cases Highly useful for Quantum Chemists in everyday work Combination of 1 atomic workflow and 4 times the Science case 1

Workflow Interoperability Challenges 17

Formal Description of WEs and WFs formal description of workflows WF = :{WFabs, WFcnr, WFcnf, WFeng} where WFabs - abstract workflow WFcnr - concrete workflow WFcnf - workflow configuration WFeng - workflow engine formal description of workflow engines WE = :{WEabs, WEcnc, WEcnf} where WEabs - abstract workflow engine WEcnr - concrete workflow engine WEcnf - workflow engine configuration

Workflow Interoperability: Coarse-Grained Interoperability (1) CGI concept = workflow engine integration Submission Service Workflow Engine B Distributed Computing Infrastructure Workflow Engine A Workflow Repository Submission Service Client

Workflow Interoperability: Coarse-Grained Interoperability (2) native workflows: J1, J2, J3 non-native workflows: WF4 - black boxes which are managed as legacy code applications

SHIWA Simulation Platform = SSP 21

SHIWA Simulation Platform SHIWA Portal WS-PGRADE Workflow engine SHIWA Repository SHIWA Submission Service DCI Bridge API cloud cluster desktop grid service grid super computer process submit publish/ search PGRADE workflows Galaxy Kepler MOTEUR Taverna etc. workflows export/import search sftp ftp http/ https SRM S3 Data Avenue data transfer SHIWA Science Gateway

SHIWA Portal: Editing Workflow 23 23

SHIWA Portal: Configuring Workflow 24

SHIWA Portal: Executing Workflow 25

SHIWA Workflow Repository 26

Workflow Interoperability: CGI Usage Scenario (1) workflow developer create migrate execute publish WFs SHIWA Simulation Platform Execute native WF SHIWA Portal DCIs Export WF Import WF SHIWA Repository SHIWA Submission Service Workflow Engines domain researcher import execute WFs Import WF Community Gateway

Workflow Interoperability: CGI Usage Scenario (2) SHIWA Simulation Platform SHIWA Portal DCIs SHIWA Repository Execute non-native WF SHIWA Submission Service Workflow Engines researcher Import WF Execute native WF import execute WFs Community Portal community gateway

Workflow Interoperability: CGI Usage Scenario + Taverna WF workflow developer SHIWA Simulation Platform Run WF CGI support: ASKALON Dispel4Py Galaxy (???) GWES Kepler MOTEUR Pegasus PGRADE ProActive Taverna Triana SHIWA Portal DCIs Search repo Import WF Execute non-native WF Execute non-native WF Upload WF myExperiment Repository SHIWA Repository Retrieve WF Execute non-native WF SHIWA Submission Service Taverna Engine 29

Distributed Computing Infrastructure Interoperability workflow for DCI B Distributed Computing Infrastructure Interoperability J2 J1 J4 J3 jobs in non-JSDL J2 J1 J4 J3 jobs in JSDL DCI support: cloud cluster desktop grid service grid supercomputer Workflow Engine JSDL Translator DCI Bridge DCI Metabroker Proxy Server 30

Usage Scenario: Domain Researcher View 31

Usage Scenario: Domain Researcher View 32

Usage Scenario: Customised View 33

Usage Scenario: Customised View 34

Knowledge Transfer and Research Community Support Academic communities Astrophysics PGRADE + Taverna Computational Chemistry Galaxy + PGRADE + UNICORE Heliophysics PGRADE + Taverna Hydrometeorology PGRADE Life Sciences Galaxy + Moteur + PGRADE + Taverna Meteorology PGRADE Material Sciences PGRADE Particle Physics PGRADE Seizmology Dispel4Py + PGRADE Non-academic communities Engineering and manufacturing SMEs Business Process Simulation Discrete Event Simulation Fluid Dynamics Simulation PGRADE

Creating and Executing Workflows workflows in the repository in 2013 abstract - 123 concrete - 119 total - 242 workflows in the repository in 2016 abstract - 213 concrete - 385 total - 598 workflow execution number SHIWA Simulation Platform - 512 (dev) / 331 (test) / 181 (training) Community gateways Astro workflows - 182 (dev) / 73 (test) 203 (prod) Compchem workflows - 550 (dyn) / 400 (dock) / 300 (quan) Helio workflows - 41 (prod) Life Sciences workflows - 79 (dev) / 325 (prod)

Research Community Support Phase 1 – introduction to the workflow technology Platform: SHIWA Simulation Platform Support: workflow creation and execution Training: platform and workflow training Phase 2 – creating and running workflows Platform: SHIWA Simulation Platform or community portal Support: portal deployment + workflow porting Training: gateway deployment and management Phase 3 – combining workflows of different workflow systems Platform: SHIWA Simulation Platform + community portal Support: access to repository + submission service Training: CGI training

Architecture of the Research Infrastructure

Access to the Research Infrastructure Local Research Facility Submission Service Data Infrastructure Access Compute Infrastructure Access Community Layer Service Layer Infra Access Layer App Marketplace Community Building App Development cluster cloud plug-in super computer plug-in grid plug-in EGI Federated Cloud EGI Grid Infrastructure PRACE cluster plug-in data archive B2x services data transfer protocols Information Service data storage data base data centre Remote Research Facility Training Remote Access Computational Chemists Experimental Chemists Science Gateway Data Service

Access to the Research Infrastructure community layer will offer social media type services allowing Experimental Chemists to run experiments on remotely available research facilities. will provide to access to simulation applications to run simulations for Computational Chemists. will also support training activities and community building. service layer will connect researchers to the research facilities and e-infrastructure resources using microservices managed by a service orchestrator. will provide data service that will connect Experimental and Computational Chemists through scientific data Experimental Chemists will use the data service to manage experimental data while Computational Chemists will run simulations through the submission service using the data service. infrastructure access layer computing infrastructure access service - will manage access to major computing resources such as cloud, cluster, grid and supercomputer. data infrastructure access service - will manage data using different types of data resources, such as data archives, databases, data collections, data storages using EUDAT B2xx and MASi services, and major data transfer protocols 40

Access to Cloud Resources in the Research Infrastructure (1) 41

Access to Cloud Resources in the Research Infrastructure (2)

Data Flows in the Research Infrastructure

Acknowledgements Prof. Sonja Herres-Pawlis Dr. Jens Kruger 44

Questions?