CENTRE FOR PARALLEL COMPUTING 8th IDGF Workshop Hannover, August 17 th 2011 International Desktop Grid Federation.

Slides:



Advertisements
Similar presentations
Introduction to Grid Application On-Boarding Nick Werstiuk
Advertisements

Legacy code support for commercial production Grids G.Terstyanszky, T. Kiss, T. Delaitre, S. Winter School of Informatics, University.
Performance Testing - Kanwalpreet Singh.
Investigating Approaches to Speeding Up Systems Biology Using BOINC-Based Desktop Grids Simon J E Taylor (1) Mohammadmersad Ghorbani (1) David Gilbert.
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
Copyright © 2007, SAS Institute Inc. All rights reserved. SAS and all other SAS Institute Inc. product or service names are registered trademarks or trademarks.
® IBM Software Group © 2006 IBM Corporation Rational Software France Object-Oriented Analysis and Design with UML2 and Rational Software Modeler 04. Other.
P-GRADE and WS-PGRADE portals supporting desktop grids and clouds Peter Kacsuk MTA SZTAKI
Visual Solution to High Performance Computing Computer and Automation Research Institute Laboratory of Parallel and Distributed Systems
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Extending a molecular docking tool to run simulations on clouds Damjan Temelkovski Dr. Tamas Kiss Dr. Gabor Terstyanszky University of Westminster.
Introduction : ‘Skoll: Distributed Continuous Quality Assurance’ Morimichi Nishigaki.
EXTENDING SCIENTIFIC WORKFLOW SYSTEMS TO SUPPORT MAPREDUCE BASED APPLICATIONS IN THE CLOUD Shashank Gugnani Tamas Kiss.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
©Ian Sommerville 2006Software Engineering, 8th edition. Chapter 12 Slide 1 Distributed Systems Architectures.
SZTAKI Desktop Grid – a Hierarchical Desktop Grid System P. Kacsuk, A. Marosi, J. Kovacs, Z. Balaton, G. Gombas, G. Vida, A. Kornafeld MTA SZTAKI
AHM /09/05 AHM 2005 Automatic Deployment and Interoperability of Grid Services G.Kecskemeti, Yonatan Zetuny, G.Terstyanszky,
Using the WS-PGRADE Portal in the ProSim Project Protein Molecule Simulation on the Grid Tamas Kiss, Gabor Testyanszky, Noam.
ExTASY 0.1 Beta Testing 1 st April 2015
Connecting OurGrid & GridSAM A Short Overview. Content Goals OurGrid: architecture overview OurGrid: short overview GridSAM: short overview GridSAM: example.
Kurochkin I.I., Prun A.I. Institute for systems analysis of RAS Centre for grid-technologies and distributed computing GRID-2012, Dubna, Russia july.
The EDGeS project receives Community research funding 1 EDGeS Bridge technology to interconnect EGEE and BOINC grids Peter Kacsuk MTA.
EXPOSE GOOGLE APP ENGINE AS TASKTRACKER NODES AND DATA NODES.
1 Developing domain specific gateways based on the WS- PGRADE/gUSE framework Peter Kacsuk MTA SZTAKI Start date: Duration:
20 October 2006Workflow Optimization in Distributed Environments Dynamic Workflow Management Using Performance Data David W. Walker, Yan Huang, Omer F.
G. Terstyanszky, T. Kukla, T. Kiss, S. Winter, J.: Centre for Parallel Computing School of Electronics and Computer Science, University of.
Protein Molecule Simulation on the Grid G-USE in ProSim Project Tamas Kiss Joint EGGE and EDGeS Summer School.
1 IDGF International Desktop Grid Federation How can you benefit from joining IDGF? Hannover, Peter Kacsuk, MTA SZTAKI, EDGI.
Parameter Sweep Workflows for Modelling Carbohydrate Recognition ProSim Project Tamas Kiss, Gabor Terstyanszky, Noam Weingarten.
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
The EDGeS project receives Community research funding 1 SG-DG Bridges Zoltán Farkas, MTA SZTAKI.
Issues in (Financial) High Performance Computing John Darlington Director Imperial College Internet Centre Fast Financial Algorithms and Computing 4th.
WS-PGRADE portal and its usage in the CancerGrid project M. Kozlovszky, P. Kacsuk Computer and Automation Research Institute of the Hungarian Academy of.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
Grid programming with components: an advanced COMPonent platform for an effective invisible grid © 2006 GridCOMP Grids Programming with components. An.
The EDGeS project receives Community research funding 1 EDGeS infrastructure for the EGEE user community Peter Kacsuk MTA SZTAKI.
P-GRADE and GEMLCA.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
The EDGeS project receives Community research funding 1 Porting Applications to the EDGeS Infrastructure A comparison of the available methods, APIs, and.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Services for advanced workflow programming.
1 IDGF International Desktop Grid Federation How can you benefit from joining IDGF? Lyon, Peter Kacsuk, MTA SZTAKI, EDGI is.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Enabling the use of e-Infrastructures with.
The EDGeS project receives Community research funding 1 Porting applications for a combined EGEE/Desktop Grid platform in the framework of the EDGeS infrastructure.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
11 Introduction to EDGI Peter Kacsuk, MTA SZTAKI Start date: Duration: 27 months EDGI.
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
The EDGeS project receives Community research funding 1 The EDGeS project: Enabling Desktop Grids for e-Science P. Kacsuk MTA SZTAKI.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
The EDGeS project receives Community research funding 1 Support services for desktop grids and service grids by the EDGeS project Tamas Kiss – University.
Supporting Big Data Processing via Science Gateways EGI CF 2015, November, Bari, Italy Dr Tamas Kiss, CloudSME Project Director University of Westminster,
The EDGI project receives Community research funding 1 Desktop Grid Infrastructure and User Support Services for EGI/NGI User Communities Tamas Kiss –
RI EGI-TF 2010, Tutorial Managing an EGEE/EGI Virtual Organisation (VO) with EDGES bridged Desktop Resources Tutorial Robert Lovas, MTA SZTAKI.
1 Support for parameter study applications in the P-GRADE Portal Gergely Sipos MTA SZTAKI (Hungarian Academy of Sciences)
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
SCI-BUS project Pre-kick-off meeting University of Westminster Centre for Parallel Computing Tamas Kiss, Stephen Winter, Gabor.
Introduction to Grid and Grid applications Peter Kacsuk MTA SZTAKI
Centre for Parallel Computing Tamas Kiss Centre for Parallel Computing A Distributed Rendering Service Tamas Kiss Centre for Parallel Computing Research.
DEGISCO Desktop Grids For International Scientific Collaboration Support for application porting Budapest, Robert Lovas, MTA SZTAKI DEGISCO.
Distributed Systems Architectures. Topics covered l Client-server architectures l Distributed object architectures l Inter-organisational computing.
Peter Kacsuk, Zoltan Farkas MTA SZTAKI
The EDGI (European Desktop Grid Initiative) infrastructure and its usage for the European Grid user communities József Kovács (MTA SZTAKI)
Grid Computing.
Grid Computing Colton Lewis.
Cloud Computing By P.Mahesh
University of Westminster Centre for Parallel Computing
Introduction to the SHIWA Simulation Platform EGI User Forum,
Presentation transcript:

CENTRE FOR PARALLEL COMPUTING 8th IDGF Workshop Hannover, August 17 th 2011 International Desktop Grid Federation

CENTRE FOR PARALLEL COMPUTING Experiences with the University of Westminster Desktop Grid S C Winter, T Kiss, G Terstyanszky, D Farkas, T Delaitre

CENTRE FOR PARALLEL COMPUTING Contents Introduction to Westminster Local Desktop Grid (WLDG) – Architecture, deployment management – EDGeS Application Development Methodology (EADM) Application examples Conclusions

CENTRE FOR PARALLEL COMPUTING Introduction to Westminster Local Desktop Grid (WLDG) New Cavendish St576 Marylebone Road559 Regent Street395 Wells Street 31 Little Titchfield St 66 Harrow Campus254

CENTRE FOR PARALLEL COMPUTING WLDG Environment DG Server on private University network Over 1500 client nodes on 6 different campuses Most machines are dual core, all running Windows Running SZTAKI Local Desktop Grid package Based on student laboratory PC’s – If not used by student  switch to DG mode – If no more work from DG server  shutdown (Green policy)

CENTRE FOR PARALLEL COMPUTING The DG Scenario

CENTRE FOR PARALLEL COMPUTING WLDG: ZENworks deployment BOINC clients installed automatically and maintained by specifically developed Novell ZENworks objects – MSI file has been created to generate a ZENworks object that installs the client software. – BOINC Client Install Shield Executable converted into an MSI package (/a switch on the BOINC Client executable) – BOINC client is part of the generic image installed on all lab PC’s throughout the University – Guaranteed that any newly purchased and installed PC automatically becomes part of the WLDG All clients registered under same user account

CENTRE FOR PARALLEL COMPUTING EDGeS Application Development Methodology (EADM) Generic methodology for DG application porting Motivation: Special focus required when porting/developing an application to a SG/DG platform Defines how the recommended software tools, eg. developed by EDGeS, can aid this process Supports iterative methods: – well-defined stages suggest a logical order – but (since in most cases process is non-linear) allows revisiting and revising results of previous stages, at any point

CENTRE FOR PARALLEL COMPUTING EADM – Defined Stages 1.Analysis of current application 2.Requirements analysis 3.Systems design 4.Detailed design 5.Implementation 6.Testing 7.Validation 8.Deployment 9.User support, maintenance & feedback

CENTRE FOR PARALLEL COMPUTING Application Examples Digital Alias-Free Signal Processing AutoDock Molecular Modelling

CENTRE FOR PARALLEL COMPUTING Digital Alias-Free Signal Processing (DASP) Users: Centre for Systems Analysis – University of Westminster Traditional DSP based on Uniform sampling – Suffers from aliasing Aim: Digital Alias-free Signal Processing (DASP) – One solution is Periodic Non-uniform Sampling (PNS) The DASP application designs PNS sequences Selection of optimal sampling sequence is computationally expensive process – A linear equation has to be solved and a large number of solutions (~10 10 ) compared. The analyses of the solutions are independent from each other  suitable for DG parallelisation

CENTRE FOR PARALLEL COMPUTING DASP - Parallelisation

CENTRE FOR PARALLEL COMPUTING DASP – Performance test results Period T (factor) Sequential DG worst DG median DG best # of work units Speedup (best case) # of nodes involved (median) 1813 min9 min7 min4 min hr 29 min 111 min 43 min20 min hr 40 min 5h 1min 3 hr 24 min 2 hr 31 min ~820 hrn/a 17 hr 54 min

CENTRE FOR PARALLEL COMPUTING DASP – Addressing the performance issues Inefficient load balancing – solutions of the equation should be grouped based on the execution time required to analyse individual solutions Inefficient work unit generation – some of the solutions should be divided into subtasks (more work units) – Limits to the possible speed-up User-community/application developers to consider redesigning the algorithm

CENTRE FOR PARALLEL COMPUTING AutoDock Molecular Modelling Users: Dept of Molecular & Applied Biosciences, UoW AutoDock: a suite of automated docking tools designed to predict how small molecules, such as substrates or drug candidates, bind to a receptor of known 3D structure application components: – AutoDock performs the docking of the ligand to a set of grids describing the target protein – AutoGrid pre-calculates these grids

CENTRE FOR PARALLEL COMPUTING Need for Parallelisation One run of AutoDock finishes in a reasonable time on a single PC However, thousands of scenarios have to be simulated and analysed to get stable and meaningful results. – AutoDock has to be run multiple times with the same input files but with random factors – Simulations runs are independent from each other – suitable for DG AutoGrid does not require Grid resources

CENTRE FOR PARALLEL COMPUTING AutoDock component workflow gpf file pdb file (ligand) pdb file (receptor) prepare_ligand4.py prepare_receptor4.py pdbqt file AUTOGRIDAUTODOCK map files Pamela dpf file AUTODOCK dlg files SCRIPT1SCRIPT2 best dlg files pdb file

CENTRE FOR PARALLEL COMPUTING Computational workflow in P-GRADE receptor.pdb ligand.pdb Autogrid executables, Scripts (uploaded by the developer, don’t change it) gpf descriptor file dpf descriptor file output pdb file Number of work units 1.The Generator job creates specified numbered of AutoDock jobs. 2.The AutoGrid job creates pdbqt files from the pdb files, runs the autogrid application and generates the map files. Zips them into an archive file. This archive will be the input of all AutoDock jobs. 3.The AutoDock jobs are running on the Desktop Grid. As output they provide dlg files. 4.The Collector job collects the dlg files. Takes the best results and concatenates them into a pdb file. dlg files

CENTRE FOR PARALLEL COMPUTING AutoDock – Performance test results

CENTRE FOR PARALLEL COMPUTING DG Drawbacks: The “Tail” Problem Jobs >> Nodes Jobs ≈ Nodes

CENTRE FOR PARALLEL COMPUTING Tackling the Tail Problem Augment the DG infrastructure with more reliable nodes, eg. service grid or cloud resources Redesign scheduler to detect tail and resubmit tardy tasks to SG or cloud

CENTRE FOR PARALLEL COMPUTING Cloudbursting: Indicative Results

CENTRE FOR PARALLEL COMPUTING AutoDock - Conclusions CygWin on Windows implementation inhibited performance – can be improved using (eg.) DG to EGEE bridge Cloudbursting AutoDock is black-box legacy application – source code not available – code-based improvement not possible

CENTRE FOR PARALLEL COMPUTING Further Applications Ultrasound Computer Tomography - Forschungszentrum Karlsruhe EMMIL – E-marketplace optimization - SZTAKI Anti-Cancer Drug Research (CancerGrid) - SZTAKI Integrator of Stochastic Differential Equations in Plasmas - BIFI Distributed Audio Retrieval - Cardiff University Cellular Automata based Laser Dynamics - University of Sevilla Radio Network Design – University of Extramadura An X-ray diffraction spectrum analysis - University of Extramadura DNA Sequence Comparison and Pattern Discovery - Erasmus Medical Center -PLINK - Analysis of genotype/phenotype data - Atos Origin 3D video rendering - University of Westminster

CENTRE FOR PARALLEL COMPUTING Conclusions – Performance Issues Performance enhancements – accrue from cyclical enterprise level hardware and software upgrades Are countered by performance degradation – arising from shared nature of resources Need to focus on robust performance measures – in face of random unpredictable run-time behaviours

CENTRE FOR PARALLEL COMPUTING Conclusions – Load Balancing Strategies Heterogranular workflows – Tasks can differ widely in size and run times – Performance prediction, based eg. on previous runs, can inform mapping (up to a point).. –.. but after this, may need to re-engineer code (white box only) –.. or consider offloading bottleneck tasks to reliable resources Homogranular workflows – Classic example: parameter sweep problem – Fine grain problems (#Tasks >> #Nodes) help smooth out the overall performance, but.. –.. tail problem can be significant (especially if #Tasks ≈ #Nodes) – Smart detection of delayed tasks coupled with speculative duplication

CENTRE FOR PARALLEL COMPUTING Conclusions – Deployment Issues Integration within enterprise desktop management environment has many advantages, eg. – PC’s and applications are continually upgraded – Hosts and licenses are “free” on the DG … but, also some drawbacks: – No direct control Typical environments can be slack and dirty Corporate objectives can override DG service objectives Examples: current UoW Win7 deployment, green agenda – Service relationship, based on trust DG bugs can easily damage trust relationship, if not caught quickly Example: recent GenWrapper bug – Non-dedicated resource Must give way to priority users, eg. students

CENTRE FOR PARALLEL COMPUTING The End Any questions?