rvGAHP – Push-Based Job Submission Using Reverse SSH Connections

Slides:



Advertisements
Similar presentations
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
Advertisements

Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
A Computation Management Agent for Multi-Institutional Grids
Condor and GridShell How to Execute 1 Million Jobs on the Teragrid Jeffrey P. Gardner - PSC Edward Walker - TACC Miron Livney - U. Wisconsin Todd Tannenbaum.
1 High Performance Computing at SCEC Scott Callaghan Southern California Earthquake Center University of Southern California.
MultiJob PanDA Pilot Oleynik Danila 28/05/2015. Overview Initial PanDA pilot concept & HPC Motivation PanDA Pilot workflow at nutshell MultiJob Pilot.
CyberShake Study 14.2 Technical Readiness Review.
Design and Implementation of a Single System Image Operating System for High Performance Computing on Clusters Christine MORIN PARIS project-team, IRISA/INRIA.
Authors: Weiwei Chen, Ewa Deelman 9th International Conference on Parallel Processing and Applied Mathmatics 1.
NSF Geoinformatics Project (Sept 2012 – August 2014) Geoinformatics: Community Computational Platforms for Developing Three-Dimensional Models of Earth.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
BOSCO Architecture Derek Weitzel University of Nebraska – Lincoln.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
BOF: Megajobs Gracie: Grid Resource Virtualization and Customization Infrastructure How to execute hundreds of thousands tasks concurrently on distributed.
CyberShake Study 15.4 Technical Readiness Review.
CyberShake Study 2.3 Technical Readiness Review. Study re-versioning SCEC software uses year.month versioning Suggest renaming this study to 13.4.
Fig. 1. A wiring diagram for the SCEC computational pathways of earthquake system science (left) and large-scale calculations exemplifying each of the.
Pegasus: Running Large-Scale Scientific Workflows on the TeraGrid Ewa Deelman USC Information Sciences Institute
Faucets Queuing System Presented by, Sameer Kumar.
CyberShake Study 15.3 Science Readiness Review. Study 15.3 Scientific Goals Calculate a 1 Hz map of Southern California Produce meaningful 2 second results.
Scheduling MPI Workflow Applications on Computing Grids Juemin Zhang, Waleed Meleis, and David Kaeli Electrical and Computer Engineering Department, Northeastern.
Southern California Earthquake Center CyberShake Progress Update 3 November 2014 through 4 May 2015 UGMS May 2015 Meeting Philip Maechling SCEC IT Architect.
Experiences Running Seismic Hazard Workflows Scott Callaghan Southern California Earthquake Center University of Southern California SC13 Workflow BoF.
Southern California Earthquake Center SCEC Collaboratory for Interseismic Simulation and Modeling (CISM) Infrastructure Philip J. Maechling (SCEC) September.
Funded by the NSF OCI program grants OCI and OCI Mats Rynge, Gideon Juve, Karan Vahi, Gaurang Mehta, Ewa Deelman Information Sciences Institute,
Ensieea Rizwani An energy-efficient management mechanism for large-scale server clusters By: Zhenghua Xue, Dong, Ma, Fan, Mei 1.
1 1.Used AWP-ODC-GPU to run 10Hz Wave propagation simulation with rough fault rupture in half-space with and without small scale heterogeneities. 2.Used.
Rapid Centroid Moment Tensor (CMT) Inversion in 3D Earth Structure Model for Earthquakes in Southern California 1 En-Jui Lee, 1 Po Chen, 2 Thomas H. Jordan,
Southern California Earthquake Center CyberShake Progress Update November 3, 2014 – 4 May 2015 UGMS May 2015 Meeting Philip Maechling SCEC IT Architect.
Welcome to the CME Project Meeting 2013 Philip J. Maechling Information Technology Architect Southern California Earthquake Center.
HTCondor’s Grid Universe Jaime Frey Center for High Throughput Computing Department of Computer Sciences University of Wisconsin-Madison.
1 USC Information Sciences InstituteYolanda Gil AAAI-08 Tutorial July 13, 2008 Part IV Workflow Mapping and Execution in Pegasus (Thanks.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
Job submission overview Marco Mambelli – August OSG Summer Workshop TTU - Lubbock, TX THE UNIVERSITY OF CHICAGO.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
SCEC CyberShake on TG & OSG: Options and Experiments Allan Espinosa*°, Daniel S. Katz*, Michael Wilde*, Ian Foster*°,
Overview of Scientific Workflows: Why Use Them?
CyberShake Study 2.3 Readiness Review
CyberShake Study 16.9 Science Readiness Review
Simplify Your Science with Workflow Tools
Duncan MacMichael & Galen Deal CSS 534 – Autumn 2016
Dynamic Deployment of VO Specific Condor Scheduler using GT4
Example: Rapid Atmospheric Modeling System, ColoState U
Design rationale and status of the org.glite.overlay component
Simplify Your Science with Workflow Tools
Seismic Hazard Analysis Using Distributed Workflows
HTCondor and LSST Stephen Pietrowicz Senior Research Programmer National Center for Supercomputing Applications HTCondor Week May 2-5, 2017.
Scott Callaghan Southern California Earthquake Center
Architecture & System Overview
CyberShake Study 16.9 Discussion
High-F Project Southern California Earthquake Center
Philip J. Maechling (SCEC) September 13, 2015
Grid Canada Testbed using HEP applications
Haiyan Meng and Douglas Thain
University of Southern California
CyberShake Study 17.3 Science Readiness Review
Pegasus Workflows on XSEDE
Laura Bright David Maier Portland State University
Wide Area Workload Management Work Package DATAGRID project
CyberShake Study 2.2: Science Review Scott Callaghan 1.
Overview of Workflows: Why Use Them?
A General Approach to Real-time Workflow Monitoring
CyberShake Study 14.2 Science Readiness Review
Southern California Earthquake Center
Process/Code Migration and Cloning
Southern California Earthquake Center
CyberShake Study 18.8 Technical Readiness Review
CyberShake Study 2.2: Computational Review Scott Callaghan 1.
CyberShake Study 18.8 Planning
Presentation transcript:

rvGAHP – Push-Based Job Submission Using Reverse SSH Connections Scott Callaghan, Gideon Juve, Karan Vahi, Philip J. Maechling, Thomas H. Jordan and Ewa Deelman WORKS ‘17 November 13, 2017 Southern California Earthquake Center

Southern California Earthquake Center Motivation Scientific workflows continue to scale up Require automated resource provisioning on largest available clusters Two main approaches: Push-based: requests initiate from workflow queue Pull-based: requests initiate from remote resource Both have limitations Push-based: not viable on systems with two-factor authentication Pull-based: risk of high overhead with heterogeneous workloads rvGAHP: alternative approach Southern California Earthquake Center 4/27/2019

Southern California Earthquake Center Push-Based Approach Resource requests are on-demand, initiating from runtime manager Request waits in remote queue, starts up, runs workflow task Efficient use of remote resources: nodes only used when workflow tasks are being run Requires automated authentication to remote system Southern California Earthquake Center 4/27/2019

Southern California Earthquake Center Pull-Based Approach Resource requests initiate from remote resource (‘pilot job’) Resources advertised to workflow submission host for task scheduling Requires authentication from remote system to workflow host Can incur overhead Waiting for workflow task Mismatch of pilot job to task size and length Southern California Earthquake Center 4/27/2019

Southern California Earthquake Center Summary of Approaches Push-based Potential issues with automated authentication Little overhead: node-hours burned are used for workflow work Pull-based Authentication performed from remote system to workflow host Risk of high overhead, especially for heterogeneous workloads Would like solution with overhead characteristics of push-based, but with authentication flexibility of pull-based Southern California Earthquake Center 4/27/2019

Southern California Earthquake Center rvGAHP Overview Developed new solution, reverse GAHP (rvGAHP) Implementation of text-based GAHP protocol Push-based approach Does not require automated authentication to remote resource, just to workflow submission host Does require daemon running on remote resource Integrates cleanly with HTCondor Southern California Earthquake Center 4/27/2019

Southern California Earthquake Center rvGAHP Details rvgahp_server is started on remote resource rvgahp_server establishes ssh connection to workflow submission host and starts rvgahp_proxy When remote GAHP job is submitted, GridManager launches rvgahp_client rvgahp_client, via rvgahp_proxy, starts appropriate GAHP process on remote resource A socketpair is set up between remote GAHP process and ssh process, establishing communication between GridManager and remote GAHP process Another ssh process and rvgahp_proxy are started by the rvgahp_server (not shown) GridManager condor_ft-gahp File transfer fork() fork() rvgahp_client rvgahp_server PBS/LSF/SGE socketpair fork() qsub/qstat/qdel UNIX socket fork() rvgahp_proxy ssh batch_gahp ssh socketpair GAHP Workflow Submission Host Remote Resource Southern California Earthquake Center 4/27/2019

Southern California Earthquake Center Application Overview Probabilistic Seismic Hazard Analysis (PSHA) What will ground motion be in the next 50 years? Building engineers Insurance companies Disaster planners PSHA is performed by Assembling a list of earthquakes Calculating the shaking of each Combining shaking with probability SCEC CyberShake project – simulation-based approach to step 2 Southern California Earthquake Center 4/27/2019

Resource Requirements Computational requirements per site Stage Code Type Nodes Walltime Mesh Creation MPI CPU 240 0.4 hrs SGT Generation MPI GPU 800 1.0 Seismogram Synthesis 10 Other Sequential 1 0.1 – 2 200-400 sites required to complete study Parallelism dominated by SGT generation 2 jobs of 800 GPU nodes x ~1 hour Cost dominated by seismogram synthesis 240 nodes for ~10 hours CyberShake workflow schematic Southern California Earthquake Center 4/27/2019 Stage Code Type Output Data Titan Node Hours Mesh Creation MPI CPU 120 GB 100 SGT Generation MPI GPU 1500 GB 1490 Seismogram Synthesis 30 GB 2440 Other Sequential <1 GB 10 Total 1650 GB 4030

Southern California Earthquake Center CyberShake Workflows Each ‘run’ for a single site consists of a workflow Software stack: Pegasus-WMS to create workflow description, plan for execution HTCondor for runtime execution Globus GRAM for communication between workflow host and remote system CyberShake study consists of running hundreds of sites Site parallelism of 10-30 Able to split CyberShake studies to run on multiple resources Southern California Earthquake Center 4/27/2019

Current Solutions for CyberShake CyberShake requires 800-node GPU jobs 6.5x faster than CPU version Targeted NCSA Blue Waters and OLCF Titan For Blue Waters: push solution works well When jobs appear in HTCondor queue, GRAM submits them over the network to be translated and scheduled in Blue Waters queue For Titan: push solution unavailable due to authentication policy Two-factor with fob Proxies only issued for data transfer nodes Decided to try pull-based approach with HTCondor pilot jobs Southern California Earthquake Center 4/27/2019

CyberShake with Pull-Based Approach In 2015, used pull solution using HTCondor pilot jobs Pilot daemon running on login node queried HTCondor queue on workflow submission host for jobs When idle jobs found, pilot job submitted to Titan queue Because of variability in task sizes, 4 different sizes of pilot jobs Introduced complexity and possibility of errors Pull solution resulted in only 68% resource utilization Scheduling overhead Tradeoff between pilot jobs terminating before workflow tasks, and pilot jobs sitting idle Southern California Earthquake Center 4/27/2019

CyberShake with rvGAHP In 2017, performed 31-day CyberShake study on Titan using rvGAHP rvgahp_server daemon ran on Titan login node 13,334 jobs submitted through rvGAHP 10 minutes – 9 hours walltime 1 – 240 nodes wide 450,000 node-hours total Increased resource utilization from 68% to 97% Savings of ~130,000 node hours Decreased delay per job from 9.2 to 0.6 hrs Southern California Earthquake Center 4/27/2019

Southern California Earthquake Center Conclusion rvGAHP provides new approach for remote resource provisioning Reduced overhead of push-based approaches Runs on systems requiring two-factor authentication Requires little of application developers already using HTCondor Very effective for real-world seismic hazard workflow application Increases efficiency on current systems Opens up new systems for scientific workflows Southern California Earthquake Center 4/27/2019

Southern California Earthquake Center Thanks! Southern California Earthquake Center 4/27/2019 14