Run-time Adaptation of Grid Data Placement Jobs George Kola, Tevfik Kosar and Miron Livny Condor Project, University of Wisconsin.

Slides:



Advertisements
Similar presentations
Network Resource Broker for IPTV in Cloud Computing Lei Liang, Dan He University of Surrey, UK OGF 27, G2C Workshop 15 Oct 2009 Banff,
Advertisements

Cross-site data transfer on TeraGrid using GridFTP TeraGrid06 Institute User Introduction to TeraGrid June 12 th by Krishna Muriki
Starfish: A Self-tuning System for Big Data Analytics.
Windows® Deployment Services
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
Top Causes for Poor Application Performance Case Studies Mike Canney.
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & LSU Stork Data Scheduler: Current Status and Future Directions Sivakumar Kulasekaran.
Esma Yildirim Department of Computer Engineering Fatih University Istanbul, Turkey DATACLOUD 2013.
Condor Project Computer Sciences Department University of Wisconsin-Madison Stork An Introduction Condor Week 2006 Milan.
6th Biennial Ptolemy Miniconference Berkeley, CA May 12, 2005 Distributed Computing in Kepler Ilkay Altintas Lead, Scientific Workflow Automation Technologies.
Reliable I/O on the Grid Douglas Thain and Miron Livny Condor Project University of Wisconsin.
MULTICOMPUTER 1. MULTICOMPUTER, YANG DIPELAJARI Multiprocessors vs multicomputers Interconnection topologies Switching schemes Communication with messages.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Resource Management Reading: “A Resource Management Architecture for Metacomputing Systems”
GridFTP Guy Warner, NeSC Training.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Harnessing the Capacity of Computational.
Workflow Management in Condor Gökay Gökçay. DAGMan Meta-Scheduler The Directed Acyclic Graph Manager (DAGMan) is a meta-scheduler for Condor jobs. DAGMan.
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA.
Building a Parallel File System Simulator E Molina-Estolano, C Maltzahn, etc. UCSC Lab, UC Santa Cruz. Published in Journal of Physics, 2009.
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
Development Timelines Ken Kennedy Andrew Chien Keith Cooper Ian Foster John Mellor-Curmmey Dan Reed.
Globus GridFTP and RFT: An Overview and New Features Raj Kettimuthu Argonne National Laboratory and The University of Chicago.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Peter F. Couvares (based on material from Tevfik Kosar, Nick LeRoy, and Jeff Weber) Associate Researcher, Condor Team Computer Sciences Department University.
STORK: Making Data Placement a First Class Citizen in the Grid Tevfik Kosar and Miron Livny University of Wisconsin-Madison March 25 th, 2004 Tokyo, Japan.
Chapter 8-2 : Multicomputers Multiprocessors vs multicomputers Multiprocessors vs multicomputers Interconnection topologies Interconnection topologies.
George Kola Computer Sciences Department University of Wisconsin-Madison DiskRouter: A Mechanism for High.
Cluster 2004 San Diego, CA A Client-centric Grid Knowledgebase George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison September 23 rd,
LEGS: A WSRF Service to Estimate Latency between Arbitrary Hosts on the Internet R.Vijayprasanth 1, R. Kavithaa 2,3 and Raj Kettimuthu 2,3 1 Coimbatore.
Tevfik Kosar Computer Sciences Department University of Wisconsin-Madison Managing and Scheduling Data.
Flexibility, Manageability and Performance in a Grid Storage Appliance John Bent, Venkateshwaran Venkataramani, Nick Leroy, Alain Roy, Joseph Stanley,
Storage Research Meets The Grid Remzi Arpaci-Dusseau.
STORK: Making Data Placement a First Class Citizen in the Grid Tevfik Kosar University of Wisconsin-Madison May 25 th, 2004 CERN.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
University of Maryland Towards Automated Tuning of Parallel Programs Jeffrey K. Hollingsworth Department of Computer Science University.
GridFTP Richard Hopkins
Hepix LAL April 2001 An alternative to ftp : bbftp Gilles Farrache In2p3 Computing Center
SAN DIEGO SUPERCOMPUTER CENTER Inca Control Infrastructure Shava Smallen Inca Workshop September 4, 2008.
1 Stork: State of the Art Tevfik Kosar Computer Sciences Department University of Wisconsin-Madison
A Fully Automated Fault- tolerant System for Distributed Video Processing and Off­site Replication George Kola, Tevfik Kosar and Miron Livny University.
Scott Koranda, UWM & NCSA 14 January 2016www.griphyn.org Lightweight Data Replicator Scott Koranda University of Wisconsin-Milwaukee & National Center.
Aneka Cloud ApplicationPlatform. Introduction Aneka consists of a scalable cloud middleware that can be deployed on top of heterogeneous computing resources.
GNEW2004 CERN March 2004 R. Hughes-Jones Manchester 1 Lessons Learned in Grid Networking or How do we get end-2-end performance to Real Users ? Richard.
Bulk Data Transfer Activities We regard data transfers as “first class citizens,” just like computational jobs. We have transferred ~3 TB of DPOSS data.
Douglas Thain, John Bent Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau, Miron Livny Computer Sciences Department, UW-Madison Gathering at the Well: Creating.
GridFTP Guy Warner, NeSC Training Team.
1 GridFTP and SRB Guy Warner Training, Outreach and Education Team, Edinburgh e-Science.
George Kola Computer Sciences Department University of Wisconsin-Madison Data Pipelines: Real Life Fully.
Reliable and Efficient Grid Data Placement using Stork and DiskRouter Tevfik Kosar University of Wisconsin-Madison April 15 th, 2004.
LACSI 2002, slide 1 Performance Prediction for Simple CPU and Network Sharing Shreenivasa Venkataramaiah Jaspal Subhlok University of Houston LACSI Symposium.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Mario Reale – GARR NetJobs: Network Monitoring Using Grid Jobs.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
BDTS and Its Evaluation on IGTMD link C. Chen, S. Soudan, M. Pasin, B. Chen, D. Divakaran, P. Primet CC-IN2P3, LIP ENS-Lyon
Campana (CERN-IT/SDC), McKee (Michigan) 16 October 2013 Deployment of a WLCG network monitoring infrastructure based on the perfSONAR-PS technology.
Fermilab Cal Tech Lambda Station High-Performance Network Research PI Meeting BNL Phil DeMar September 29, 2005.
Scott Koranda, UWM & NCSA 20 November 2016www.griphyn.org Lightweight Replication of Heavyweight Data Scott Koranda University of Wisconsin-Milwaukee &
Introduction to Distributed Platforms
Example: Rapid Atmospheric Modeling System, ColoState U
Condor – A Hunter of Idle Workstation
R. Hughes-Jones Manchester
Gregory Kesden, CSE-291 (Storage Systems) Fall 2017
Gregory Kesden, CSE-291 (Cloud Computing) Fall 2016
A Framework for Automatic Resource and Accuracy Management in A Cloud Environment Smita Vijayakumar.
Supporting Fault-Tolerance in Streaming Grid Applications
湖南大学-信息科学与工程学院-计算机与科学系
Outline Problem DiskRouter Overview Details Real life DiskRouters
Outline What users want ? Data pipeline overview
STORK: A Scheduler for Data Placement Activities in Grid
Presentation transcript:

Run-time Adaptation of Grid Data Placement Jobs George Kola, Tevfik Kosar and Miron Livny Condor Project, University of Wisconsin

Run-time Adaptation of Grid Data Placement Jobs 2Introduction Grid presents a continuously changing environment Data intensive applications are being run on the grid Data intensive applications have two parts –Data placement part –Computation part

Run-time Adaptation of Grid Data Placement Jobs 3 Data Placement Stage in data Stage out data Compute Data placement Data placement encompasses data transfer, staging, replication, data positioning, space allocation and de-allocation A Data Intensive Application

Run-time Adaptation of Grid Data Placement Jobs 4Problems Insufficient automation –Failures –No tuning – tuning is difficult ! Lack of adaptation to changing environment –Failure of one protocol while others are functioning –Changing network characteristics

Run-time Adaptation of Grid Data Placement Jobs 5 Current Approach Fedex Hand tuning Network Weather Service –Not useful for high-bandwidth, high- latency networks TCP Auto-tuning –16-bit windows size and window scale option limitations

Run-time Adaptation of Grid Data Placement Jobs 6 Our Approach Full automation Continuously monitor environment characteristics Perform tuning whenever characteristics change Ability to dynamically and automatically choose an appropriate protocol Ability to switch to alternate protocol incase of failure

Run-time Adaptation of Grid Data Placement Jobs 7 Memory Profiler Disk Profiler Network Profiler Disk Profiler Memory Profiler Monitoring Host 2 Tuning Infra- structure Monitoring Host 1 Network Parameters Data Placement Scheduler Disk Parameters Memory Parameters Disk Parameters Memory Parameters Data Transfer Parameters The Big Picture

Run-time Adaptation of Grid Data Placement Jobs 8 Monitoring Host 2 Tuning Infra- structure Memory Profiler Disk Profiler Network Profiler Disk Profiler Memory Profiler Monitoring Host 1 Network Parameters Data Placement Scheduler Disk Parameters Memory Parameters Disk Parameters Memory Parameters Data Transfer Parameters The Big Picture

Run-time Adaptation of Grid Data Placement Jobs 9Profilers Memory Profiler –Optimal memory block-size and incremental block- size Disk Profiler –Optimal disk block-size and incremental block-size Network Profiler –Determines bandwidth, latency and the number of hops between a given pair of hosts –Uses pathrate, traceroute and diskrouter bandwidth test tool

Run-time Adaptation of Grid Data Placement Jobs 10 Memory Profiler Disk Profiler Network Profiler Disk Profiler Memory Profiler Monitoring Host 2 Tuning Infra- structure Monitoring Host 1 Network Parameters Data Placement Scheduler Disk Parameters Memory Parameters Disk Parameters Memory Parameters Data Transfer Parameters The Big Picture Parameter Tuner 1

Run-time Adaptation of Grid Data Placement Jobs 11 Parameter Tuner Generates optimal parameters for data transfer between a given pair of hosts Calculates TCP buffer size as the bandwidth-delay product Calculates the optimal disk buffer size based on TCP buffer size Uses a heuristic to calculate the number of tcp streams No of streams = 1+ No of hops with latency > 10ms Rounded to an even number

Run-time Adaptation of Grid Data Placement Jobs 12 Memory Profiler Disk Profiler Network Profiler Disk Profiler Memory Profiler Monitoring Host 2 Tuning Infra- structure Monitoring Host 1 Network Parameters Data Placement Scheduler Disk Parameters Memory Parameters Disk Parameters Memory Parameters Data Transfer Parameters The Big Picture

Run-time Adaptation of Grid Data Placement Jobs 13 Data Placement Scheduler Data placement is a real-job A meta-scheduler (e.g. DAGMan) is used to co-ordinate data placement and computation Sample data placement job [ dap_type = “transfer” ; src_url = “diskrouter://slic04.sdsc.edu/s/s1” ; dest_url=“diskrouter://quest2.ncsa.uiuc.edu/d/d1”;]

Run-time Adaptation of Grid Data Placement Jobs 14 Data Placement Scheduler Used Stork, a prototype data placement scheduler Tuned parameters are fed to Stork Stork uses the tuned parameters to adapt data placement jobs

Run-time Adaptation of Grid Data Placement Jobs 15Implementation Profilers are run as remote batch jobs on respective hosts Parameter tuner is also a batch job An instance of parameter tuner is run for every pair of nodes involved in data transfer Monitoring and tuning infrastructure is coordinated by DAGMan

Run-time Adaptation of Grid Data Placement Jobs 16 Disk Profiler Parameter Tuner Memory Profiler Network Profiler Lightweight Network Profiler Parameter Tuner This part executes periodically This part executes at startup and whenever requested Coordinating DAG

Run-time Adaptation of Grid Data Placement Jobs 17Scalability There is no centralized server Parameter tuner can be run on any computation resource Profiler data is 100s of bytes per host There can be multiple data placement schedulers

Run-time Adaptation of Grid Data Placement Jobs 18 Memory Profiler Disk Profiler Network Profiler Disk Profiler Memory Profiler Monitoring Host 2 Tuning Infra- structure Monitoring Host 1 Network Parameters Data Placement Scheduler Disk Parameters Memory Parameters Disk Parameters Memory Parameters Data Transfer Parameters The Big Picture

Run-time Adaptation of Grid Data Placement Jobs 19Scalability There is no centralized server Parameter tuner can be run on any computation resource Profiler data is 100s of bytes per host There can be multiple data placement schedulers

Run-time Adaptation of Grid Data Placement Jobs 20 Dynamic Protocol Selection Determines the protocols available on the different hosts Creates a list of hosts and protocols in ClassAd format e.g. [ hostname=“quest2.ncsa.uiuc.edu” ; protocols=“diskrouter,gridftp,ftp” ] [ hostname=“nostos.cs.wisc.edu” ; protocols=“gridftp,ftp,http” ]

Run-time Adaptation of Grid Data Placement Jobs 21 Dynamic Protocol Selection [ dap_type = “transfer” ; src_url = “any://slic04.sdsc.edu/s/data1” ; dest_url=“any://quest2.ncsa.uiuc.edu/d/data1” ; ] Stork determines an appropriate protocol to use for the transfer In case of failure, Stork chooses another protocol

Run-time Adaptation of Grid Data Placement Jobs 22 Alternate Protocol Fallback [ dap_type = “transfer”; src_url = “diskrouter://slic04.sdsc.edu/s/data1”; dest_url=“diskrouter://quest2.ncsa.uiuc.edu/d/data1”; alt_protocols=“nest-nest, gsiftp-gsiftp”; ] In case of diskrouter failure, Stork will switch to other protocols in the order specified

Run-time Adaptation of Grid Data Placement Jobs 23 Real World Experiment DPOSS data had to be transferred from SDSC located in San Diego to NCSA located at Chicago SDSC (slic04.sdsc.edu) NCSA (quest2.ncsa.uiuc.edu) Transfer

Run-time Adaptation of Grid Data Placement Jobs 24 SDSC (slic04.sdsc.edu) GridFTP Management Site ( skywalker.cs.wisc.edu) NCSA (quest2.ncsa.uiuc.edu) Control flow Data flow StarLight (ncdm13.sl.startap.net) WAN DiskRouter Real World Experiment

Run-time Adaptation of Grid Data Placement Jobs 25 Data Transfer from SDSC to NCSA using Run-time Protocol Auto-tuning Time Transfer Rate (MB/s) Auto-tuning turned on Network outage

Run-time Adaptation of Grid Data Placement Jobs 26 Parameter Tuning Parameter Before auto- tuning After auto- tuning Parallelism 1 TCP stream 4 TCP streams Block size 1 MB TCP buffer size 64 KB 256 KB

Run-time Adaptation of Grid Data Placement Jobs 27 Testing Alternate Protocol Fall-back [ dap_type = “transfer”; src_url = “diskrouter://slic04.sdsc.edu/s/data1”; dest_url=“diskrouter://quest2.ncsa.uiuc.e du/d/data1”; alt_protocols=“nest-nest, gsiftp-gsiftp”; ]

Time Transfer Rate (MB/s) DiskRouter server killed DiskRouter server restarted 1234 Testing Alternate Protocol Fall-back

Run-time Adaptation of Grid Data Placement Jobs 29Conclusion Run-time adaptation has a significant impact (20 times improvement in our test case) The profiling data has the potential to be used for data mining Network misconfigurations Network outages Dynamic protocol selection and alternate protocol fall-back increase resilence and improve overall throughput

Run-time Adaptation of Grid Data Placement Jobs 30 Questions ? For more information you can contact: George Kola : Kola : Tevfik Kosar: Kosar: Project web pages: Stork: DiskRouter: