1 Stork: State of the Art Tevfik Kosar Computer Sciences Department University of Wisconsin-Madison

Slides:



Advertisements
Similar presentations
Categories of I/O Devices
Advertisements

Operating System.
Lecture 11: Operating System Services. What is an Operating System? An operating system is an event driven program which acts as an interface between.
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & LSU Stork Data Scheduler: Current Status and Future Directions Sivakumar Kulasekaran.
Condor Project Computer Sciences Department University of Wisconsin-Madison Stork An Introduction Condor Week 2006 Milan.
A Computation Management Agent for Multi-Institutional Grids
04/14/2008CSCI 315 Operating Systems Design1 I/O Systems Notice: The slides for this lecture have been largely based on those accompanying the textbook.
I/O Hardware n Incredible variety of I/O devices n Common concepts: – Port – connection point to the computer – Bus (daisy chain or shared direct access)
I/O Systems CS 3100 I/O Hardware1. I/O Hardware Incredible variety of I/O devices Common concepts ◦Port ◦Bus (daisy chain or shared direct access) ◦Controller.
The Difficulties of Distributed Data Douglas Thain Condor Project University of Wisconsin
1 Exploring Data Reliability Tradeoffs in Replicated Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh Matei Ripeanu.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
GridFTP Guy Warner, NeSC Training.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Harnessing the Capacity of Computational.
Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison Condor and Grid Challenges.
Workflow Management in Condor Gökay Gökçay. DAGMan Meta-Scheduler The Directed Acyclic Graph Manager (DAGMan) is a meta-scheduler for Condor jobs. DAGMan.
1 Programming Languages Tevfik Koşar Lecture - I January 17 th, 2006.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
ALICE data access WLCG data WG revival 4 October 2013.
July Lecture 4: Grid Data Management1 Grid Data Management.
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA.
I/O Systems I/O Hardware Application I/O Interface
Enabling Data Intensive Science with PetaShare Tevfik Kosar Center for Computation & Technology Louisiana State University April 6, 2007.
◦ What is an Operating System? What is an Operating System? ◦ Operating System Objectives Operating System Objectives ◦ Services Provided by the Operating.
Why GridFTP? l Performance u Parallel TCP streams, optimal TCP buffer u Non TCP protocol such as UDT u Order of magnitude greater l Cluster-to-cluster.
File and Object Replication in Data Grids Chin-Yi Tsai.
Reliable Data Movement using Globus GridFTP and RFT: New Developments in 2008 John Bresnahan Michael Link Raj Kettimuthu Argonne National Laboratory and.
INVITATION TO COMPUTER SCIENCE, JAVA VERSION, THIRD EDITION Chapter 6: An Introduction to System Software and Virtual Machines.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
Parallel Programming on the SGI Origin2000 With thanks to Igor Zacharov / Benoit Marchand, SGI Taub Computer Center Technion Moshe Goldberg,
Peter F. Couvares (based on material from Tevfik Kosar, Nick LeRoy, and Jeff Weber) Associate Researcher, Condor Team Computer Sciences Department University.
Condor Week Summary March 14-16, 2005 Madison, Wisconsin.
STORK: Making Data Placement a First Class Citizen in the Grid Tevfik Kosar and Miron Livny University of Wisconsin-Madison March 25 th, 2004 Tokyo, Japan.
Chapter 13: I/O Systems. 13.2/34 Chapter 13: I/O Systems I/O Hardware Application I/O Interface Kernel I/O Subsystem Transforming I/O Requests to Hardware.
George Kola Computer Sciences Department University of Wisconsin-Madison DiskRouter: A Mechanism for High.
Cluster 2004 San Diego, CA A Client-centric Grid Knowledgebase George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison September 23 rd,
July 11-15, 2005Lecture3: Grid Job Management1 Grid Compute Resources and Job Management.
Tevfik Kosar Computer Sciences Department University of Wisconsin-Madison Managing and Scheduling Data.
Storage Research Meets The Grid Remzi Arpaci-Dusseau.
STORK: Making Data Placement a First Class Citizen in the Grid Tevfik Kosar University of Wisconsin-Madison May 25 th, 2004 CERN.
O PERATING S YSTEM. What is an Operating System? An operating system is an event driven program which acts as an interface between a user of a computer,
A Fully Automated Fault- tolerant System for Distributed Video Processing and Off­site Replication George Kola, Tevfik Kosar and Miron Livny University.
Computer Studies (AL) I/O Management. Reference Silberschatz, Galvin, Gagne “Operating System Concepts 6 th edition”, 2003, Wiley Stallings, “Operating.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Bulk Data Transfer Activities We regard data transfers as “first class citizens,” just like computational jobs. We have transferred ~3 TB of DPOSS data.
Silberschatz, Galvin, and Gagne  Applied Operating System Concepts Module 12: I/O Systems I/O hardwared Application I/O Interface Kernel I/O.
GridFTP Guy Warner, NeSC Training Team.
George Kola Computer Sciences Department University of Wisconsin-Madison Data Pipelines: Real Life Fully.
Reliable and Efficient Grid Data Placement using Stork and DiskRouter Tevfik Kosar University of Wisconsin-Madison April 15 th, 2004.
Major OS Components CS 416: Operating Systems Design, Spring 2001 Department of Computer Science Rutgers University
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Run-time Adaptation of Grid Data Placement Jobs George Kola, Tevfik Kosar and Miron Livny Condor Project, University of Wisconsin.
Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison Job Delegation and Planning.
Condor DAGMan: Managing Job Dependencies with Condor
Module 12: I/O Systems I/O hardware Application I/O Interface
Chapter 13: I/O Systems Modified by Dr. Neerja Mhaskar for CS 3SH3.
Operating Systems (CS 340 D)
US CMS Testbed.
Operating System Concepts
13: I/O Systems I/O hardwared Application I/O Interface
CS703 - Advanced Operating Systems
Outline Problem DiskRouter Overview Details Real life DiskRouters
Outline What users want ? Data pipeline overview
STORK: A Scheduler for Data Placement Activities in Grid
Lecture 15 Reading: Bacon 7.6, 7.7
Language Processors Application Domain – ideas concerning the behavior of a software. Execution Domain – Ideas implemented in Computer System. Semantic.
Chapter 13: I/O Systems I/O Hardware Application I/O Interface
Chapter 13: I/O Systems I/O Hardware Application I/O Interface
Module 12: I/O Systems I/O hardwared Application I/O Interface
Presentation transcript:

1 Stork: State of the Art Tevfik Kosar Computer Sciences Department University of Wisconsin-Madison

2 The Imminent Data “deluge” Moore’s Law outpaced by growth of scientific data! Exponential growth of scientific data –2000 : ~0.5 Petabyte –2005 : ~10 Petabytes –2010 : ~100 Petabytes –2015 : ~1000 Petabytes “I am terrified by terabytes” -- Anonymous “I am petrified by petabytes” -- Jim Gray

3 Bioinformatics: BLAST High Energy Physics: LHC Astronomy: LSST 2MASS SDSS DPOSS GSC-II WFCAM VISTA NVSS FIRST GALEX ROSAT OGLE... LSST 2MASS SDSS DPOSS GSC-II WFCAM VISTA NVSS FIRST GALEX ROSAT OGLE... Educational Technology: WCER EVP 500 TB/year 2-3 PB/year 11 PB/year 20 TB - 1 PB/year

4 How to access and process distributed data? TB PB

5 BUS CPU DISK MEMORY I/O PROCESSOR HARDWARE LEVEL CONTROLLER DMA I/O Management in the History HARDWARE LEVEL BUS CPU DISK MEMORY I/O PROCESSOR CONTROLLER DMA

6 HARDWARE LEVEL BUS CPU DISK MEMORY I/O PROCESSOR CONTROLLER DMA OPERATING SYSTEMS LEVEL I/O CONTROL SYSTEM CPU SCHEDULER I/O SCHEDULER I/O SUBSYSTEM OPERATING SYSTEMS LEVEL I/O CONTROL SYSTEM CPU SCHEDULER I/O SCHEDULER I/O SUBSYSTEM I/O Management in the History

7 HARDWARE LEVEL BUS CPU DISK MEMORY I/O PROCESSOR CONTROLLER DMA OPERATING SYSTEMS LEVEL I/O CONTROL SYSTEM CPU SCHEDULER I/O SCHEDULER I/O SUBSYSTEM DISTRIBUTED SYSTEMS LEVEL BATCH SCHEDULERS I/O Management in the History

8 HARDWARE LEVEL BUS CPU DISK MEMORY I/O PROCESSOR CONTROLLER DMA OPERATING SYSTEMS LEVEL I/O CONTROL SYSTEM CPU SCHEDULER I/O SCHEDULER I/O SUBSYSTEM DISTRIBUTED SYSTEMS LEVEL BATCH SCHEDULERS DATA PLACEMENT SCHEDULER I/O Management in the History

9 Stage-in Execute job jStage-out Stage-in Execute job jStage-out Release input space Release output space Allocate space for input & output data JOB iJOB kJOB i JOB k Individual Jobs JOB j get put Stage-in Stage-out Stage-in Stage-out Release input space Release output space Allocate space for input & output data Compute Jobs Data placement Jobs Release input spaceRelease output space Allocate space for input & output data

10 Separation of Jobs Data A A.stork Data B B.stork Job C C.condor ….. Parent A child B Parent B child C Parent C child D, E ….. DAG specification Workflow Manager A C B D E F Compute Job Queue C Data Job Queue E

11 Stork: Data Placement Scheduler First scheduler specialized for data movement/placement. De-couples data placement from computation. Understands the characteristics and semantics of data placement jobs. Can make smart scheduling decisions for reliable and efficient data placement. A prototype is already implemented and deployed at several sites. Now distributed with Condor Developers Release v

12 Support for Heterogeneity Provides uniform access to different data storage systems and transfer protocols. Acts as an IOCS for distributed systems. Multilevel Policy Support Protocol translation: using Stork Disk Cacheusing Stork Memory Buffer [ Type = “Transfer”; Src_Url = “srb://ghidorac.sdsc.edu/kosart.condor/x.dat”; Dest_Url = “nest://turkey.cs.wisc.edu/kosart/x.dat”; …… Max_Retry = 10; Restart_in = “2 hours”; ] [ICDCS’04]

13 Dynamic Protocol Selection [ dap_type = “transfer”; src_url = “drouter://slic04.sdsc.edu/tmp/test.dat”; dest_url = “drouter://quest2.ncsa.uiuc.edu/tmp/test.dat”; alt_protocols = “gsiftp-gsiftp, nest-nest”; or: src_url = “any://slic04.sdsc.edu/tmp/test.dat”; dest_url = “any://quest2.ncsa.uiuc.edu/tmp/test.dat”; ] DiskRouter crashes DiskRouter resumes Traditional Scheduler: 48 Mb/s Using Stork: 72 Mb/s [ICDCS’04]

14 Run-time Auto-tuning [ link = “slic04.sdsc.edu – quest2.ncsa.uiuc.edu”; protocol = “gsiftp”; bs = 1024KB;// I/O block size tcp_bs = 1024KB;// TCP buffer size p= 4; // number of parallel streams ] Before Tuning: parallelism = 1 block_size = 1 MB tcp_bs = 64 KB After Tuning: parallelism = 4 block_size = 1 MB tcp_bs = 256 KB Traditional Scheduler (without tuning) 0.5 MB/s Using Stork (with tuning) 10 MB/s [AGridM’03] GridFTP

15 Controlling Throughput Increasing concurrency/parallelism does not always in crease transfer rate Effect on local area and wide are is different Concurrency and parallelism have slightly different impacts on transfer rate Wide Area Local Area [Europar’04]

16 Controlling CPU Utilization Concurrency and parallelism have totally opposite impacts on CPU utilization at the server side. Client Server [Europar’04]

17 Detecting and Classifying Failures Check DNS Server Check DNS Check Network S F Check Protocol Check Host DNS Server error No DNS entry Network Outage Host Down Protocol Unavailable F F F F S S S Test Transfer Transfer Failed S F Check Credentials Not Authenticated F Check File Source File Does Not Exist F S S S Transient Permanent Transient Permanent POLICIES [Grid’04]

min 99.7% Detecting Hanging Transfers Collecting job execution time statistics –Fit a distribution Detect and avoid –black holes –hanging transfers Eg. for normal distribution: 99.7% of job execution times should lie between [(avg-3*stdev), (avg+3*stdev)] [Cluster’04]

19 Stork can also: Allocate/de-allocate (optical) network links Allocate/de-allocate storage space Register/un-register files to Meta Data Catalog Locate physical location of a logical file name Control concurrency levels on storage servers You can refer to [ICDCS’04][JPDC’05][AGridM’03]

20 Apply to Real Life Applications

21 DPOSS Astronomy Pipeline

22 UniTree not responding Diskrouter reconfigured and restarted SDSC cache reboot & UW CS Network outage Software problem Failure Recovery

23 End-to-end Processing of 3 TB DPOSS Astronomy Data Traditional Scheduler: 2 weeks Using Stork: 6 days

24 Summary Stork provides solutions for the data placement needs of the Grid community. It is ready to fly! Now distributed with Condor developers release v All basic features you will need are included in the initial release. More features coming in the future releases.

25 Thank you for listening.. Questions?