Information and Scheduling: What's available and how does it change Jennifer M. Schopf Argonne National Lab.

Slides:



Advertisements
Similar presentations
Network Weather Service Sathish Vadhiyar Sources / Credits: NWS web site: NWS papers.
Advertisements

Dead Reckoning Objectives – –Understand what is meant by the term dead reckoning. –Realize the two major components of a dead reckoning protocol. –Be capable.
Lincoln University Canterbury New Zealand Evaluating the Parallel Performance of a Heterogeneous System Elizabeth Post Hendrik Goosen formerly of Department.
Receiver-driven Layered Multicast S. McCanne, V. Jacobsen and M. Vetterli SIGCOMM 1996.
GridFlow: Workflow Management for Grid Computing Kavita Shinde.
Performance Prediction Engineering Francine Berman U. C. San Diego Rich Wolski U. C. San Diego and University of Tennessee This presentation will probably.
AQM for Congestion Control1 A Study of Active Queue Management for Congestion Control Victor Firoiu Marty Borden.
A Grid Resource Broker Supporting Advance Reservations and Benchmark- Based Resource Selection Erik Elmroth and Johan Tordsson Reporter : S.Y.Chen.
Computational Astrophysics: Methodology 1.Identify astrophysical problem 2.Write down corresponding equations 3.Identify numerical algorithm 4.Find a computer.
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
The Network Weather Service: A Distributed Resource Performance Forecasting Service for Metacomputing, Rich Wolski, Neil Spring, and Jim Hayes, Journal.
Performance Evaluation
1 Introduction to Load Balancing: l Definition of Distributed systems. Collection of independent loosely coupled computing resources. l Load Balancing.
1 Lecture 24: Interconnection Networks Topics: communication latency, centralized and decentralized switches (Sections 8.1 – 8.5)
Characterizing and Predicting TCP Throughput on the Wide Area Network Dong Lu, Yi Qiao, Peter Dinda, Fabian Bustamante Department of Computer Science Northwestern.
The Effects of Systemic Packets Loss on Aggregate TCP Flows Thomas J. Hacker May 8, 2002 Internet 2 Member Meeting.
1 Reading Report 9 Yin Chen 29 Mar 2004 Reference: Multivariate Resource Performance Forecasting in the Network Weather Service, Martin Swany and Rich.
1 Enabling Large Scale Network Simulation with 100 Million Nodes using Grid Infrastructure Hiroyuki Ohsaki Graduate School of Information Sci. & Tech.
1 Interconnects Shared address space and message passing computers can be constructed by connecting processors and memory unit using a variety of interconnection.
Lecture 2 Process Concepts, Performance Measures and Evaluation Techniques.
10/19/2015Erkay Savas1 Performance Computer Architecture – CS401 Erkay Savas Sabanci University.
Scheduling policies for real- time embedded systems.
1 Nasser Alsaedi. The ultimate goal for any computer system design are reliable execution of task and on time delivery of service. To increase system.
11 Experimental and Analytical Evaluation of Available Bandwidth Estimation Tools Cesar D. Guerrero and Miguel A. Labrador Department of Computer Science.
Performance.
Time-Series Forecasting Overview Moving Averages Exponential Smoothing Seasonality.
1 Grid Scheduling Cécile Germain-Renaud. 2 Scheduling Job –A computation to run on a machine –Possibly with network access e.g. input/output file (coarse.
A High-Throughput Path Metric for Multi- Hop Wireless Routing Douglas S. J. De Couto, Daniel Aguayo, John Bicket, Robert Morris MIT Computer Science and.
Embedded System Lab 김해천 Thread and Memory Placement on NUMA Systems: Asymmetry Matters.
Resource Predictors in HEP Applications John Huth, Harvard Sebastian Grinstein, Harvard Peter Hurst, Harvard Jennifer M. Schopf, ANL/NeSC.
1 CS/COE0447 Computer Organization & Assembly Language CHAPTER 4 Assessing and Understanding Performance.
1 On Dynamic Parallelism Adjustment Mechanism for Data Transfer Protocol GridFTP Takeshi Itou, Hiroyuki Ohsaki Graduate School of Information Sci. & Tech.
Lecture 4 TTH 03:30AM-04:45PM Dr. Jianjun Hu CSCE569 Parallel Computing University of South Carolina Department of.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Chapter 5: Process Scheduling.
Measuring the Capacity of a Web Server USENIX Sympo. on Internet Tech. and Sys. ‘ Koo-Min Ahn.
CS 484 Load Balancing. Goal: All processors working all the time Efficiency of 1 Distribute the load (work) to meet the goal Two types of load balancing.
1 Real-Time Scheduling. 2Today Operating System task scheduling –Traditional (non-real-time) scheduling –Real-time scheduling.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Static Process Scheduling
The New Policy for Enterprise Networking Robert Bays Chief Scientist June 2002.
Dynamic Placement of Virtual Machines for Managing SLA Violations NORMAN BOBROFF, ANDRZEJ KOCHUT, KIRK BEATY SOME SLIDE CONTENT ADAPTED FROM ALEXANDER.
Network Weather Service. Introduction “NWS provides accurate forecasts of dynamically changing performance characteristics from a distributed set of metacomputing.
Performance Computer Organization II 1 Computer Science Dept Va Tech January 2009 © McQuain & Ribbens Defining Performance Which airplane has.
LACSI 2002, slide 1 Performance Prediction for Simple CPU and Network Sharing Shreenivasa Venkataramaiah Jaspal Subhlok University of Houston LACSI Symposium.
Holding slide prior to starting show. Scheduling Parametric Jobs on the Grid Jonathan Giddy
Resource Characterization Rich Wolski, Dan Nurmi, and John Brevik Computer Science Department University of California, Santa Barbara VGrADS Site Visit.
Chapter 7 Packet-Switching Networks Shortest Path Routing.
OPERATING SYSTEMS CS 3502 Fall 2017
Introduction to Load Balancing:
Defining Performance Which airplane has the best performance?
Mean Value Analysis of a Database Grid Application
Morgan Kaufmann Publishers
Chapter 6: CPU Scheduling
Lottery Scheduling Ish Baid.
CSCI1600: Embedded and Real Time Software
湖南大学-信息科学与工程学院-计算机与科学系
CprE 458/558: Real-Time Systems
Chapter 6: CPU Scheduling
Module 5: CPU Scheduling
3: CPU Scheduling Basic Concepts Scheduling Criteria
Chapter5: CPU Scheduling
Chapter 6: CPU Scheduling
Chapter 6: CPU Scheduling
Module 5: CPU Scheduling
Parallel Programming in C with MPI and OpenMP
Chapter 6: CPU Scheduling
CSCI1600: Embedded and Real Time Software
Requirements of Computing in Network
Module 5: CPU Scheduling
Approximate Mean Value Analysis of a Database Grid Application
Presentation transcript:

Information and Scheduling: What's available and how does it change Jennifer M. Schopf Argonne National Lab

Oct 20, Information and Scheduling l How a scheduler work is closely tied to the information available l Choice of algorithm dependent on accessible data

Oct 20, This Talk l What approaches expect form information l What data is actually available, and some open questions l How data changes l What to do about changing data

Oct 20, NB l I’m speaking (pessimistically) from my own background l We’ve heard some talks earlier today (for example PACE) which address some of these problems l I still think these are interesting open issues to think about

Oct 20, Information systems (NOTE: taken from my standard MDS2 talk) l Information is always old –Time of flight, changing system state –Need to provide quality metrics l Distributed system state is hard to obtain –Information is not contemporaneous (thanks j.g.) –Complexity of global snapshot l Components will fail l Scalability and overhead –Approaches are changed for scalability, this will affect the information available

Oct 20, Scheduling approaches assume l A lot of data is available l All information is accurate l Values don’t change

Oct 20, Example: System data 1. The bandwidth bij : the maximum data rate in bits per second. 2. The flow fij : the effective data rate in bits per second on the link. 3. The utilization uij : the utilization is represented as the ratio of the effective flow to bandwidth, uij = fij / bij 4. The length lij : the Euclidean distance between its end peers. 5. The cost C ij : the cost can be defined as a function of the link length and its bandwidth, C ij = S*(lij/bij), where S is a constant value. 6. T i the processor speed of the peer, which is the number of work units that the peer can execute per unit of time. etc.

Oct 20, Example: Application information 1. B i is the number of work units (in terms of computations) in the task. So, the number of time units that the task ti needs in order to be executed on peer vk are B i/Tk 2. u i, is the number of packets required to transfer the task. Thus, the task ti needs u iw/bij work units to be transferred from peer vi to the peer vj, assuming that these two peers are direct neighbors and the condition of the network is ideal. l 3. Implicit: exact mapping of tasks and data in a DAG l etc…

Oct 20, What some people expect l Perfect bandwidth info l Number of operations in an application l Scalar value of computer “power” l Mapping of “power” to applications l Perfect load information

Oct 20, Bandwidth data l Network Weather Service (Wolski, UCSB) –64k probe BW data –Latency data –Predictions l Pinger (Les Cotrell, SLAC) –Create long term baselines for expectations on means/medians and variability for response time, throughput, packet loss l Predicting TCP performance –Allen Downey – l But what do Grid applications need?

Oct 20, Perfect Bandwidth Data 64 k probes don’t look like large file transfers LBL-ANL GridFTP (approximately 400 transfers at irregular intervals) end-to-end bandwidth and NWS (approximately 1,500 probes every five minutes) probe bandwidth for the two-week August’01 dataset.

Oct 20, Predicting Large File Transfers l Vazhkudai and Schopf: use GridFTP logs and some background data - NWS, ioStat (HPDC 2002) –Error rate of ~15% l M. Faerman A. Su, R. Wolski, and F. Berman (HPDC 99) –Similar results for SARA data l Hu and Schopf: use an AI learning technique on GridFTP log files only (not published yet) –Picks best place to get a file from 60-80% of time, using averages only gives you ~50% “best chosen” l This topic needs much more study!

Oct 20, Data Generally Available From an Application l What some scheduling approaches want: –Number of ops in an application –Exact execution time on a platform –Perfect models of applications

Oct 20, Application Data Currently Available l Bad models of applications l No models of applications –Some work (Propehsy, Taylor at Texas A&M) does logging to create models l Many interesting applications have non- deterministic run times l User estimates of application run time (historically) off by 20%+ l We need to be able to figure out ways to do predictions of application run times WITHOUT models

Oct 20, Scalar value of computer “power” l MDS2 gives me: –CPU vendor, model and version –CPU speed –OS name, release and version –RAM size –Node count –CPU count l Where is “compute power” in this data?

Oct 20, What is compute “power” l I could get benchmark data, but what’s the right benchmark(s) to use? l Computer “power” simply isn’t scalar, especially in a Grid environment l Goal is really to understand how an application will run on a machine Given three different benchmarks, 3 different platforms will perform very differently – one best on BM1, another best on BM2

Oct 20, Mapping “power” to applications l Many scheduling approaches assume “power” is a scalar – just multiply it by the set application time and we’re set l Only problem: –Power isn’t a scalar –No one knows absolute application run times –Mapping will NOT be straight forward l We need a way to estimate application time on a contended system

Oct 20, Perfect Load Information l MDS2 gives me: –Basic queue data –Host load 5/10/15 min avg –Last value only

Oct 20, Load Predictions l Network weather service –12+ prediction techniques –Work on any time series –Expect regularly arriving data l Only a prediction of the next value –*I* want to know what load is going to be like in 20 mins –Or the AVERAGE over the next 20 mins?

Oct 20, Information and Scheduling l What approaches expect us to have l What we actually have access to l How it changes l What to do about changing data

Oct 20, Dedicated SOR Experiments l Platform- 2 Sparc 2’s. 1 Sparc 5, 1 Sparc 10 l 10 mbit ethernet connection l Quiescent machines and network l Prediction within 3% before memory spill

Oct 20, Non-dedicated SOR results l Available CPU on workstations varied from.43 to.53

Oct 20, SOR with Higher Variance in CPU Availability

Oct 20, Improving predictions l Available CPU has range of / l Prediction should also have a range

Oct 20, Scheduling needs to consider variance Conservative Scheduling: Using Predicted Variance to Improve Scheduling Decisions in Dynamic Environments –Lingyun Yang, Jennifer M. Schopf, Ian Foster –To appear at SC'03, November 15-21, 2003, Phoenix, Arizona, USA – scheduling.pdfwww.mcs.anl.gov/~jms/Pubs/lingyun-SC- scheduling.pdf

Oct 20, Scheduling with Variance l Summary: Scheduling with variance can give better mean performance and less variance in overall execution time

Oct 20, Lessons: l We need work predicting large file transfers – NOT bandwidth l We need to be able to figure out ways to do predictions of application run times WITHOUT models l We need predictions over time periods – not just a next value l We need a way to represent “power” of a machine, that takes variance into account l We need a way to map power to application behavior l We need better scheduling approaches that take variance into account

Oct 20, Contact Information l Jennifer M. Schopf l l –Links to some of the publications mentioned –Links to the co-edited book “Grid resource Management: State of the Art and Future Trends”