1 Reading Report 8 Yin Chen 24 Mar 2004 Reference: A Network Performance Tool for Grid Environments, Craig A. Lee, Rich Wolski, Carl Kesselman, Ian Foster,

Slides:



Advertisements
Similar presentations
Agenda of Week V. Forecasting
Advertisements

Part II – TIME SERIES ANALYSIS C3 Exponential Smoothing Methods © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Congestion Control Created by M Bateman, A Ruddle & C Allison As part of the TCP View project.
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
4-1 Operations Management Forecasting Chapter 4 - Part 2.
Fast Pattern-Based Throughput Prediction for TCP Bulk Transfers Tsung-i (Mark) Huang Jaspal Subhlok University of Houston GAN ’ 05 / May 10, 2005.
Measurements of Congestion Responsiveness of Windows Streaming Media (WSM) Presented By:- Ashish Gupta.
Time Series Analysis Autocorrelation Naive & Simple Averaging
Receiver-driven Layered Multicast S. McCanne, V. Jacobsen and M. Vetterli SIGCOMM 1996.
T T18-03 Exponential Smoothing Forecast Purpose Allows the analyst to create and analyze the "Exponential Smoothing Average" forecast. The MAD.
The Network Weather Service A Distributed Resource Performance Forecasting Service for Metacomputing Rich Wolski, Neil T. Spring and Jim Hayes Presented.
Chapter 12 - Forecasting Forecasting is important in the business decision-making process in which a current choice or decision has future implications:
On the Constancy of Internet Path Properties Yin Zhang, Nick Duffield AT&T Labs Vern Paxson, Scott Shenker ACIRI Internet Measurement Workshop 2001 Presented.
MOVING AVERAGES AND EXPONENTIAL SMOOTHING
CHAPTER 3 Forecasting.
Lecture 3 Forecasting CT – Chapter 3.
Chapter 3 Forecasting McGraw-Hill/Irwin
Chapter 13 Forecasting.
1 Emulating AQM from End Hosts Presenters: Syed Zaidi Ivor Rodrigues.
Part II – TIME SERIES ANALYSIS C2 Simple Time Series Methods & Moving Averages © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Forecasting McGraw-Hill/Irwin Copyright © 2012 by The McGraw-Hill Companies, Inc. All rights reserved.
Statistical Forecasting Models
Ns Simulation Final presentation Stella Pantofel Igor Berman Michael Halperin
Jennifer Rexford Princeton University MW 11:00am-12:20pm Wide-Area Traffic Management COS 597E: Software Defined Networking.
Applied Business Forecasting and Planning
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. 3 Forecasting.
Slides 13b: Time-Series Models; Measuring Forecast Error
Fall, 2012 EMBA 512 Demand Forecasting Boise State University 1 Demand Forecasting.
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Demand Management and Forecasting
1 Reading Report 9 Yin Chen 29 Mar 2004 Reference: Multivariate Resource Performance Forecasting in the Network Weather Service, Martin Swany and Rich.
TIME SERIES by H.V.S. DE SILVA DEPARTMENT OF MATHEMATICS
3-1 McGraw-Hill/Irwin Operations Management, Seventh Edition, by William J. Stevenson Copyright © 2002 by The McGraw-Hill Companies, Inc. All rights reserved.
Operations Management
3-1Forecasting William J. Stevenson Operations Management 8 th edition.
3-1Forecasting. 3-2Forecasting FORECAST:  A statement about the future value of a variable of interest such as demand.  Forecasts affect decisions and.
DSc 3120 Generalized Modeling Techniques with Applications Part II. Forecasting.
1 DSCI 3023 Forecasting Plays an important role in many industries –marketing –financial planning –production control Forecasts are not to be thought of.
Time series Model assessment. Tourist arrivals to NZ Period is quarterly.
1 1 Slide Forecasting Professor Ahmadi. 2 2 Slide Learning Objectives n Understand when to use various types of forecasting models and the time horizon.
Lesson 5.1 Forecasting Moving Averages Exponential Smoothing.
Forecasting. 預測 (Forecasting) A Basis of Forecasting In business, forecasts are the basis for budgeting and planning for capacity, sales, production and.
Forecasting Chapter 9. Copyright © 2013 Pearson Education, Inc. publishing as Prentice Hall Define Forecast.
Reid & Sanders, Operations Management © Wiley 2002 Forecasting 8 C H A P T E R.
4-1 Operations Management Forecasting Chapter 4 - Part 2.
1 1 Chapter 6 Forecasting n Quantitative Approaches to Forecasting n The Components of a Time Series n Measures of Forecast Accuracy n Using Smoothing.
Forecasting is the art and science of predicting future events.
CHAPTER 12 FORECASTING. THE CONCEPTS A prediction of future events used for planning purpose Supply chain success, resources planning, scheduling, capacity.
Network Weather Service. Introduction “NWS provides accurate forecasts of dynamically changing performance characteristics from a distributed set of metacomputing.
3-1Forecasting CHAPTER 3 Forecasting McGraw-Hill/Irwin Operations Management, Eighth Edition, by William J. Stevenson Copyright © 2005 by The McGraw-Hill.
Distance Vector Routing
3-1Forecasting William J. Stevenson Operations Management 8 th edition.
MBF1413 | Quantitative Methods Prepared by Dr Khairul Anuar 8: Time Series Analysis & Forecasting – Part 1
Performance Comparison of Ad Hoc Network Routing Protocols Presented by Venkata Suresh Tamminiedi Computer Science Department Georgia State University.
McGraw-Hill/Irwin Copyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 3 Forecasting.
Demand Management and Forecasting Chapter 11 Portions Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Forecast 2 Linear trend Forecast error Seasonal demand.
Chapter 15 Forecasting. Forecasting Methods n Forecasting methods can be classified as qualitative or quantitative. n Such methods are appropriate when.
Forecasting. Model with indicator variables The choice of a forecasting technique depends on the components identified in the time series. The techniques.
Chapter 11 – With Woodruff Modications Demand Management and Forecasting Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 3 Lecture 4 Forecasting. Time Series is a sequence of measurements over time, usually obtained at equally spaced intervals – Daily – Monthly –
T T18-02 Weighted Moving Average Forecast Purpose Allows the analyst to create and analyze the "Weighted Moving Average" forecast for up to 5.
Lecture 9 Forecasting. Introduction to Forecasting * * * * * * * * o o o o o o o o Model 1Model 2 Which model performs better? There are many forecasting.
Confidence Intervals Cont.
Forecasting Methods Dr. T. T. Kachwala.
Fast Pattern-Based Throughput Prediction for TCP Bulk Transfers
FORCASTING AND DEMAND PLANNING
MBF1413 | Quantitative Methods Prepared by Dr Khairul Anuar
Presentation transcript:

1 Reading Report 8 Yin Chen 24 Mar 2004 Reference: A Network Performance Tool for Grid Environments, Craig A. Lee, Rich Wolski, Carl Kesselman, Ian Foster, pdf.pdf

2 Overview Network bandwidth discovery and allocation is a serious issue in Grid.  How to know which hosts have suitable bandwidth?  How should application initially configure and adapt to network conditions?  Difficulties -- Running application can monitor their own performance but cannot measure other parts of the environment.  Need a tool for performance monitoring and prediction. Gloperf  Makes sample, end-to-end TCP measurements  No special host permissions is required, and independent on link topology  Addressed problems of Accuracy vs. Instrusiveness Portability Scalability Fault tolerance Security Measurement Policy Data Discovery, Access and Usebility

3 The Gloperf Design Four Function Areas  Sensors : a set of gloperfd daemons  Collation of data : data is collected into Globus Metacomputing Directory Service(MDS)  Production of derived data  Access and use of data Working Mechanism  When Globus is started on a host, one gloperfd will register to MDS  Then, goes into main control loop while ( no termination signal ) { query MDS for all other gloperfds to do peer and group discovery; filter list of gloperfds on version number and groups; build a randomized list of tests that does a bandwidth test and a latency test to each filtered peer; for each element of the list { perform the test; write the results in the MDS; wait n minutes; }  Measurement based on netperf Uses a netperf “TCP_STREAM” test TCP/IP overhead Uses a netperf “TCP_RR” measure latency  One bandwidth or latency test in the randomized list is done every n min. (n = 5)

4 Performance Results Highly accurate means minimally intrusive. Frequent measurements can increase accurate, but also increase the intrusiveness. Can be improved by using shorter tests. The experimental version of Gloperf makes one test per min between a pair of hosts  A cycle of 5 test times are used : 1,2,5,7,10 sec.  Collect 5, interleaved time series of measurements, with 5 min between tests.  To evaluate the effect of test period, compare the error in this 5-min data with the error produced when the same data is subsampled to produce a longer test period.  i.e., subsample every 12 th point to effectively produce a one-hour test period.

5 Performance Results Figure1 shows the bandwidth measurements made between sleipnir.aero.org (near Los Angeles) and yukon.mcs.anl.gov (near Chicago) for approximately a 48 hour period.  The 1-sec. tests are the lowest on average and longer test times produce a higher measured bandwidth. Figure 2 is the histograms of sleipnir- yukon data.  TCP coming out of slow start.  5,7, and 10-sec. tests all share the same mode.  This indicates that tests longer than 5- sec. are NOT necessary on this link in accordance with the bandwidth-delay product governing TCP slow start.

6 Performance Results NOT all links are well-behaved. Figure 5 shows a similar experiment between jupiter.isi.edu (in Los Angeles) and yukon.mcs.anl.gov (near Chicago) that exhibits a volatile mean. Only the 1 and 10-sec. tests are shown. Figure 6 is the histograms of jupiter-yukon.

7 Performance Results To assess the “error”, require to compare with “true” values. Need prediction model based on Gloperf data. The simplest prediction model assumes that the last measurement predicts the next and the error is the difference between them. By using the last measurement predictor, can assess the error by computing the Mean Absolute Error (MAE) and assess the variability by comparing that with the Mean Squared Error (MSE). Can also compute the MAE and MSE for subsampled data. Since when subsampling every n-th point, there are n different possible subsampled time series, thus, for a given subsampling period, compute the MAE/MSE for every possible series.

8 Performance Results Figure 3 shows sleipnir-yukon MAE/MSE as a function of test time.  The longer test times “average-out” some of the variability seen in the shorter test times resulting in a significantly lower MAE and MSE.  An increase in the MAE for the 2-sec. tests. May caused by packet loss At two seconds, TCP is in the process of leaving slow start which causes the strata to be spread out wider, producing a higher MAE. Once out of slow start with a longer test time, the strata are more narrowly constrained and an individual packet drop has less effect. Figure 4 shows the same data as a function of test period.  The test period has NO effect on the MAE. Testing once an hour is just as accurate as testing once every five-min, since the original data have relatively stationary means.  In this situation, infrequent measurement is adequate and preferable.

9 Performance Results Figure 7 shows jupiter-yukon MAE/ MSE by test time using the last measurement predictor and a random predictor. Even in a “noisy” environment, infrequent measurements still produce a BETTER bandwidth estimate than random guessing.

10 Performance Results One drawback of the last measurement predictor is that an individual measurement may not be representative of the subsequent “true” bandwidth. i.e., the measurement happened to hit a transient peak or trough in the trace. In a long-lived operational environment, an extremely large number of samples would be collected and a new measurement would NOT perceptibly change the mean. A running exponential average could be used to track the current “load” based on a weighted sum of the current measurement and the previous “load”. This allows current peaks and troughs in the trace to be weighted against recent past measurement with an exponential decay. Figure 8 shows jupiter-yukon MAE/ MSE by test time for both the last measurement predictor and an exponential average predictor with a weight of 0.35 on the most recent measurement.

11 Performance Results Work with the Network Weather Service showed that simple forecasting techniques such as exponential smoothing can improve prediction accuracy dramatically over simple use of monitor data alone. In this case, exponential average data shows a clear improvement in the prediction.  Figure 9 shows exponential average data plotted as a function of test period.  Once out of slow start, the MAE only rises approximately 10% from testing every 5-min to testing once an hour. (Compare with Figure 4)

12 Performance Results Rise the test period to once every 24 hours. Used a trace generated by the NWS that contained high-frequency tests over a 28.8 day period. Subsampeld this trace to produce a trace with 7163 tests at a test period of 300 sec. over the same 28.8 days. Figure 10 shows the MAE and MSE for test periods of 5,10,30,45,60 min and 2,4,8,12,24 hours.  Beyond one hour, the MAE dose start to rise more quickly.  But, the average worst error measuring once a day only amounts to less than20% of the actual bandwidth observed.  Note, interleave time series do not produce the same MAE and MSE error readings. Two series shifted only by one measurement can generate different overall forecasting errors. A frequency analysis does NOT expose such periodicities. Figure 11 is the histogram of exponential predictor MAEs from subsampled NWS trace.  Despite the variability, the worst-case sequence with a one-day measurement period produced an error of only slightly worse than 25% of the actual bandwidth observed.  Also note that the 24 hour period exhibits the widest degree of variability leading is to conjecture that cyclic frequency components are the cause of the variability.

13 Performance Discussion Longer test period do not erode accuracy. Longer test times tend to average-out variation, and also increase the intrusiveness. The overhead of Gloperf in total number of bytes transferred per hour is linearly related to the test period.  Going from 5 min, to 60 min, represents a 12x reduction in testing data volume  Going from 1 hour to 24 hours represents a 24x redoction.  On fast, under-utilized links, Gloperf can push a larger data volume than is nucessary. What is the best metric for “intrusiveness”  The number of bytes transferred  The time spent sending additional traffic  The bandwidth actually denied to other flows  The congestion precipitated at routers along the path But not all of these metrics are easy to capture. End …