1 Reading Report 7 Yin Chen 22 Mar 2004 Reference: Customer View of Internet Service Performance: Measurement Methodology and Metrics, Cross-Industry Working.

Slides:



Advertisements
Similar presentations
Data and Computer Communications
Advertisements

1 MM3 - Reliability and Fault tolerance in Networks Service Level Agreements Jens Myrup Pedersen.
G. Alonso, D. Kossmann Systems Group
1 Routing Techniques in Wireless Sensor networks: A Survey.
A Flexible Model for Resource Management in Virtual Private Networks Presenter: Huang, Rigao Kang, Yuefang.
11 TROUBLESHOOTING Chapter 12. Chapter 12: TROUBLESHOOTING2 OVERVIEW  Determine whether a network communications problem is related to TCP/IP.  Understand.
1 © Fluke networks 2004 Everett WAMonday, May 18, 2015 Application Performance & Network Analysis Improving the end user experience.
Multiple constraints QoS Routing Given: - a (real time) connection request with specified QoS requirements (e.g., Bdw, Delay, Jitter, packet loss, path.
Evaluating Hypotheses
Lecture 9 Output Analysis for a Single Model. 2  Output analysis is the examination of data generated by a simulation.  Its purpose is to predict the.
1 Drafting Behind Akamai (Travelocity-Based Detouring) AoJan Su, David R. Choffnes, Aleksandar Kuzmanovic, and Fabian E. Bustamante Department of Electrical.
User-level Internet Path Diagnosis R. Mahajan, N. Spring, D. Wetherall and T. Anderson.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
CHAPTER 6 Statistical Analysis of Experimental Data
BCOR 1020 Business Statistics Lecture 21 – April 8, 2008.
Department of Electronic Engineering City University of Hong Kong EE3900 Computer Networks Transport Protocols Slide 1 Transport Protocols.
8-1 Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall Chapter 8 Confidence Interval Estimation Statistics for Managers using Microsoft.
Copyright ©2011 Pearson Education 8-1 Chapter 8 Confidence Interval Estimation Statistics for Managers using Microsoft Excel 6 th Global Edition.
Total Quality Management BUS 3 – 142 Statistics for Variables Week of Mar 14, 2011.
Chapter 8 Introduction to Hypothesis Testing
Network Measurement Bandwidth Analysis. Why measure bandwidth? Network congestion has increased tremendously. Network congestion has increased tremendously.
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
EHealth Network Monitoring Network Tool Presentation J. Gaston Senior Network Design Seminar Professor Morteza Anvari 10 December 2004.
X-bar and R Control Charts
Monté Carlo Simulation MGS 3100 – Chapter 9. Simulation Defined A computer-based model used to run experiments on a real system.  Typically done on a.
Chapter 9 Numerical Integration Numerical Integration Application: Normal Distributions Copyright © The McGraw-Hill Companies, Inc. Permission required.
Maputo, Mozambique, April 2014 QoS Framework for Broadband Internet Service Country Case: Mauritius Roubee GADEVADOO, Engineer/Licensing Officer,
Draft-constantine-ippm-tcp-throughput-tm-00.txt 1 TCP Throughput Testing Methodology IETF 76 Hiroshima Barry Constantine
A User Experience-based Cloud Service Redeployment Mechanism KANG Yu.
Reading Report 14 Yin Chen 14 Apr 2004 Reference: Internet Service Performance: Data Analysis and Visualization, Cross-Industry Working Team, July, 2000.
Confidence Intervals and Hypothesis Testing - II
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
Fundamentals of Hypothesis Testing: One-Sample Tests
23-Support Protocols and Technologies Dr. John P. Abraham Professor UTPA.
Chapter 4. After completion of this chapter, you should be able to: Explain “what is the Internet? And how we connect to the Internet using an ISP. Explain.
Error Analysis Accuracy Closeness to the true value Measurement Accuracy – determines the closeness of the measured value to the true value Instrument.
Application-Layer Anycasting By Samarat Bhattacharjee et al. Presented by Matt Miller September 30, 2002.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Confidence Interval Estimation Basic Business Statistics 11 th Edition.
Confidence Interval Estimation
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
POSTECH DP&NM Lab. Internet Traffic Monitoring and Analysis: Methods and Applications (1) 2. Network Monitoring Metrics.
1 Network Monitoring Mi-Jung Choi Dept. of Computer Science KNU
© 2010 AT&T Intellectual Property. All rights reserved. AT&T, the AT&T logo and all other AT&T marks contained herein are trademarks of AT&T Intellectual.
Computer Networks with Internet Technology William Stallings
ECO-DNS: Expected Consistency Optimization for DNS Chen Stephanos Matsumoto Adrian Perrig © 2013 Stephanos Matsumoto1.
1 Chapters 8 Overview of Queuing Analysis. Chapter 8 Overview of Queuing Analysis 2 Projected vs. Actual Response Time.
ERCOT Planning WMS 10/20/2010 Target Reserve Margin and Effective Load Carrying Capability of Installed Wind Capacity for the ERCOT System – Methodology.
Wide-scale Botnet Detection and Characterization Anestis Karasaridis, Brian Rexroad, David Hoeflin In First Workshop on Hot Topics in Understanding Botnets,
Chapter 7 Sampling Distributions Statistics for Business (Env) 1.
CEN st Lecture CEN 4021 Software Engineering II Instructor: Masoud Sadjadi Monitoring (POMA)
Lecture 9 Chap 9-1 Chapter 2b Fundamentals of Hypothesis Testing: One-Sample Tests.
TELE202 Lecture 6 Routing in WAN 1 Lecturer Dr Z. Huang Overview ¥Last Lecture »Packet switching in Wide Area Networks »Source: chapter 10 ¥This Lecture.
Service Level Monitoring. Measuring Network Delay, Jitter, and Packet-loss  Multi-media applications are sensitive to transmission characteristics of.
Chap 8-1 Chapter 8 Confidence Interval Estimation Statistics for Managers Using Microsoft Excel 7 th Edition, Global Edition Copyright ©2014 Pearson Education.
High-Speed Policy-Based Packet Forwarding Using Efficient Multi-dimensional Range Matching Lakshman and Stiliadis ACM SIGCOMM 98.
Teknik Routing Pertemuan 10 Matakuliah: H0524/Jaringan Komputer Tahun: 2009.
C HAPTER 4  Hypothesis Testing -Test for one and two means -Test for one and two proportions.
Development of a QoE Model Himadeepa Karlapudi 03/07/03.
1 Transport Layer: Basics Outline Intro to transport UDP Congestion control basics.
Hypothesis Testing. Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean μ = 120 and variance σ.
Network Layer Lecture Network Layer Design Issues.
TCP/IP1 Address Resolution Protocol Internet uses IP address to recognize a computer. But IP address needs to be translated to physical address (NIC).
Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 8 th Edition Chapter 9 Hypothesis Testing: Single.
Prof. Alfred J Bird, Ph.D., NBCT Office – Science 3rd floor – S Office Hours – Monday and Thursday.
Sampling and Sampling Distributions. Sampling Distribution Basics Sample statistics (the mean and standard deviation are examples) vary from sample to.
Manajemen Jaringan, Sukiswo ST, MT 1 Network Monitoring Sukiswo
Prof. Alfred J Bird, Ph.D., NBCT Office – McCormick 3rd floor 607 Office Hours – Monday 3:00 to 4:00 and.
Parameter Estimation.
Authors – Johannes Krupp, Michael Backes, and Christian Rossow(2016)
Presentation transcript:

1 Reading Report 7 Yin Chen 22 Mar 2004 Reference: Customer View of Internet Service Performance: Measurement Methodology and Metrics, Cross-Industry Working Team, Oct,

2 Problems Overview The expectations of reliability are driving customers to negotiate with their Internet Service providers (ISPs) for guarantees that will meet customer requirements for specific Qos levels. Difficulties  Remote networks that extend beyond the responsibility of the customer’s ISP can dictate application-level service quality.  Reaching agreement can be complex and time-consuming.  NO agreed-upon methodologies for measuring and monitoring The work intend to provide a common set of metrics and a common measurement methodology that can be used to assess, monitor, negotiate, and test compliance of service quality. The hope is that the application of the metrics and methodology will lead to improved Internet performance and foster greater cooperation between customers and service providers.

3 Related Work Internet Engineering Task Force (IETF) Automotive Industry Action Group (AIAG) T1A1 ITU-T Study Group 13 National Internet Measurement Infrastructure (NIMI) Surveyor

4 Measurement Methodology Requirement Isolate sources of problems Provide meaningful results Not required new infrastructure Avoid unnecessary or duplicate measurements / traffic Be auditable Be robust

5 Measurement Methodology Measurement Architecture Test point : collects performance data or responds to measurement queries. Measurement agent :communicates with test points to conduct the measurements or collect data. Dissemination agent : provides the results of the measurements.

6 Metric Definitions Performance Metrics  Packet Loss  Round Trip Delay Reliability Metrics  Reachability  Network Service Availability  Duration of Outage  Time Between Outages Ancillary Metrics  Network Resource Utilization  DNS Performance DNS query loss DNS response time Aggregation of Measurements  Measurement values V m  Measurement interval I m  Baseline value B m  Baseline spread S m  Aggregation interval I a  Aggregate value F

7 Metric Definitions Performance Metrics Packet Loss  Defined as the fraction of packets sent from a measurement agent to a test point for which the measurement agent does not receive an acknowledgment from the test point.  It includes packets that are not received by the test point AND acknowledgments that are lost before returning to the measurement agent.  Acknowledgments that do not arrive within a predefined round trip delay at the measurement agent are also considered lost. Round Trip Delay  Defined as the interval between the time a measurement agent application sends a packet to a test point and the time it receives acknowledgment that the packet was received by the test point.  It includes any queuing delays at the end-points or the intermediate hosts.  But it does NOT include any DNS lookup times by the measurement application.

8 Metric Definitions Reliability Metrics Difficult to specify service levels based directly on reliability. Instead use terms such as “availability”. NO common definitions and computational methodology exist, difficult to negotiate reliability guarantees—and harder to compare quality of service based on quoted values. From a customer perspective, there are three components to service reliability:  Can the customer reach the service?  If so, is the service available?  If not, how frequently and for how long do the outages last?

9 Metric Definitions Reliability Metrics (Cont.) Reachability  A test point is considered reachable from a measurement agent if the agent can send packets to the test point and, within a short predefined time interval, receive acknowledgment from the test point that the packet was received.  In most instances, can consider the PING test as a sufficient metric of reachability.  If each measurement sample consists of multiple PINGs, the test point is considered reachable if the measurement agent receives at least one acknowledgment from it. Network Service Availability  The network between a measurement agent and a test point is considered available at a given time t if, during a specified time interval D around t, the measured packet loss rate and the round trip delays are both BELOW predefined thresholds.  Network service availability -- defined as the fraction of time the network is available from a specified group (one or more) of measurement agents to a specified group of test points. It depends on network topology.

10 Metric Definitions Reliability Metrics (Cont.) Duration of Outage  Defined as the difference between the time a service becomes unavailable and the time it is restored.  Because of the statistical nature of Internet traffic, the duration over which service is measured to be unavailable should exceed some minimum threshold before it is declared an outage.  Similarly, when service is restored after an outage, it should stay available for some minimum duration before the outage is declared over. Time Between Outages  Defined as the difference between the start times of two consecutive outages.

11 Metric Definitions Ancillary Metrics Ancillary metrics -- often needed to interpret the results obtained from direct measurement of performance or availability.  Network Resource Utilization The percentage of a particular part of the network infrastructure used during a given time interval. It is a dimensionless number calculated by dividing the amount of the particular resource used during a given time interval by the total theoretically available amount of that particular resource during that same interval. Measuring resource utilization is especially important for KEY resources that include links and routers. BOTH utilization peaks and percentiles must be monitored.  DNS Performance DNS has become an increasingly important part of the Internet, almost all applications now use it to resolve host names to IP addresses. As a result, application-level response times can be SLOW if DNS performance is bad. Define DNS performance measures using two metrics  DNS query loss -- defined as the fraction of DNS queries made by a measurement agent for which the measurement agent does not receive a response from the DNS server within a predetermined time. (Analogous to the packet loss metric).  DNS response time -- defined as the interval between the time a measurement agent application sends a DNS query to a DNS server and the time it receives a response from the server providing the result of the query. (Analogous to the round trip delay metric.

12 Metric Definitions Tradeoffs Between Metrics The metrics are related, but provide different types of information. Good results with one metric may be a poor showing in another. Actions taken to improve one metric may have a negative effect on others. Most metrics depend on the utilization of resources in the network.

13 Metric Definitions Aggregation of Measurements Although individual measurements can be used to detect operational problems in near real time, metrics of interest usually need to be aggregated over time to obtain valid estimates of performance. Due to system complexity, it is difficult to predict the performance that can be achieved a priori; it is thus necessary to compute baselines that can be used for setting quality-of-service targets and for comparing performance. Statistical aggregates such as means or standard deviations are not appropriate to quantify performance of data networks because the underlying primary metrics have “heavy-tailed” distributions that are not represented well by those aggregates. These metrics can be more appropriately represented by percentiles and order statistics such as the median.

14 Metric Definitions Aggregation of Measurements (Cont.) Measurement values V m  Each measurement sample M results in a value V m.  In most cases, V m will itself be computed from a set of measurements.  i.e., V m could be the fraction of responses received when a given host is pinged 10 times at one-second intervals, or it could be the median round trip delay computed from the returned responses.  i.e., V m could represent the median packet loss measured between a group of sites in North America and a group in Europe. Measurement interval I m  I m is the interval between measurement samples.  If measurements occur at random times, I m is the expected value of the interval associated with the measurement.  i.e., a measurement may be taken every five minutes (periodic) or at intervals that are Poisson distributed with expected time of arrival equal to five minutes.  Note that I m defines the temporal resolution of the measurements—i.e., events that are shorter than I m in duration are likely to be missed by the measurements.

15 Metric Definitions Aggregation of Measurements (Cont.) Baseline value B m  A baseline value represents the expected value of M.  Baselines may be static (i.e., time invariant) or dynamic (time dependent).  Baselines may also be dependent on service load and other system parameters.  It is expected that under normal circumstances, baselines will be computed from historical records of measurement samples.  As noted above, the sample mean is a poor baseline value, and the median is probably a better baseline. Baseline spread S m  The baseline spread is a measure of the normal variability associated with M.  The baseline spread may be static or dynamic.  The spread should be computed using quartiles or percentiles rather than the standard deviation.  A measurement is considered to be within baseline if |V m - B m | / S m ≤ T m (T m is a threshold )  If the underlying measurement distributions are significantly asymmetric, the baseline spread may be specified in terms of an upper specification limit U m and a lower specification limit L m.

16 Metric Definitions Aggregation of Measurements (Cont.) Aggregation interval I a  I a is the time over which multiple measurement samples are aggregated to create the metric A which is representative of system behavior over that time.  Typically, aggregation may be provided over hourly, daily, weekly, monthly, quarterly or annual intervals.  Note that aggregation intervals may be disjoint, i.e., aggregation may occur only at peak times or during business hours. Aggregate value F  The aggregate value F a is the fraction of measurements that are within baseline over the aggregation interval.  If N measurements are taken over the aggregation interval, and N b are within baseline, the aggregate value is F a = N b / N  Bounds may be placed on F a to specify “acceptable” service behavior. Measurements that return illegal or unknown values (e.g., if all packets are lost in a round trip delay measurement) should normally be considered out of baseline for computing the aggregate value.  Note that while aggregation intervals used to compute F a are likely to be large for monitoring and planning purposes, alarms may be generated by as few as one or two sequential measurements if they are sufficiently out of baseline.  The work argued that an aggregate value is NOT calculated by averaging the measurement values, because of the long time intervals involved in aggregation, such values do NOT provide meaningful conclusions about service quality.  The baseline value(s) can be used for historical comparisons if they are dynamic.

17 Example of Computation of Metrics Scenario  The customer has some resources that it wishes to make available for users on the Internet. Methodology  The measurement agents will sample at some given interval, i.e.,5,10,15 or 30 min  A measurement agent can PING each test point and the use the results to measure round trip delay, packet loss, and reachability.  These measurements can then be used to determine availability during the sampling interval.  When a test point is not reachable, the measurement agent records the time that the outage occurred.  When the outage has been repaired, duration of the outage can be calculated.  When another outage occurs, time between outages can be recorded.  Baseline Measurement first needs to take place for a few weeks (a minimum of two) to establish baseline values and baseline spreads for all metrics. The objective is to establish norms of network behavior. Measurement that fall sufficiently far outside the baseline can be flagged as problems. Comparisons between the measurements made by agents on the customer’s ISP and agents on OTHER ISPs can be used to decide if the problem is due to the customer’s ISP or beyond it. Difficulties  Gathering data from agents on OTHER ISPs.  What to do when problems are found, since ISPs have little direct control to fix problems on OTHER ISPs.