Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-011 Draft-geib-ippm-metrictest-01. Incorporates RFC2330 philosophy, draft-bradner and draft-morton.

Slides:



Advertisements
Similar presentations
IPPM WG. Note Well Any submission to the IETF intended by the Contributor for publication as all or part of an IETF Internet-Draft or RFC and any statement.
Advertisements

ISO TC209 WG1, Nov. 11, 2005.Page 1 ISO TC209 WG1 N Statistical Issues - reissue.ppt ISO Statistical Issues Mark Varney, Statistician.
TSVWG #1 IETF-92 (Dallas) 24 th March 2015 Gorry Fairhurst David Black WG chairs.
Exercising these ideas  You have a description of each item in a small collection. (30 web sites)  Assume we are looking for information about boxers,
What’s the Problem Web Server 1 Web Server N Web system played an essential role in Proving and Retrieve information. Cause Overloaded Status and Longer.
R.G. Cole - AT&T Labs1rperfmon BOF Active Probes for Performance Monitoring (APPM) Draft: posted to DISMAN WG list Authors: Cole, R., Kalbfleisch, C. and.
1 Confidence Intervals for Means. 2 When the sample size n< 30 case1-1. the underlying distribution is normal with known variance case1-2. the underlying.
Design, goal of design, design process in SE context, Process of design – Quality guidelines and attributes Evolution of software design process – Procedural,
Nov. 2010Geib/Morton/Fardid/Steinmitz / draft-ietf-metrictest-011 Testing Standards Track Metrics Draft-ietf-ippm-metrictest-01 Geib, Morton, Fardid, Steinmitz.
Today Evaluation Measures Accuracy Significance Testing
IETF 68 – SIMPLE WG SIMPLE Problem Statement draft-ietf-simple-interdomain-scaling-analysis-00 Avshalom Houri – IBM Tim Rang - Microsoft Edwin Aoki – AOL.
CHAPTER 21: Comparing Two Proportions
Lecture 15. IGP and MPLS D. Moltchanov, TUT, Spring 2008 D. Moltchanov, TUT, Spring 2015.
WG RAQMON Internet-Drafts RMON MIB WG Meeting Washington, Nov. 11, 2004.
Draft-campbell-dime-load- considerations-01 IETF 92 DIME Working Group Meeting Dallas, Texas.
Welcome to the Unit 8 Seminar Dr. Ami Gates
Framework for Performance Metric Development draft-morton-perf-metrics-framework-01.txt Alan Clark IETF 70 PMOL WG.
POSTECH DP&NM Lab. Internet Traffic Monitoring and Analysis: Methods and Applications (1) 2. Network Monitoring Metrics.
(Long-Term) Reporting Metrics: Different Points of View Al Morton Gomathi Ramachandran Ganga Maguluri November 2010 draft-ietf-ippm-reporting-metrics-04.
1 Advancing Metrics on the Standards Track: RFC 2680 (1-way Loss) Test Plan and Results draft-ietf-ippm-testplan-rfc Len Ciavattone, Rüdiger Geib,
1 IPFIX Protocol Specifications IPFIX IETF-59 March 3, 2004 Benoit Claise Mark Fullmer Reinaldo Penno Paul Calato Stewart Bryant Ganesh Sadasivan.
Performance monitoring -- Nicolas Simar Performance monitoring TF-NGN meeting, Rome (Italy), Nicolas Simar, Network.
1 draft-ietf-ippm-loss-episode-metrics-00 Loss Episode Metrics for IPPM Nick Duffield, Al Morton, AT&T Joel Sommers, Colgate University IETF 79, Beijing,
1 Miscellaneous Capabilities for IP Network Infrastructure IETF 64 Vancouver, BC, Canada November 2005.
One-way ANOVA: - Comparing the means IPS chapter 12.2 © 2006 W.H. Freeman and Company.
CHAPTER 37 Presentation of Data 2. Time Series A TIME SERIES is a set of readings taken at TIME INTERVALS. A TIME SERIES is often used to monitor progress.
1 Monami6 Working Group IETF 66 July 2006 Montréal, Canada Thierry Ernst (INRIA) Nicolas Montavont (ENST Bretagne)
Pair Development Framework Monvorath (Molly) Phongpaibul.
TFRC for Voice: the VoIP Variant Sally Floyd, Eddie Kohler. March draft-ietf-dccp-tfrc-voip-01.txt
Delay Variation Applicability Statement draft-morton-ippm-delay-var-as-02 March 21, 2007 Al Morton Benoit Claise “
1 draft-duffield-ippm-burst-loss-metrics-01.txt Nick Duffield, Al Morton, AT&T Joel Sommers, Colgate University IETF 76, Hiroshima, Japan 11/10/2009.
Software testing techniques Software testing techniques Statistical Testing Presentation on the seminar Kaunas University of Technology.
IETF-90 (Toronto) DHC WG Meeting Wednesday, July 23, GMT IETF-90 DHC WG1 Last Updated: 07/21/ :10 EDT.
RBSP Radiation Belt Storm Probes RBSP Radiation Belt Storm Probes 12/25/20151 Flight Software Template for Instrument Critical Design Review Gary M. Heiligman.
Delay Variation Applicability Statement draft-morton-ippm-delay-var-as-03 July 24, 2007 Al Morton Benoit Claise.
How Good is a Model? How much information does AIC give us? –Model 1: 3124 –Model 2: 2932 –Model 3: 2968 –Model 4: 3204 –Model 5: 5436.
Chapter 11: The ANalysis Of Variance (ANOVA)
Privecsg Privacy Recommendation PAR Proposal Date: [ ] Authors: NameAffiliationPhone Juan Carlos ZúñigaInterDigital
EGEE is a project funded by the European Union under contract IST Study of Performance Standards, kick off (Task 1.1.1) Robert Stoy DFN EGEE.
IPFIX MIB Status Managed Object for IP Flow Export A Status Report Thomas Dietz Atsushi Kobayashi
July 2010Geib/Morton/Fardid/Steinmitz / draft-ietf-metrictest-001 Testing Standards Track Metrics Draft-ietf-ippm-metrictest-00 Geib, Morton, Fardid, Steinmitz.
TSVWG IETF-89 (London) 5 th & 7 th March 2014 Gorry Fairhurst David Black James Polk WG chairs 1.
IETF-89 (London) DHC WG Meeting Monday, March 3, GMT IETF-89 DHC WG1 Last Updated: 02/27/ EST.
1 Content-Aware Device Benchmarking Methodology (draft-hamilton-bmwg-ca-bench-meth-05) BMWG Meeting IETF-79 Beijing November 2010 Mike Hamilton
Issue EAPoL-Key message generation at WTP or AC Issue 199, summarized as:...the WTP maintains the KeyRSC while the AC requires this information to.
64th IETF Vancouver November 2005 ASON-Compatible Signaling.
1 Advancing Metrics on the Standards Track: RFC 2680 (Loss) Test Plan and Results draft-morton-ippm-testplan-rfc Len Ciavattone, Rüdiger Geib, Al.
Privecsg Privacy Recommendation PAR Proposal Date: [ ] Authors: NameAffiliationPhone Juan Carlos ZúñigaInterDigital
1 Life Cycle Assessment A product-oriented method for sustainability analysis UNEP LCA Training Kit Module k – Uncertainty in LCA.
draft-morton-ippm-testplan-rfc
IP Flow Information eXport (IPFIX)
Authors: Scott Poretsky, Quarry Technologies Brent Imhoff, LightCore
RFC6374 Synonymous Flow Labels draft-bryant-mpls-rfc6374-sfl-03
Target for Today Know what can go wrong with a survey and simulation
Nick Duffield, Al Morton, AT&T Joel Sommers, Colgate University
Benchmarking Framework draft-constantine-bmwg-traffic-management-02
Globecom 2003 December 1-5, San Francisco, California
Statistical Testing Jonas Abromaitis IFM-0/2.
(Long-Term) Reporting Metrics: Different Points of View
Chapter 8: Inference for Proportions
Chapter 8 Inference for Proportions
Introduction to Summary Statistics
IP Performance Specifications - Progress and Next Steps
This teaching material has been made freely available by the KEMRI-Wellcome Trust (Kilifi, Kenya). You can freely download,
draft-ietf-ippm-multipoint-alt-mark-00
Chapter 11: The ANalysis Of Variance (ANOVA)
IQA 2018 QA Basics September 11, 2018 Day 1 Workshop 1.
WGC-2 DG Meeting Towards a Guidance on Groundwater Chemical Status and Threshold Values 14:00 – 16:00 21 April 2008 Ljubljana, Slovenia.
Testing Standards Track Metrics Draft-geib-ippm-metrictest-02
Presentation and project
Presentation transcript:

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-011 Draft-geib-ippm-metrictest-01. Incorporates RFC2330 philosophy, draft-bradner and draft-morton. Inputs and ideas on which this draft is based on:  Draft-morton: Compare the single implementation against the metric specification.  Philosophy RFC2330: IPPM metric implementations measuring simultaneously along an identical path, should result in the same measurement. Validate for a single implementation as well as for different compatible implemenatations. Apply Anderson Darling K-sample test with 95 % confidence (see RFC’s 2330 & 2679).  To be conforming to a metric specification, publish the smallest resolution under which Anderson Darling k- sample test was passed with 95% confidence.  Document the chosen implementation options (and be aware of the possibly resulting limitations for a statistical test comparing different implementations).  Draft-morton (and IETF in general): improve IPPM metric specifications based on implementation experience before promoting them to standards.

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-012 Draft-geib-ippm-metrictest-01. Identical networking conditions for repeated measurements. Metric implementations will be operated in real networks. Metric compliance should be tested under live network conditions too. Identical networking conditions for multiple flows can be reached by:  Setting up a tunnel using IP/MPLS transport between two sites.  Simultaneously measure with 5 or more flows per implementation.  Ensure that the test set up doesn’t interfere with the metric measurement. Metric Implement. A Instance 1 Metric Implement. A Instance 2 Tunnel terminati on 1 Tunnel terminati on 2 Internet Example: „repeating“ measurements under identical network conditions with a single implementation by measuring with two parallel flows.

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-013 Draft-geib-ippm-metrictest-01. Some results with two instances of a single implementation. Unless stated otherwise: two implementations, partially sharing a path, same packet size and same queue. Resolution is 1  s. Data has been normalised to the same average value (ADK is sensitive to variations in averages too). AD2 (95%) test. For more details on measurement set up see ietf_75_ippm_geib_metrictest.pdf.  Single instance, different packet sizes, different queues, low load, normalised on same mean.  Two instances, same queue, same packet size, moderate load, not normalised. DelayMean [ms]Standard Dev. [ms]AD2 Test Flow A passed Flow B JitterMean [ms]Standard Dev. [ms]AD2 Test Path A passed Path B

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-014 Draft-geib-ippm-metrictest-01. More results (1).  Two instances, same queue, same packet size, moderate load, not normalised, 32 samples only.  Single instance, single queue, low load, results split into four contiguous sets of data (“repeated measurement single implementation”), not normalised. Pckt LossMean [pckts]Standard Dev. [pckts]AD2 Test Path A passed Path B DelayMean [ms]Standard Dev. [ms]ADK passedAD2 failed Interval A With B,CWith D Interval B With A, CWith D Interval C With A, C, D Interval D With CWith A, B

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-015 Draft-geib-ippm-metrictest-01. More results (2).  Two instances, same queue, same packet size, low load, data has been normalised. ADK test passed after limiting temporal resolution to 25  s.  Single instance, different packet sizes, different queues, low load, normalised on same mean. ADK test passed after limiting temporal resolution to 150  s. DelayMean [ms]Standard Dev. [ms]AD2 Test Path A100.9 failed Path B DelayMean [ms]Standard Dev. [ms]AD2 Test Flow A failed Flow B

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-016 Draft-geib-ippm-metrictest-01. Next steps.  If the concept is comprehensible and makes sense, is there support to go on?  If “yes”: complete the draft by adding more of Al’s ideas, add figures and so on.  Get a review by design team (name volunteers).  Improve draft again, resubmit and suggest as WG draft.  If the answer is “no” – read the draft and suggest changes.

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-017 Backup

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-018 Draft-geib-ippm-metrictest-01. Prior work: RFC2330 repeatability (precision). RFC2330: „A methodology for a metric should [be] repeatable: if the methodology is used multiple times under identical conditions, the same measurements should result in the same measurements.” Draft-geib: This demands a high precision. By measuring a metric multiple times, probes are drawn from the underlying (and unknown) distribution of networking conditions. Source: Wikipedia High precision, low accuracy High accuracy, low precision

Nov 2009Geib/Morton/ Hasslinger/Fardid / draft-geib-metrictest-019 Draft-geib-ippm-metrictest-01. Prior work: RFC2330/2679 ADK sample (95% confidence). RFC2330: „A methodology for a given metric exhibits continuity if, for small variations in conditions, it results in small variations in the resulting measurements.” Using a different metric implementation under otherwise identical (network) conditions should be a “small variation”. The sample distribution of metric implementation A is taken as the „given“ distribution against which the sample distribution of metric implementation B is compared by a goodness of fit test (proposal: Anderson-Darling k-test). RFC2330 provides guidelines on testing for goodness of fit for calibration (quotes):  Summarizing measurements using histograms, the “EDF” is preferred.  IPPM goodness-of-fit tests are done using 5% significance (see also RFC2679).  …recommends the Anderson-Darling EDF test.