Kapsalakis Giorgos - AM: 1959 HY459 - Internet Measurements Fall 2010.

Slides:



Advertisements
Similar presentations
Online Course Design Online Course Design EVALUATION AND REVISION
Advertisements

Networking Essentials Lab 3 & 4 Review. If you have configured an event log retention setting to Do Not Overwrite Events (Clear Log Manually), what happens.
Module 12: Auditing SQL Server Environments
1 Reading Log Files. 2 Segment Format
Chapter 7 – Transport Layer Protocols
Bug Isolation via Remote Program Sampling Ben Liblit, Alex Aiken, Alice X.Zheng, Michael I.Jordan Presented by: Xia Cheng.
Distributed Systems Fall 2010 Time and synchronization.
Internet Networking Spring 2006 Tutorial 12 Web Caching Protocols ICP, CARP.
Internet Traffic Patterns Learning outcomes –Be aware of how information is transmitted on the Internet –Understand the concept of Internet traffic –Identify.
Discrete-Event Simulation: A First Course Steve Park and Larry Leemis College of William and Mary.
Looking Forward Mike Goodchild. Where is ESRI going? 9.0 –massively expanded toolbox –script management and metadata –Python, JScript, Perl –visual modeling.
Network Traffic Measurement and Modeling CSCI 780, Fall 2005.
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #13 Web Caching Protocols ICP, CARP.
Security in Databases. 2 Srini & Nandita (CSE2500)DB Security Outline review of databases reliability & integrity protection of sensitive data protection.
20101 Synchronization in distributed systems A collection of independent computers that appears to its users as a single coherent system.
Internet Networking Spring 2002 Tutorial 13 Web Caching Protocols ICP, CARP.
The Basics of Software Testing
On the Difficulty of Scalably Detecting Network Attacks Kirill Levchenko with Ramamohan Paturi and George Varghese.
Security in Databases. 2 Outline review of databases reliability & integrity protection of sensitive data protection against inference multi-level security.
Computer Security and Penetration Testing
Swami NatarajanJuly 14, 2015 RIT Software Engineering Reliability: Introduction.
1 The World Wide Web. 2  Web Fundamentals  Pages are defined by the Hypertext Markup Language (HTML) and contain text, graphics, audio, video and software.
Measurement System Analysis Kevin B. Craner Boise State University October 6, 2003.
Monté Carlo Simulation MGS 3100 – Chapter 9. Simulation Defined A computer-based model used to run experiments on a real system.  Typically done on a.
Hands-On Microsoft Windows Server 2008 Chapter 11 Server and Network Monitoring.
FIREWALL TECHNOLOGIES Tahani al jehani. Firewall benefits  A firewall functions as a choke point – all traffic in and out must pass through this single.
SIAM Life Sci/Soc. for Math Biol. Summer Biocalculus: Reflecting the needs of the students Stephen J. Merrill Dept of MSCS Marquette University.
Today’s Lecture application controls audit methodology.
CSE 486/586 CSE 486/586 Distributed Systems PA Best Practices Steve Ko Computer Sciences and Engineering University at Buffalo.
1 Issues in Benchmarking Intrusion Detection Systems Marcus J. Ranum.
Rob Sherwood CS244 Lecture 8: Sound Strategies For Internet Measurement.
COEN 252 Computer Forensics Collecting Network-based Evidence.
 To explain the importance of software configuration management (CM)  To describe key CM activities namely CM planning, change management, version management.
Protocol(TCP/IP, HTTP) 송준화 조경민 2001/03/13. Network Computing Lab.2 Layering of TCP/IP-based protocols.
Using Web Cameras & OneNote to Record Student Performance Increasing the Effectiveness of Deliberate and Reflective Practice Exercises in Skills Courses.
Event Management & ITIL V3
Auditing Information Systems (AIS)
Fundamentals of Proxying. Proxy Server Fundamentals  Proxy simply means acting on someone other’s behalf  A Proxy acts on behalf of the client or user.
1 CS 502: Computing Methods for Digital Libraries Lecture 19 Interoperability Z39.50.
1 RECONSTRUCTION OF APPLICATION LAYER MESSAGE SEQUENCES BY NETWORK MONITORING Jaspal SubhlokAmitoj Singh University of Houston Houston, TX Fermi National.
Chapter 10 Verification and Validation of Simulation Models
Remote Data Access with OPeNDAP Dr. Dennis Heimbigner Unidata netCDF Workshop October 25, 2012.
Lesson 8—Networking BASICS1 Networking BASICS Managing a Local Area Network Unit 2 Lesson 8.
Chapter 24 Transport Control Protocol (TCP) Layer 4 protocol Responsible for reliable end-to-end transmission Provides illusion of reliable network to.
1 Microsoft Windows 2000 Network Infrastructure Administration Chapter 4 Monitoring Network Activity.
ASP-2-1 SERVER AND CLIENT SIDE SCRITPING Colorado Technical University IT420 Tim Peterson.
Firewalls A brief introduction to firewalls. What does a Firewall do? Firewalls are essential tools in managing and controlling network traffic Firewalls.
LECTURE 02: EVALUATING MODELS January 27, 2016 SDS 293 Machine Learning.
1 Chapter Overview Monitoring Access to Shared Folders Creating and Sharing Local and Remote Folders Monitoring Network Users Using Offline Folders and.
T EST T OOLS U NIT VI This unit contains the overview of the test tools. Also prerequisites for applying these tools, tools selection and implementation.
1 CS 430: Information Discovery Lecture 26 Architecture of Information Retrieval Systems 1.
IP packet filtering Breno de Medeiros. Florida State University Fall 2005 Packet filtering Packet filtering is a network security mechanism that works.
Distributed Systems Lecture 5 Time and synchronization 1.
Troubleshooting Windows Vista Lesson 11. Skills Matrix Technology SkillObjective DomainObjective # Troubleshooting Installation and Startup Issues Troubleshoot.
Science, Measurement, Uncertainty and Error1 Science, Measurements, Uncertainty and Error.
Testing and Debugging PPT By :Dr. R. Mall.
Introduction to Networking
Internet Networking recitation #12
Chapter 10 Verification and Validation of Simulation Models
It is great that we automate our tests, but why are they so bad?
The Application Lifecycle
CS 1111 Introduction to Programming Fall 2018
Lecture 4 Communication Network Protocols
Traffic Analysis– Wireshark Simple Example
System Testing.
Performance Evaluation of Computer Networks
Performance Evaluation of Computer Networks
Protocol Application TCP/IP Layer Model
Internet Research Group at Clemson University
Presentation transcript:

Kapsalakis Giorgos - AM: 1959 HY459 - Internet Measurements Fall 2010

Strategic Areas: Errors & imperfections. Dealing with large volumes of data. Ensuring reproducible analysis. Community datasets.

Precision Precision: Maximum exactness that a tool’s design permits. When reporting measurements, indicate the tool’s precision When does extra effort matter? Unfortunately no crisp rules Discuss whether or not it matters

Meta-data Preserving the information during the course of analysis Many convenient data formats lack a way to annotate measurements, e.g. Today’s web server log, Tcpdump Simple, line-oriented structure no API available for retrieving or setting the value

Meta-data Uniform format for the meta-data is needed Data have a lifetime beyond what the researcher initially envisions Good data is hard to gather Revisiting datasets in new contexts It is beneficial to retain meta-data information even when doing so is not of immediate benefit.

Accuracy How well does the measured abstraction indeed match the actual phenomenon? Much broader problem than precision E.g. Clocks can: Arbitrarily off from true time Jump backward or forward Run arbitrarily fast or slow Fail to move E.g. Packet filters can: Fail to record packets (“drops”) Fail to report drops Report drops that did not occur Reorder packets Duplicate packets Record wrong packets Measurement tools are often deficient in adequately recording failure information.

Misconception Errors in equating what we are actually measuring with what we wish to measure E.g. Computing the distribution of TCP connection sizes By capturing SYN and FIN packets Using the difference between the sequence numbers Miss/fail to include packets SYNs or FINs Very largest connections started before we start tracing Not terminated when we finish General strategy against misconceptions To gather more than one type of dataset From a different location From a different time Seek out early peer review

Calibration A set of techniques that help with detecting problems of inaccuracy, misconception, errors in analysis. Examining outliers and spikes Employing self-consistency checks Measuring facets of the same phenomenon different ways and comparing Evaluating synthetic data Not achieve perfection but achieve confidence

Examining outliers and spikes Outliers unusually low or high values Spikes values that repeat a great deal Represent “corner cases” at the extremes of measurement where problems often manifest Reflect measurement errors, analysis errors, misconception

Employing self-consistency checks Testing whether properties that must hold do in fact hold, e.g. How to find evidence of packet filter drops: Use cumulative TCP ACK properties To see whether each ACK present in a trace at the point in time it was sent All of the data up to the sequence number of ACK has been seen previously in the trace

Comparing multiple measurements Measuring the same phenomenon different ways and comparing the results Conducting additional measurement Multiple versions of analysis

Evaluating Synthetic Data Hand-edit some measurements to introduce changes Test whether software processes it correctly Monte Carlo simulations Simulate multiple random instances

Dealing with large volume of data Large scale leads to a number of potential difficulties System limitations such as disk space, maximum file sizes, number of files on a volume, …. Many systems used for statistical analysis have upper bounds on the amount of data they can process Strategy : Indepth analysis of small subsets

Ensuring reproducible analysis The analysis derived from the data is reproducible Strategy: Structured analysis Single master script Cataloging notebook Benefits of structuring analysis Reproduce the results, Minimizing the headaches Explore the analysis of the data in a consistent fashion

Making datasets publicly available Dearth of publicly available datasets Strategy: periodically analyze ongoing measurements let’s you discover when data acquisition broken ensures you’re collecting (some) meta-data Problems Reluctance, legal impediments Privacy, security and business sensitivities - anonymization Strategy: package analysis for “data reduction requests” send data analysis software to dataset holder they run it, inspect results, & return them