Characteristic Studies of Distributed Systems Maryam Rahmaniheris & Daniel Uhlig 1.

Slides:



Advertisements
Similar presentations
Running Your Startup on Amazon Web Services Alex Iskold Founder/CEO AdaptiveBlue Feature Writer ReadWriteWeb.
Advertisements

A Measurement Study of Peer-to-Peer File Sharing Systems Presented by Cristina Abad.
Creating HIPAA-Compliant Medical Data Applications with Amazon Web Services Presented by, Tulika Srivastava Purdue University.
Cloud Service Models and Performance Ang Li 09/13/2010.
Ivan Pleština Amazon Simple Storage Service (S3) Amazon Elastic Block Storage (EBS) Amazon Elastic Compute Cloud (EC2)
JamesRH  7 major AWS Services (  Amazon E-Commerce Service (ECS)  Amazon.
PROVENANCE FOR THE CLOUD (USENIX CONFERENCE ON FILE AND STORAGE TECHNOLOGIES(FAST `10)) Kiran-Kumar Muniswamy-Reddy, Peter Macko, and Margo Seltzer Harvard.
Web Server Benchmarking Using the Internet Protocol Traffic and Network Emulator Carey Williamson, Rob Simmonds, Martin Arlitt et al. University of Calgary.
Low Cost, Scalable Proteomics Data Analysis Using Amazon's Cloud Computing Services and Open Source Search Algorithms Brian D. Halligan, Ph.D. Medical.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 6 2/13/2015.
Measurement, Modeling, and Analysis of a Peer-2-Peer File-Sharing Workload Presented For Cs294-4 Fall 2003 By Jon Hess.
Measurement, Modeling and Analysis of a Peer-to-Peer File-Sharing Workload Krishna Gummadi, Richard Dunn, Stefan Saroiu Steve Gribble, Hank Levy, John.
1 CS 425 / ECE 428 Distributed Systems Fall 2014 Indranil Gupta (Indy) Measurement Studies Lecture 23 Reading: See links on website All Slides © IG.
Streaming Video Traffic: Characterization and Network Impact Kobus van der Merwe Shubho Sen Chuck Kalmanek
Peer to Peer (P2P) Networks and File sharing. By: Ryan Farrell.
1 School of Computing Science Simon Fraser University, Canada Modeling and Caching of P2P Traffic Mohamed Hefeeda Osama Saleh ICNP’06 15 November 2006.
Fresh Analysis of Streaming Media Stored on the Web Rabin Karki M.S. Thesis Presentation Advisor: Mark Claypool Reader: Emmanuel Agu 10 Jan, 2011.
The War Between Mice and Elephants Presented By Eric Wang Liang Guo and Ibrahim Matta Boston University ICNP
Authors: Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox Publish: HPDC'10, June 20–25, 2010, Chicago, Illinois, USA ACM Speaker: Jia Bao Lin.
An Analysis of Internet Content Delivery Systems Stefan Saroiu, Krishna P. Gommadi, Richard J. Dunn, Steven D. Gribble, and Henry M. Levy Proceedings of.
Measurement, Modeling, and Analysis of a Peer-to-Peer File sharing Workload Krishna P. Gummadi, Richard J. Dunn, Stefan Saroiu, Steven D. Gribble, Henry.
Exploiting Content Localities for Efficient Search in P2P Systems Lei Guo 1 Song Jiang 2 Li Xiao 3 and Xiaodong Zhang 1 1 College of William and Mary,
A. Frank 1 Internet Resources Discovery (IRD) Peer-to-Peer (P2P) Technology (1) Thanks to Carmit Valit and Olga Gamayunov.
Tracking the Evolution of Web Traffic: Felix Hernandez-Campos, Kevin Jeffay F. Donelson Smith IEEE/ACM International Symposium on Modeling, Analysis.
Nikolay Tomitov Technical Trainer SoftAcad.bg.  What are Amazon Web services (AWS) ?  What’s cool when developing with AWS ?  Architecture of AWS 
An Introduction to Cloud Computing. The challenge Add new services for your users quickly and cost effectively.
Introduction to Amazon Web Services (AWS)
Cloud Computing الحوسبة السحابية. subject History of Cloud Before the cloud Cloud Conditions Definition of Cloud Computing Cloud Anatomy Type of Cloud.
Types of Operating System
EE616 Technical Project Video Hosting Architecture By Phillip Sutton.
1 One-Click Hosting Services: A File-Sharing Hideout Demetris Antoniades Evangelos P. Markatos ICS-FORTH Heraklion,
Presentation on Osi & TCP/IP MODEL
INSTALLING MICROSOFT EXCHANGE SERVER 2003 CLUSTERS AND FRONT-END AND BACK ‑ END SERVERS Chapter 4.
Cloud Computing Characteristics A service provided by large internet-based specialised data centres that offers storage, processing and computer resources.
Cloud Computing Dave Elliman 11/10/2015G53ELC 1. Source: NY Times (6/14/2006) The datacenter is the computer!
Sharing Information across Congestion Windows CSE222A Project Presentation March 15, 2005 Apurva Sharma.
1 CS 425 Distributed Systems Fall 2011 Slides by Indranil Gupta Measurement Studies All Slides © IG Acknowledgments: Jay Patel.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
Paper Survey of DHT Distributed Hash Table. Usages Directory service  Very little amount of information, such as URI, metadata, … Storage  Data, such.
Efficient P2P Search by Exploiting Localities in Peer Community and Individual Peers A DISC’04 paper Lei Guo 1 Song Jiang 2 Li Xiao 3 and Xiaodong Zhang.
A Utility-based Approach to Scheduling Multimedia Streams in P2P Systems Fang Chen Computer Science Dept. University of California, Riverside
GFS. Google r Servers are a mix of commodity machines and machines specifically designed for Google m Not necessarily the fastest m Purchases are based.
Networking Basics CCNA 1 Chapter 11.
Performance Testing Test Complete. Performance testing and its sub categories Performance testing is performed, to determine how fast some aspect of a.
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-1.
INTERNET TECHNOLOGIES Week 10 Peer to Peer Paradigm 1.
1 CS 425 / ECE 428 Distributed Systems Fall 2013 Indranil Gupta (Indy) Measurement Studies Lecture 22 Nov Reading: See links on website All Slides.
09/13/04 CDA 6506 Network Architecture and Client/Server Computing Peer-to-Peer Computing and Content Distribution Networks by Zornitza Genova Prodanoff.
#16 Application Measurement Presentation by Bobin John.
Amazon Web Services. Amazon Web Services (AWS) - robust, scalable and affordable infrastructure for cloud computing. This session is about:
1 Design and Implementation of a High-Performance Distributed Web Crawler Polytechnic University Vladislav Shkapenyuk, Torsten Suel 06/13/2006 석사 2 학기.
Course: Cluster, grid and cloud computing systems Course author: Prof
Getting the Most out of Scientific Computing Resources
Getting the Most out of Scientific Computing Resources
Operating Systems : Overview
Scale and Performance in the CoBlitz Large-File Distribution Service
Amazon Storage- S3 and Glacier
Web Caching? Web Caching:.
AWS COURSE DEMO BY PROFESSIONAL-GURU. Amazon History Ladder & Offering.
Replication Middleware for Cloud Based Storage Service
VDN: Virtual Machine Image Distribution Network for Cloud Data Centers
Providing Secure Storage on the Internet
Building a Database on S3
Operating Systems : Overview
AWS Cloud Computing Masaki.
Discretized Streams: A Fault-Tolerant Model for Scalable Stream Processing Zaharia, et al (2012)
Operating Systems : Overview
Operating Systems : Overview
Operating Systems : Overview
Presentation transcript:

Characteristic Studies of Distributed Systems Maryam Rahmaniheris & Daniel Uhlig 1

Distributed System Implementations Characteristics of systems in real world How users and systems behave  P2P  Track clients and traffic on Kazaa  Explore and model usage  Cloud Computing  Experiment with Amazon Web Services  Measure performance metrics 2

Krishna P. Gummadi, Richard J. Dunn, Stefan Saroiu, Steven D. Gribble, Henry M. Levy, And John Zahorjan Measurement, Modeling, and Analysis of Peer-to-Peer File-Sharing Work Load Peer To Peer Characteristics Presented by Daniel Uhlig 3

Modeling of P2P Systems Observe clients and users behavior in real world Used large P2P system: Kazaa  Closed P2P protocal  Fasttrack network (still exists?)  Peaked at 2.3 million users in 2003 (more than Napster’s peak)  Now subscription DRM music store 4

Observing P2P Recorded all Kazaa traffic at U. of Washington  60,000 faculty, staff, and students 203 day trace (late spring semester past the end of fall semester) Protected privacy by making data anonymous Analyze data, develop model  compared to web: Is there a better comparison? 5

Data Log Recorded all Kazaa traffic – Incoming and outgoing data transfer and searches – HTTP traffic with username in header – KazaaLite showed up 60 days into trace Username hardcoded so used IP to differentiated. 20 terabytes of incoming data 1.6 Million Requests 25,000 users 8.85 TB of unique objects Paper used requests from university peers to external peers. Will this bias results? 6

Kazaa Objects Fetch at most once  94% of Kazaa files vs 57% of web objects  99% of Kazaa files fetch at most twice  Clients download objects just once What is an object?  Authors assume immutable object that is unique  Same song Still fetch at most once? - Different filename? - Different bitrate? - Different Length? - Encoded by different user? 7

Users are Patient 8 Web users want instant access, P2P users will wait P2P users are patient Small Objects – 30%  1 hr – 10% nearly a day Large Objects – 50%  1 day – 20% wait 1 week Is this accurate since client automatically restarts requests?

Users Slow Down Took 30 day traces of ‘new’ users Bytes requested decrease with age Possible reasons?  Loss of interest  New P2P app  New ID 9

Users Slow Down Users leave the system forever Users request less data as they age Core clients have constant activity level But, less data request 10

Client Activity How to measure activity: – Logged in vs downloading? Average session length 2.4 minutes – Sessions could be split with a short term break in transfer Activity over lifetime = 5.5% Average transfer 17 minutes, but average session 2.4 minutes Many transactions fail, looking for new host peer 11

Workload Large (>100 MB) vsSmall (<10 MB) files Request Volume vsTransfer volume Audio clips vs Video clips 12

Object Dynamics 13 Clients fetch objects at-most-once Popular objects quickly cycle New objects are most popular Most requests are for old objects Small ObjectsLarge Objects Top 10Top 100Top 10Top 100 Overlap between first and last 30 days0 of 105 of 1001 of 1044 of 100 # of popular objects that a less 30 days old6 of 1073 of 952 of 947 of 56 How does the idea of distinct objects affect this?

Zipf Distribution where most popular objects are most fetched. Classical result for web pages 14

Non-Zipf Authors propose that Kazaa traffic is NOT modeled by Zipf distribution. P2P differences from web – ‘fetch-at-most once’ – Immutable object (cnn.com changes regularly, a multimedia file does not) Simulated model of behaviors and observed 15

Non-Zipf model Zipf seen as model in many places – Video on demand, video rental, movie tickets Non-Zipf might better explain some of these models Common characteristics – Birth of new objects – Fetch at most once – Immutable Characteristics sometimes seen – Expensive to get object – Object remains Does their non-zipf model explain everything? 16

Really Non-Zipf? Multiple copies of the same object? – Does fetch at most once still hold Requests for established files handled by internal users? (‘cached’) Are objects immutable? – Changing names – Is new album/song a new object or an update from artist? Non-Zipf in other multimedia – YouTube, Video Rental, DVD purchase, movie tickets? 17

Locality Awareness Conserve University P2P bandwidth 86% of objects requested already at U of W – Cache data (legal issues) – Redirector so request stay internal when possible Few key nodes can save significant bandwidth 18

Discussion Points What is an unique item – Does this affect the distribution of popular objects? Are objects immutable Apply ideas to other multimedia: – YouTube video popularity – Still fetch at most once – Non-Zipf for DVD rental or purchase How to define a unique object Should P2P handle large and small objects differently Caching or other forced locality vs P2P built in locality 19

Presented by: Maryam Rahmaniheris University of Illinois at Urbana-Champaign CS525-Spring2009 An Evaluation of Amazon's Grid Computing Services: EC2, S3, SQS Simon L. Garfinkel 20

Cluster Computing Building your own cluster  Costly  Space  Cooling system  Staff…  Underutilization  Overutilization Cloud computing  Computing as a utility  Time-multiplexing of resources  No need fro planning ahead  Elasticity  Amazon’s AWS  You only need a working credit card 21

Amazon AWS EC2 (Elastic Compute Cloud)  Linux virtual machines for 10 cents per CPU hour S3 (Simple Storage Service):  Data storage for 15 cents per gigabyte per month SQS (Simple Queue Service):  Messaging service for 10 cents per thousands messages 22

An Example Application- GrepTheWeb 23

An Example Application 24

AWS Interface Creating an Amazon AWS account Signing up for individual services Using REST API to start up virtual machines on EC2 and accessing data on S3  HTTP commands  GET  PUT  DELETE 25

AWS Security Accessing the account information  Stolen account password  Resetting the lost password through  Eavesdropping  Stolen password Does not provide snapshots or backups  Multiple EC2 machines  Multiple S3  Multiple AWS accounts Does not guarantee the privacy of data on S3  Encryption  Digital signatures 26

S3 Evaluation-Test Scenarios Throughput and TPS  A bucket containing 15 objects, three each in sizes:  1 byte  1 Kbyte  1 Mbyte  16 Mbyte  100 Mbyte  Different objects are used to minimize the effect of caching Measuring end-to-end performance:  A series of successive probes  The delay between the probes follows Poisson distribution  Surge experiment  No delay between the queries  Between 1 and 6 threads executing at any moment A total of 137,099 probes in 2007  Single EC2 instance to S3 (32,269)  Distributed test bed (39,269)  Surge experiment (74,808) 27

Average Daily Read Throughput A minor change in the network topology made by Amazon  Introducing additional delay b/w EC2 and S3  Small TCP window size 28

CDF of Read and Write Throughput March 22-April8 29

Results S3 performs better with larger transaction sizes  High per-transaction overhead 40% of 1-byte writes are slower than 10 TPS as opposed to 5% of 1-byte reads  Writes must be committed to at least 2 different clusters  Amazon reliability guarantee The median of 1-mbyte write bandwidth is roughly 5 times faster  Write transactions are being acknowledged when the data is written to cache  when the transaction raises to the cache size the difference disappears 30

Query Variance Low Correlation High Correlation 31

Results Lower correlation for the 1 byte transactions  Sending a second request rather than waiting  Issuing two simultaneous requests for time-critical issues Higher correlation for the 100 Mbyte transactions  Simultaneous requests, once they start providing data, are likely to take similar amount of time 32

Concurrent Performance :Improving S3 performance by issuing concurrent request to the same bucket Performance of 100MB GETs from S3 for one thread and combined threads Surge experiment Two VMs on two EC2 clusters Executing 1 thread for 10 min and then 2 for 10 min and … The experiment was run for 11 hours The Amount of BW received by each thread is cut in half The 6 threads have three times the aggregate BW of 1 thread 33

Other Experiments Availability  From 107,556 non-surge tests consisting of multiple read and write probes:  6 write retries and 3 write errors  4 read retries  100% availability with proper retry mechanism Throughput  From the analysis of 19,630 1 Mbyte and greater transactions:  No write probes with a throughput less than 10 KB/s  6 write probes with a throughput less than 100 KB/s 34

Experience with EC2 and SQS EC2 instances  Fast  Responsive  Reliable  1 unscheduled reboot, no lost data  1 instance freeze, lost data SQS  Simple API  Insert  One message at a time  4 messages per second  Remove  In batches of 256: 5 messages per second  One message at a time: 2.5 messages per second 35

Conclusion EC2 provides ready-to-go VMs at a reasonable cost S3 delivers a high performance only for transactions of size 16Mb or larger  High per-transaction overhead S3 delivers a much higher performance to EC2 than to other locations on the Internet Limited SQS throughput  4-5 transactions per second per thread High availability Security risks 36

Discussion Points High correlation for large object transactions  Load balancer  Larger # of replicas  Scheduling policy  More noticeable variance in smaller scales Why performance of SQS isn’t sufficient for scheduling tasks faster than seconds or slower than several hours? Google AppEngine or Amazon EC2  What are the best candidate applications for each of them? What are the advantages of Amazon AWS over shared resources in Grids? What are the disadvantages over dedicated clusters? What will the funders of research projects feel about providing fund to pay Amazon bills instead of building dedicated clusters? 37