Web Prefetching Between Low-Bandwidth Clients and Proxies : Potential and Performance Li Fan, Pei Cao and Wei Lin Quinn Jacobson (University of Wisconsin-Madsion)

Slides:



Advertisements
Similar presentations
Summary Cache: A Scalable Wide-Area Web Cache Sharing Protocol Li Fan, Pei Cao and Jussara Almeida University of Wisconsin-Madison Andrei Broder Compaq/DEC.
Advertisements

Dissemination-based Data Delivery Using Broadcast Disks.
Background Virtual memory – separation of user logical memory from physical memory. Only part of the program needs to be in memory for execution. Logical.
Consistency and Replication Chapter 7 Part II Replica Management & Consistency Protocols.
G. Alonso, D. Kossmann Systems Group
Latency-sensitive hashing for collaborative Web caching Presented by: Xin Qi Yong Yang 09/04/2002.
Workloads Experimental environment prototype real sys exec- driven sim trace- driven sim stochastic sim Live workload Benchmark applications Micro- benchmark.
EEC-484/584 Computer Networks Lecture 6 Wenbing Zhao
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
1 Prefetching the Means for Document Transfer: A New Approach for Reducing Web Latency 1. Introduction 2. Data Analysis 3. Pre-transfer Solutions 4. Performance.
Web Prefetch 張燕光 資訊工程系 成功大學
Internet Networking Spring 2006 Tutorial 12 Web Caching Protocols ICP, CARP.
Hypertext Transfer Protocol Kyle Roth Mark Hoover.
An Analysis of Internet Content Delivery Systems Stefan Saroiu, Krishna P. Gommadi, Richard J. Dunn, Steven D. Gribble, and Henry M. Levy Proceedings of.
Scalable and Continuous Media Streaming on Peer-to-Peer Networks M. Sasabe, N. Wakamiya, M. Murata, H. Miyahara Osaka University, Japan Presented By Tsz.
EEC-484/584 Computer Networks Discussion Session for HTTP and DNS Wenbing Zhao
Improving Proxy Cache Performance: Analysis of Three Replacement Policies Dilley, J.; Arlitt, M. A journal paper of IEEE Internet Computing, Volume: 3.
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #13 Web Caching Protocols ICP, CARP.
Internet Networking Spring 2002 Tutorial 13 Web Caching Protocols ICP, CARP.
Differentiated Multimedia Web Services Using Quality Aware Transcoding S. Chandra, C.Schlatter Ellis and A.Vahdat InfoCom 2000, IEEE Journal on Selected.
Web Caching Robert Grimm New York University. Before We Get Started  Illustrating Results  Type Theory 101.
Web Latency Reduction Via Client-side Prefetching Avinoam N. Eden; Brian W. Joh; Trevor Mudge Electrical Engineering and Conputer Science Department The.
Design and Implementation of a Server Director Project for the LCCN Lab at the Technion.
Caching And Prefetching For Web Content Distribution Presented By:- Harpreet Singh Sidong Zeng ECE Fall 2007.
1Bloom Filters Lookup questions: Does item “ x ” exist in a set or multiset? Data set may be very big or expensive to access. Filter lookup questions with.
An Intelligent Cache System with Hardware Prefetching for High Performance Jung-Hoon Lee; Seh-woong Jeong; Shin-Dug Kim; Weems, C.C. IEEE Transactions.
Web Caching Schemes For The Internet – cont. By Jia Wang.
1 The Mystery of Cooperative Web Caching 2 b b Web caching : is a process implemented by a caching proxy to improve the efficiency of the web. It reduces.
Peer-to-peer Multimedia Streaming and Caching Service by Won J. Jeon and Klara Nahrstedt University of Illinois at Urbana-Champaign, Urbana, USA.
Application Layer  We will learn about protocols by examining popular application-level protocols  HTTP  FTP  SMTP / POP3 / IMAP  Focus on client-server.
1 ENHANCHING THE WEB’S INFRASTUCTURE: FROM CACHING TO REPLICATION ECE 7995 Presented By: Pooja Swami and Usha Parashetti.
COMPUTER TERMS PART 1. COOKIE A cookie is a small amount of data generated by a website and saved by your web browser. Its purpose is to remember information.
By Ravi Shankar Dubasi Sivani Kavuri A Popularity-Based Prediction Model for Web Prefetching.
Achieving Load Balance and Effective Caching in Clustered Web Servers Richard B. Bunt Derek L. Eager Gregory M. Oster Carey L. Williamson Department of.
Hybrid Prefetching for WWW Proxy Servers Yui-Wen Horng, Wen-Jou Lin, Hsing Mei Department of Computer Science and Information Engineering Fu Jen Catholic.
Introduction The large amount of traffic nowadays in Internet comes from social video streams. Internet Service Providers can significantly enhance local.
1 Reading Report 9 Yin Chen 29 Mar 2004 Reference: Multivariate Resource Performance Forecasting in the Network Weather Service, Martin Swany and Rich.
Storage Allocation in Prefetching Techniques of Web Caches D. Zeng, F. Wang, S. Ram Appeared in proceedings of ACM conference in Electronic commerce (EC’03)
Design and Implement an Efficient Web Application Server Presented by Tai-Lin Han Date: 11/28/2000.
Infrastructure for Better Quality Internet Access & Web Publishing without Increasing Bandwidth Prof. Chi Chi Hung School of Computing, National University.
Web Cache Replacement Policies: Properties, Limitations and Implications Fabrício Benevenuto, Fernando Duarte, Virgílio Almeida, Jussara Almeida Computer.
Enabling Class of Service for CIOQ Switches with Maximal Weighted Algorithms Thursday, October 08, 2015 Feng Wang Siu Hong Yuen.
The Case for Persistent-Connection HTTP Telecommunication System LAB 최 명길 Western Research Laboratory Research Report 95/4 (Proceedings of the SIGCOMM.
1 University of Maryland Linger-Longer: Fine-Grain Cycle Stealing in Networks of Workstations Kyung Dong Ryu © Copyright 2000, Kyung Dong Ryu, All Rights.
A Measurement Based Memory Performance Evaluation of High Throughput Servers Garba Isa Yau Department of Computer Engineering King Fahd University of Petroleum.
Web Performance 성민영 SNU Computer Systems lab.. 2 차례 4 Modeling the Performance of HTTP Over Several Transport Protocols. 4 Summary Cache : A Scaleable.
-1- System Software Research Lab. Active Cache : Caching Dynamic Contents on the Web Pei Cao, Jin Zhang and Kevin Beach Middleware’ SSLAB,
Dr. Yingwu Zhu Summary Cache : A Scalable Wide- Area Web Cache Sharing Protocol.
ICOM 6115: Computer Systems Performance Measurement and Evaluation August 11, 2006.
Load-Balancing Routing in Multichannel Hybrid Wireless Networks With Single Network Interface So, J.; Vaidya, N. H.; Vehicular Technology, IEEE Transactions.
An Effective Disk Caching Algorithm in Data Grid Why Disk Caching in Data Grids?  It takes a long latency (up to several minutes) to load data from a.
Empirical Quantification of Opportunities for Content Adaptation in Web Servers Michael Gopshtein and Dror Feitelson School of Engineering and Computer.
Performance of Web Proxy Caching in Heterogeneous Bandwidth Environments IEEE Infocom, 1999 Anja Feldmann et.al. AT&T Research Lab 발표자 : 임 민 열, DB lab,
HTTP evolution - TCP/IP issues Lecture 4 CM David De Roure
Measuring the Capacity of a Web Server USENIX Sympo. on Internet Tech. and Sys. ‘ Koo-Min Ahn.
MiddleMan: A Video Caching Proxy Server NOSSDAV 2000 Brian Smith Department of Computer Science Cornell University Ithaca, NY Soam Acharya Inktomi Corporation.
Energy Efficient Prefetching and Caching Athanasios E. Papathanasiou and Michael L. Scott. University of Rochester Proceedings of 2004 USENIX Annual Technical.
Web Prefetching Lili Qiu Microsoft Research March 27, 2003.
1 COMP 431 Internet Services & Protocols HTTP Persistence & Web Caching Jasleen Kaur February 11, 2016.
09/13/04 CDA 6506 Network Architecture and Client/Server Computing Peer-to-Peer Computing and Content Distribution Networks by Zornitza Genova Prodanoff.
Web Proxy Caching: The Devil is in the Details Ramon Caceres, Fred Douglis, Anja Feldmann Young-Ho Suh Network Computing Lab. KAIST Proceedings of the.
1 Evaluation of Cooperative Web Caching with Web Polygraph Ping Du and Jaspal Subhlok Department of Computer Science University of Houston presented at.
WWW and HTTP King Fahd University of Petroleum & Minerals
E-commerce | WWW World Wide Web - Concepts
E-commerce | WWW World Wide Web - Concepts
Web Caching? Web Caching:.
Internet Networking recitation #12
Kalyan Boggavarapu Lehigh University
Outline System architecture Current work Experiments Next Steps
Lecture 1: Bloom Filters
Presentation transcript:

Web Prefetching Between Low-Bandwidth Clients and Proxies : Potential and Performance Li Fan, Pei Cao and Wei Lin Quinn Jacobson (University of Wisconsin-Madsion) SIGMETRICS 황호영, CNR Lab. Web prefetching 중에서도 Low-Bandwidth Modem Client 에 대한 논문으로, Client 와 Proxy 사이의 prefetching 에 대해서 언급하 고 있습니다. Originality 가 떨어지는 논문으 로 크게 추천하지는 않습니다. Web prefetching 중에서도 Low-Bandwidth Modem Client 에 대한 논문으로, Client 와 Proxy 사이의 prefetching 에 대해서 언급하 고 있습니다. Originality 가 떨어지는 논문으 로 크게 추천하지는 않습니다.

Communication Networks Research Lab Introduction 2. Proxy-Initiated Prefetching 3. Traces and Simulator 4. Reducing Client Latency 5. Prediction Algorithm 6. Performance 7. Implementation Experience 8. Conclusion and Critique Content

Communication Networks Research Lab. 3 One approach to reduce latency prefetching between caching proxies and browsers. The majority of the Internet population access the WWW via dial-up modem connections. The low modem bandwidth is a primary contributor to client latency. Investigate one technique to reduce latency for modem users. 1. Introduction

Communication Networks Research Lab. 4 Proxy-Initiated Prefetching The proxy can often predict what objects a user might access next. The modem link to the user often has idle periods as the user is reading the current Web document. If the objects are cached at the proxy, the proxy can utilize the idle periods to push them to user, or to have the browser pull them. Since the proxy only initiates prefetches for objects in its cache, there is no extra Internet traffic. 2. Proxy-Initiated Prefetching (1/3)

Communication Networks Research Lab. 5 Assumptions Users have idle times between requests, because users often read some parts of one document before jumping to the next one. The proxy can predict which Web pages a user will access in the near future based on reference patterns observed from many users The proxy has a cache that hold recently accessed Web pages. Proxy maintain a history structure Every time the proxy services a request, it updates the history structure, establishing the connection between past accesses made by the same user and the current request. In browser cache, assume LRU(Least-Recently-Used) algorithm 2. Proxy-Initiated Prefetching (2/3)

Communication Networks Research Lab. 6 Performance Metrics Request Savings : the number of times that a user request hits in the browser cache or the requested object is being prefetched, in percentages of the total number of user requests.  Prefetched  Cached  Partially Prefetched Latency Reduction : the reduction in client latency, in percentages Wasted Bandwidth : the sum of bytes that are prefetched but are not read by the client 2. Proxy-Initiated Prefetching (3/3)

Communication Networks Research Lab. 7 Traces We use the HTTP traces gathered from the University of California at Berkeley home dial-up populations from November 14~19, Simulator The simulator uses timing information in the traces to estimate latency seen by each modem client. The simulator assumes that each modem link has a bandwidth of 21kb/s. The simulator assumes the existence of a proxy between the modem clients and the Internet. (16GB : proxy cache size) 3. Traces and the Simulator

Communication Networks Research Lab. 8 Increase the size of browser cache Use delta compression to transfer modified Web pages between the proxy and clients Apply application-level compression to HTML pages 4. Reducing Client Latency (1/2)

Communication Networks Research Lab. 9 Cumulative distribution of user idle time in the UCB traces About 40% of the requests are preceded by 2 to 128 seconds of idle time, indicating plenty of prefetching opportunities. 4. Reducing Client Latency (2/2)

Communication Networks Research Lab. 10 The realistic prediction algorithm is based on the Prediction-by-Partial-Matching (PPM) PPM Predictors m : prefix depth (# of past accesses that are used to predict future ones) l : search depth (# of steps that the algorithm tries to predict into the future) t : threshold (only candidates whose probability of access is higher than t, 0  t  1, is considered for prefetching) Past m references are matched against the collection of trees to produce the set of URLs for the next l steps. Only URLs whose frequencies of accesses are larger than t are included. 5. Prediction Algorithms (1/4)

Communication Networks Research Lab. 11 PPM Predictors Finally, the URLs are sorted first by giving preferences to longer prefixes, and then by giving preferences to URLs with higher probability within the same prefix. Previous proposed prefetching algorithms Papadumanta and Mogul -> m always equal to 1 Krishna and Vitter -> l always equal to 1 m>1 : more contexts might improve the accuracy of the prediction l>1 : URL is not always requested as the immediate next request after another URL, but rather than within the next few requests. Best performing : m=2, l=4 5. Prediction Algorithms (2/4)

Communication Networks Research Lab. 12 History Structure The history structure is a forest of trees of a fixed depth K, where K=m+l The history encodes all sequences of accesses up to a maximum length K. The history structure is updated every time a user makes a request. being updated for user sequence A…B…C (K=3) 5. Prediction Algorithms (3/4)

Communication Networks Research Lab. 13 Every time the modem link to a user is idle, the proxy calls the predictor for the list of candidate URLs. The proxy then initiates prefetching of the objects in the order specified in the list. When the user makes a new request, the ongoing prefetching is stopped, and a new round of prediction and prefetching starts again next time. The size of history structure can be controlled using LRU algorithm. 5. Prediction Algorithms (4/4)

Communication Networks Research Lab. 14 Performance of Proxy-Initiated Prefetching 6. Performance (1/6)

Communication Networks Research Lab. 15 Assumption prefetch threshold : 50KB, 8 objects browser cache : 16MB(extended), LRU replacement algorithm Performance of Proxy-Initiated Prefetching decreasing the threshold t increases the wasted bandwidth and helps to generate enough candidates.  for l>1, t=0.25 is the best choice. increasing the search depth l increases both the latency reduction and the wasted bandwidth.  l=4 appears the best choice, as larger l makes little difference increasing the prefix depth m increases both the latency reduction and the wasted bandwidth. 6. Performance (2/6)

Communication Networks Research Lab. 16 The accuracy of the prediction algorithm 6. Performance (3/6)

Communication Networks Research Lab. 17 The accuracy of the prediction algorithm attempted : the total number of candidates suggested by the predictor prefetched : the actual number of objects that are prefetched used : the number of objects that are prefetched and actually accessed by the user The ratio between used and prefetched is the accuracy of the prediction algorithm accuracy range : 40% (2,4,0.125) ~ 73% (1,1,0.5) low threshold configurations appear to sacrifice accuracy for more prefetches. 6. Performance (4/6)

Communication Networks Research Lab. 18 Recommendations for the configuration of PPM If the highest latency reduction is the goal and some amount of wasted bandwidth can be tolerated  (2,4,0.125) is the best choice. If both high latency reduction and low wasted bandwidth are desired  (2,4,0.5) is the best choice. If limits on storage requirements make smaller m and l desirable,  (2,1,0.25) and (1,1,0.125) are good choices. 6. Performance (5/6)

Communication Networks Research Lab. 19 Effects of Implementation Variations No proxy notification upon browser cache hits : no-notice Prefetching without knowledge of the content of browser caches : oblivious Limiting the size of history structure 6. Performance (6/6)

Communication Networks Research Lab. 20 Proxy-initiated prefetching we have implemented proxy-initiated prefetching in the CERN httpd proxy software. CERN httpd uses a process-based structure and forks a new process each time a new request arrives. A separate predictor process communicates with other processes via UDP messages. The predictor runs in an infinite loop, waiting to receive updates and queries messages. The process checks a shared global array of flags to see whether the modem link is idle. If it is, starts pushing the URL objects on the existing connection. 7. Implementation Experience (1/2)

Communication Networks Research Lab. 21 At the client side, instead of modifying browsers, we set up a copy of the CERN httpd proxy. The browser requests are first sent to the local proxy. The local proxy manages its own cache, issues requests to the main proxy, and receives pushed objects. Measurement emulate modem connections on the LAN, and generate workloads that reflect typical browser behavior and Internet latencies. We have instrumented the Linux kernel to simulate modem connections on our Ethernet LAN 7. Implementation Experience (2/2)

Communication Networks Research Lab. 22 Conclusion We have investigated the potential and performance of one technique, prefetching between the low-bandwidth clients and caching proxies, and found that combined with delta compression It can reduce user-visible latency by over 23% Prediction algorithm based on the PPM compressor perform well. The technique is easy to implement and can have a considerable effect on user’s Web surfing experience. 8. Conclusion and Critique (1/2)

Communication Networks Research Lab. 23 Weakness We assumed fixed user request arrival times in simulation Our calculation of client latency is merely an estimate based on the time stamps recorded in the traces and the modem bandwidth We also does not model the proxy in detail We have not investigated the implementation of delta compression. Didn’t consider CPU overhead and delay PPM algorithm is a little different from previous proposed prefetching algorithms. 8. Conclusion and Critique (2/2)