Hierarchical Caching and Prefetching for Continuous Media Servers with Smart Disks By:Amandeep Singh Parth Kushwaha.

Slides:



Advertisements
Similar presentations
Chapter 6 I/O Systems.
Advertisements

Chapter 13: I/O Systems I/O Hardware Application I/O Interface
Chapter 1 Introduction Copyright © Operating Systems, by Dhananjay Dhamdhere Copyright © Introduction Abstract Views of an Operating System.
CPS216: Data-Intensive Computing Systems Data Access from Disks Shivnath Babu.
SE-292 High Performance Computing
1 Mobility-Based Predictive Call Admission Control and Bandwidth Reservation in Wireless Cellular Networks Fei Yu and Victor C.M. Leung INFOCOM 2001.
Disk Storage SystemsCSCE430/830 Disk Storage Systems CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu (U. Maine) Fall,
Conserving Disk Energy in Network Servers ACM 17th annual international conference on Supercomputing Presented by Hsu Hao Chen.
Paper by: Chris Ruemmler and John Wikes Presentation by: Timothy Goldberg, Daniel Sink, Erin Collins, and Tony Luaders.
- Dr. Kalpakis CMSC Dr. Kalpakis 1 Outline In implementing DBMS we need to answer How should the system store and manage very large amounts of data?
I/O Management and Disk Scheduling
MEMORY popo.
The Performance Impact of Kernel Prefetching on Buffer Cache Replacement Algorithms (ACM SIGMETRIC 05 ) ACM International Conference on Measurement & Modeling.
1 Sizing the Streaming Media Cluster Solution for a Given Workload Lucy Cherkasova and Wenting Tang HPLabs.
SE-292 High Performance Computing
SE-292 High Performance Computing Memory Hierarchy R. Govindarajan
Lecture 8: Memory Hierarchy Cache Performance Kai Bu
Miss Penalty Reduction Techniques (Sec. 5.4) Multilevel Caches: A second level cache (L2) is added between the original Level-1 cache and main memory.
Performance of Cache Memory
CS Spring 2009 CS 414 – Multimedia Systems Design Lecture 28 – Media Server (Part 3) Klara Nahrstedt Spring 2009.
Continuous Media 1 Differs significantly from textual and numeric data because of two fundamental characteristics: –Real-time storage and retrieval –High.
Distributed Multimedia Systems
Study of Hurricane and Tornado Operating Systems By Shubhanan Bakre.
CS Spring 2011 CS 414 – Multimedia Systems Design Lecture 27 – Media Server (Part 3) Klara Nahrstedt Spring 2011.
Input/Output Management and Disk Scheduling
1 Adaptive Live Broadcasting for Highly-Demanded Videos Hung-Chang Yang, Hsiang-Fu Yu and Li-Ming Tseng IEEE International Conference on Parallel and Distributed.
Computational Astrophysics: Methodology 1.Identify astrophysical problem 2.Write down corresponding equations 3.Identify numerical algorithm 4.Find a computer.
Double buffer SDRAM Memory Controller Presented by: Yael Dresner Andre Steiner Instructed by: Michael Levilov Project Number: D0713.
Locality-Aware Request Distribution in Cluster-based Network Servers 1. Introduction and Motivation --- Why have this idea? 2. Strategies --- How to implement?
Distributed Multimedia Streaming over Peer-to-Peer Network Jin B. Kwon, Heon Y. Yeom Euro-Par 2003, 9th International Conference on Parallel and Distributed.
End-to-End Analysis of Distributed Video-on-Demand Systems P. Mundur, R. Simon, and A. K. Sood IEEE Transactions on Multimedia, Vol. 6, No. 1, Feb 2004.
A Novel Video Layout Strategy for Near-Video-on- Demand Servers Shenze Chen & Manu Thapar Hewlett-Packard Labs 1501 Page Mill Rd. Palo Alto, CA
1 Storage Hierarchy Cache Main Memory Virtual Memory File System Tertiary Storage Programs DBMS Capacity & Cost Secondary Storage.
SECTIONS 13.1 – 13.3 Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin SECONDARY STORAGE MANAGEMENT.
A Server-less Architecture for Building Scalable, Reliable, and Cost-Effective Video-on-demand Systems Presented by: Raymond Leung Wai Tak Supervisor:
CS 551-Memory Management1 Learning Objectives Centralized Memory Management -review Simple Memory Model Shared Memory Model Distributed Shared Memory Memory.
Device Management.
Efficient Support for Interactive Browsing Operations in Clustered CBR Video Servers IEEE Transactions on Multimedia, Vol. 4, No.1, March 2002 Min-You.
Memory access scheduling Authers: Scott RixnerScott Rixner,William J. Dally,Ujval J. Kapasi, Peter Mattson, John D. OwensWilliam J. DallyUjval J. KapasiPeter.
Web-Conscious Storage Management for Web Proxies Evangelos P. Markatos, Dionisios N. Pnevmatikatos, Member, IEEE, Michail D. Flouris, and Manolis G. H.
1 CS222: Principles of Database Management Fall 2010 Professor Chen Li Department of Computer Science University of California, Irvine Notes 01.
An Intelligent Cache System with Hardware Prefetching for High Performance Jung-Hoon Lee; Seh-woong Jeong; Shin-Dug Kim; Weems, C.C. IEEE Transactions.
SECTIONS 13.1 – 13.3 Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin SECONDARY STORAGE MANAGEMENT.
CS Spring 2012 CS 414 – Multimedia Systems Design Lecture 34 – Media Server (Part 3) Klara Nahrstedt Spring 2012.
Introduction to Database Systems 1 The Storage Hierarchy and Magnetic Disks Storage Technology: Topic 1.
Buffer Management for Shared- Memory ATM Switches Written By: Mutlu Apraci John A.Copelan Georgia Institute of Technology Presented By: Yan Huang.
CS4432: Database Systems II Data Storage (Better Block Organization) 1.
Data Management for Decision Support Session-5 Prof. Bharat Bhasker.
Toolbox for Dimensioning Windows Storage Systems Jalil Boukhobza, Claude Timsit 12/09/2006 Versailles Saint Quentin University.
E0262 MIS - Multimedia Playback Systems Anandi Giridharan Electrical Communication Engineering, Indian Institute of Science, Bangalore – , India.
CPU Cache Prefetching Timing Evaluations of Hardware Implementation Ravikiran Channagire & Ramandeep Buttar ECE7995 : Presentation.
TRACK-ALIGNED EXTENTS: MATCHING ACCESS PATTERNS TO DISK DRIVE CHARACTERISTICS J. Schindler J.-L.Griffin C. R. Lumb G. R. Ganger Carnegie Mellon University.
1 I/O Management and Disk Scheduling Chapter Categories of I/O Devices Human readable Used to communicate with the user Printers Video display terminals.
1. Memory Manager 2 Memory Management In an environment that supports dynamic memory allocation, the memory manager must keep a record of the usage of.
Distributing Layered Encoded Video through Caches Authors: Jussi Kangasharju Felix HartantoMartin Reisslein Keith W. Ross Proceedings of IEEE Infocom 2001,
임규찬. 1. Abstract 2. Introduction 3. Design Goals 4. Sample-Based Scheduling for Parallel Jobs 5. Implements.
An I/O Simulator for Windows Systems Jalil Boukhobza, Claude Timsit 27/10/2004 Versailles Saint Quentin University laboratory.
The Vesta Parallel File System Peter F. Corbett Dror G. Feithlson.
CPSC 404, Laks V.S. Lakshmanan1 External Sorting Chapter 13: Ramakrishnan & Gherke and Chapter 2.3: Garcia-Molina et al.
Lecture 40: Review Session #2 Reminders –Final exam, Thursday 3:10pm Sloan 150 –Course evaluation (Blue Course Evaluation) Access through.
Storing and Serving Multimedia. What is a Media Server? A scalable storage manager Allocates multimedia data optimally among disk resources Performs memory.
Multimedia Retrieval Architecture Electrical Communication Engineering, Indian Institute of Science, Bangalore – , India Multimedia Retrieval Architecture.
CS Spring 2009 CS 414 – Multimedia Systems Design Lecture 27 – Media Server (Part 2) Klara Nahrstedt Spring 2009.
COMPUTER SYSTEMS ARCHITECTURE A NETWORKING APPROACH CHAPTER 12 INTRODUCTION THE MEMORY HIERARCHY CS 147 Nathaniel Gilbert 1.
Part IV I/O System Chapter 12: Mass Storage Structure.
Memory Management.
CS 414 – Multimedia Systems Design Lecture 31 – Media Server (Part 5)
Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin
Persistence: hard disk drive
13.3 Accelerating Access to Secondary Storage
Presentation transcript:

Hierarchical Caching and Prefetching for Continuous Media Servers with Smart Disks By:Amandeep Singh Parth Kushwaha

Index: Introduction Media Servers Proposed Algorithms -Sweep and Prefetch(S&P) -Gradual Prefetching -Grouped Periodic Multiground Prefetching(GPMP) Performance Evaluation The Experiment Results Conclusion Future Use References

Introduction Due to increase in CPU performance,I/O systems have become the performance bottleneck. Reason for this performance bottleneck is mechanical movement of disk head. Several algorithms are introduced that exploit emerging smart disk technologies and increase data throughput on media servers.

Media Server media server is a device that simply stores and shares audio,video media files Media server offers client access to media types such as video,audio and so on. To avoid glitches media server must retrieve data from secondary memory at specific rate.

Working of Media server To avoid glitches media server require double buffer in main memory to ensure a data streams continuous playback. Data read from the disk fills half of buffer while other half is use to play video When data in first buffer is consumed,a switch occurs and media server use the second buffer to play video and uses the empty buffer for storage.

Contd.. Media server serve multiple requests concurrently Therefore media server serves streams in rounds, which is of specific length During a round system reads one block for each stream.

Contd..

Each request which is to served in a round is added to service list A resolution mechanism maps each requested video block to disk block I/O controller routes this requested disk block list to disk drive Then the low level schedular schedules the corresponding disk block for each drive,reducing disk head positioning overhead

contd.. The disk head then transfers the requested data from the disk surface to the disk buffer cache. With the help of I/O bus this buffer cache data transferred to server RAM

Contd.. Several algorithms are introduced to achieve maximum throughput, which is maximum number of streams served by disk. Two important factors that limits maximum throughput are: 1.Time required to transfer data from the disk to main memory 2.Cost of cache memory

Proposed Algorithms In these proposed algorithm we increase maximum throughput while keeping round duration constant,with relatively low memory requirement. To improve the drives maximum stream throughput,our algorithm use caching and prefetching techniques. It actually increase throughput because prefetched requests are served from the disks cache,without head –positioning –overhead.

SWEEP AND PREFETCH ALGORITHM Data blocks that are retrieved from the disk are called randomly retrieved blocks.This causes more head-positioning overhead In this algorithm disk head prefetches blocks when it reads adjacent blocks and stores them in a cache.This cause no head-positioning overhead

CONTD.. In this algorithm maximum reandomly retrieved blocks are 25 in one round The ratio of randomly retrieved block that can be exchanged with prefetched blocks is 5/8.It means for retrieving 8 prefetched blocks required as much time to retrieve 5 random blocks

Contd.. During round 1 and round2 of fig(a) all the blocks are randomly retrieved. During round 1of fig(b) last five blocks are served through high level cache buffer As the ratio of randomly retrieved blocks exchanged with prefetched blocks is 5/8 so we can exchange last 5 blocks by prefetching 8 blocks from the disk.

Contd.. During round2 of the fig(2) first 8 blocks are already prefetched so we dont have to retrived these blocks from disk. In place of these 8 preftched blocks we can add 3 randomly accessed blocks and 8 prefetched blocks

Issues With Sweep & Prefetch: Until maximum throughput is achieved S&P services without prefetch. Requires 3 cache buffers for each stream that is cached at the higher level. This requires 3 blocks from each stream. – 2 for the double buffer. – 1 for the multi-disk controller.

Thus each streams block will be skipped in some round. Creating extra startup latency.

Gradual Prefetching: Force the server to work under S&P all the time. Dont consider the number of concurrent streams. At any time disk head prefetches half of all supported streams. For every 2 newly admitted streams, 1 will have its next block prefetched in the first round

It works with maximum number of supported streams so always time to prefetch an additional block for half of all new streams. Hence, no stream will have extra start-up latency and triple buffering is no longer required.

Random retrievals (v) = 19 Prefetched blocks (p) = 9 v ~ ~25 Strings Serviced = 19 Round 1 Gradual prefetching

v = 17 p = 12 v ~ ~25 Strings = 26 Round 2 Gradual Prefetching

v = 16 p = 14 v ~ ~25 Strings = 28 Round 3 Gradual Prefetching

v = 16 p = 14 v ~16+8 ~25 Strings = 30 Round 4 Gradual Prefetching

Grouped Periodic Multi-ground Prefetching: This algorithm temporarily stores prefetched blocks in the hosts cache. epoch: epoch or virtual round is the total duration of fixed number of actual rounds. At the system level, GPMP offers finer grained disk requests per stream, more flexible configuration and higher service quality, because of lower start-up latencies.

During GPMP round, media server serves all streams. Delivers all blocks sustaining playback of the supported streams to host. These blocks dont have to be retrieved from disk in each round.

During a round, group containing fraction of the supported stream is randomly retrieved. Time remains in round for u prefetches for each stream.(u=num. of blocks prefetched for one stream). In next round, blocks sustaining playback of next group are read from disk with u prefetched blocks for each stream.

All streams are served in each round as N-v rounds that were not retrieved from disk are in cache having been prefetched in previous rounds. The u prefetched blocks sustain for u rounds. After which same streams read from disk in given round are read again.

Epoch length is k = u+1 Total number of supported streams are N=v(u+1) Played blocks are immediately discarded.

Performance evaluation To analytically evaluate the performance of the algorithms the continuous media server is simulated for a pregenerated workload similar to that of a typical server use. DiskSim is used. It does not address onboard cache issues like ability to enable disable caching, the ability to switch prefetching on and off and ability to change disk cache segmentation. DiskSim code is modified to add these capabilities.

The Experiment Use of video library that holds videos that are 90 to 120 min. long which are ordered according to popularity and follow Zipfian distribution. Newly videos are with a Poisson distribution. Disk partitioning is assumed. S&P algorithms are implemented using trace generator to pass hints and evaluate algorithms.

Results

Testing Sweep and S&P for throughput for: Disk Cache sizes (d): 2, 4, 8, 12 Mbytes Rounds length (r): 0.5, 1, 1.5, 2, 3, 4 sec. 20 to 70 % improvements in throughput for r= 0.5 to 1.5 sec and d=2 to 12 Mbytes. Longer r explode memory requirements. GPMP outperforms Sweep getting higher throughput but poor start-up latency.

In this figure the evaluation of a single disk media server is shown. r= 0.25 to 1 sec. Low request arrival rate=> low value of k=> lower start-up latencies. As request arrival rate increases, queuing delays occur and GPMP configurations become more beneficial for higher throughput.

Conclusion The techniques presented introduce higher(60 to 70%) throughput compared to Sweep strategies for continuous blocks. Sweep does not exploit on board buffers. There are no prefetching techniques by disk manufacturers for media retrievals that account for concept of rounds. Parallel transfer of I/O requests and other disk to buffer transfers.

Expected future use: The current technology trends suggest that these techniques will show even better results for future disk products because transfer rates will improve and more powerful controllers on bigger embedded caches are certain to follow.

References: Paper on Hierarchical Caching and Prefetching for Continuous Media Servers with Smart disks -Stavros Harizopoulos (Carnegie Mellon University) Costas Harizakis and Peter Triantafillou (Technical University of Crete)