The IEEE International Conference on Cluster Computing 2010

Slides:



Advertisements
Similar presentations
Introduction to cloud computing Jiaheng Lu Department of Computer Science Renmin University of China
Advertisements

Efficient Event-based Resource Discovery Wei Yan*, Songlin Hu*, Vinod Muthusamy +, Hans-Arno Jacobsen +, Li Zha* * Chinese Academy of Sciences, Beijing.
MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
2P13 Week 11. A+ Guide to Managing and Maintaining your PC, 6e2 RAID Controllers Redundant Array of Independent (or Inexpensive) Disks Level 0 -- Striped.
Availability in Globally Distributed Storage Systems
Ceph: A Scalable, High-Performance Distributed File System Priya Bhat, Yonggang Liu, Jing Qin.
 Need for a new processing platform (BigData)  Origin of Hadoop  What is Hadoop & what it is not ?  Hadoop architecture  Hadoop components (Common/HDFS/MapReduce)
End-to-End Analysis of Distributed Video-on-Demand Systems Padmavathi Mundur, Robert Simon, and Arun K. Sood IEEE Transactions on Multimedia, February.
Efficient Content Location Using Interest-based Locality in Peer-to-Peer Systems Presented by: Lin Wing Kai.
Performance Evaluation of Load Sharing Policies on a Beowulf Cluster James Nichols Marc Lemaire Advisor: Mark Claypool.
Performance Evaluation of Peer-to-Peer Video Streaming Systems Wilson, W.F. Poon The Chinese University of Hong Kong.
Inexpensive Scalable Information Access Many Internet applications need to access data for millions of concurrent users Relational DBMS technology cannot.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google∗
The Hadoop Distributed File System: Architecture and Design by Dhruba Borthakur Presented by Bryant Yao.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
1 The Google File System Reporter: You-Wei Zhang.
An adaptive framework of multiple schemes for event and query distribution in wireless sensor networks Vincent Tam, Keng-Teck Ma, and King-Shan Lui IEEE.
An Architecture for Video Surveillance Service based on P2P and Cloud Computing Yu-Sheng Wu, Yue-Shan Chang, Tong-Ying Juang, Jing-Shyang Yen speaker:
Managing Service Metadata as Context The 2005 Istanbul International Computational Science & Engineering Conference (ICCSE2005) Mehmet S. Aktas
Presenter: Dipesh Gautam.  Introduction  Why Data Grid?  High Level View  Design Considerations  Data Grid Services  Topology  Grids and Cloud.
Location-aware MapReduce in Virtual Cloud 2011 IEEE computer society International Conference on Parallel Processing Yifeng Geng1,2, Shimin Chen3, YongWei.
W HAT IS H ADOOP ? Hadoop is an open-source software framework for storing and processing big data in a distributed fashion on large clusters of commodity.
Introduction to Apache Hadoop Zibo Wang. Introduction  What is Apache Hadoop?  Apache Hadoop is a software framework which provides open source libraries.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
BFTCloud: A Byzantine Fault Tolerance Framework for Voluntary-Resource Cloud Computing Yilei Zhang, Zibin Zheng, and Michael R. Lyu
Distributed File System By Manshu Zhang. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
The exponential growth of data –Challenges for Google,Yahoo,Amazon & Microsoft in web search and indexing The volume of data being made publicly available.
CEPH: A SCALABLE, HIGH-PERFORMANCE DISTRIBUTED FILE SYSTEM S. A. Weil, S. A. Brandt, E. L. Miller D. D. E. Long, C. Maltzahn U. C. Santa Cruz OSDI 2006.
On Optimizing the Backoff Interval for Random Access Scheme Zygmunt J. Hass and Jing Deng IEEE Transactions on Communications, Dec 2003.
Performance Evaluation of Image Conversion Module Based on MapReduce for Transcoding and Transmoding in SMCCSE Speaker : 吳靖緯 MA0G IEEE.
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
GreenSched: An Energy-Aware Hadoop Workflow Scheduler
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
CARDIO: Cost-Aware Replication for Data-Intensive workflOws Presented by Chen He.
Zibin Zheng DR 2 : Dynamic Request Routing for Tolerating Latency Variability in Cloud Applications CLOUD 2013 Jieming Zhu, Zibin.
Eduardo Gutarra Velez. Outline Distributed Filesystems Motivation Google Filesystem Architecture The Metadata Consistency Model File Mutation.
Ceph: A Scalable, High-Performance Distributed File System
PROP: A Scalable and Reliable P2P Assisted Proxy Streaming System Computer Science Department College of William and Mary Lei Guo, Songqing Chen, and Xiaodong.
CS Spring 2009 CS 414 – Multimedia Systems Design Lecture 30 – Media Server (Part 5) Klara Nahrstedt Spring 2009.
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
HADOOP DISTRIBUTED FILE SYSTEM HDFS Reliability Based on “The Hadoop Distributed File System” K. Shvachko et al., MSST 2010 Michael Tsitrin 26/05/13.
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
Data Consolidation: A Task Scheduling and Data Migration Technique for Grid Networks Author: P. Kokkinos, K. Christodoulopoulos, A. Kretsis, and E. Varvarigos.
Experiments in Utility Computing: Hadoop and Condor Sameer Paranjpye Y! Web Search.
{ Tanya Chaturvedi MBA(ISM) Hadoop is a software framework for distributed processing of large datasets across large clusters of computers.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Cloud Distributed Computing Environment Hadoop. Hadoop is an open-source software system that provides a distributed computing environment on cloud (data.
Distributed File System. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
INTRODUCTION TO HADOOP. OUTLINE  What is Hadoop  The core of Hadoop  Structure of Hadoop Distributed File System  Structure of MapReduce Framework.
Architecture for Resource Allocation Services Supporting Interactive Remote Desktop Sessions in Utility Grids Vanish Talwar, HP Labs Bikash Agarwalla,
Seminar On Rain Technology
1 Student Date Time Wei Li Nov 30, 2015 Monday 9:00-9:25am Shubbhi Taneja Nov 30, 2015 Monday9:25-9:50am Rodrigo Sanandan Dec 2, 2015 Wednesday9:00-9:25am.
An Introduction to GPFS
Presenter: Yue Zhu, Linghan Zhang A Novel Approach to Improving the Efficiency of Storing and Accessing Small Files on Hadoop: a Case Study by PowerPoint.
Introduction to Distributed Platforms
Advanced Topics in Concurrency and Reactive Programming: Case Study – Google Cluster Majeed Kassis.
CSS534: Parallel Programming in Grid and Cloud
Gregory Kesden, CSE-291 (Storage Systems) Fall 2017
Gregory Kesden, CSE-291 (Cloud Computing) Fall 2016
Auburn University COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques (2) Dr. Xiao Qin Auburn University.
Unistore: Project Updates
The Basics of Apache Hadoop
Outline Midterm results summary Distributed file systems – continued
Hadoop Technopoints.
CLUSTER COMPUTING.
Harrison Howell CSCE 824 Dr. Farkas
Presentation transcript:

The IEEE International Conference on Cluster Computing 2010 CDRM: A Cost-effective Dynamic Replication Management Scheme for Cloud Storage Cluster Qingsong Wei Data Storage Institute, A-STAR, Singapore Bharadwaj Veeravalli, Bozhao Gong National University of Singapore Lingfang Zeng, Dan Feng Huazhong University of Science & Technology, China

Cost-effective Dynamic Replication Management (CDRM) Agenda Outline Introduction Problem Statement Cost-effective Dynamic Replication Management (CDRM) Evaluation Conclusion

1, Introduction Outline HDFS Architecture Clients Name Node Network Meta Data Name Node Network Control Node Disk Disks Node Disk Disks Node Disk Disks Node Disk Disks Node Disk Disks Node Disk Disks Data Blocks Data Nodes

1. Introduction In the HDFS, files are striped into date blocks across multiple data nodes to enable parallel access. B1 B2 … Bm Data Striping However, Block may be unaccessible due to date node unavailable. If one of the blocks is unavailable, so as the whole file. Failure is normal instead of exception in large scale storage cloud system. Fault tolerance is required in such a system. … Node1 Node2 Noden Page 4 of 19

1. Introduction Data nodes Clients 1 1 2 5 2 2 4 3 4 5 4 3 5 Replication is used in HDFS. When one data node fails, the data is still accessible from the replicas and storage service need not be interrupted. Besides fault tolerance, replicas among data nodes can be used to balance workload. Page 5 of 19 5

2. Problem Statement Current replication managements Treat all data as same: same replica number for all data Treat all storage nodes as same Fixed and Static 1 2 3 4 5 5 1 2 3 4 4 5 1 2 3 High cost & Poor load balance Page 6 of 19 6

2. Problem Statement Replica number is critical to management cost. More replica, more cost. 1 2 1 5 2 2 4 The block 5 is modified 3 5 4 4 3 5 Update to maintain consistency Because large number of blocks are stored in system, even a small increase of replica number can result in a significant increase of management cost in the overall system. Then, how many minimal replica should be kept in the system to satisfy availability requirement? Page 7 of 19 7

2. Problem Statement Replica placement influences intra-request parallelism. Client File (B1, B2, B3) B3 B2 B1 Requests Blocked If B1 has a replica in Datanode 2 which has free sessions, the access can be immediately serviced. B3 B2 B1 B1 Data Node1 Data Node2 Data Node3 Sessionmax=3 Sessionfree=1 Sessionmax=3 Sessionfree=2 Sessionmax=2 Sessionfree=0 Page 8 of 19

2. Problem Statement Replica placement also influences inter-request parallelism. Client1 Client2 B3 B2 B1 B1 Requests How to place these replicas among Data nodes clusters in a balance way to improve access parallelism? B1 B2 B1 B3 Data Node1 Data Node2 Data Node3 Sessionmax=3 Sessionfree=0 Sessionmax=3 Sessionfree=1 Sessionmax=2 Sessionfree=0 Page 9 of 19

3. Cost-effective Dynamic Replication Management System Model pj : popularity sj : size rj : replica number tj : access latency requirement (p1, s1, r1, t1) (pj, sj, rj, tj) (pM, sM, rM, tM) …… …… B1 Bj BM Total arrival rate: λ λi : req. arr. rate τi : average ser. time fi : failure rate ci : max sessions Node1 … Nodei … NodeN Two key issues to be addressed 1, Minimal replica number 2, How to place these replicas in balanced way (λ1, τ1, f1, c1) (λi, τi, fi, ci) (λN, τN, fN, cN) Data has different attributes Data nodes are different Page 10 of 19

3. Cost-effective Dynamic Replication Management Availability Suppose file F is striped into m blocks {b1 , b2 ,…, bm}. To retrieve whole file F, we must get all the m blocks. Availability is modeled as function of replica number. Suppose the expected availability for file F is Aexpect, which defined by users. To satisfy the availability requirement for a given file, we get Minimum replicas can be calculated from above Eq. for a given expected availability. Page 11 of 19

3. Cost-effective Dynamic Replication Management Blocking Probability Blocking probability is used as criterion to place replicas among data nodes to improve load balance . An data node Si is modeled as M/G/ci system with arrival rate λi and service time τi, and accordingly, the blocking probability of data node Si can be given to be Replica placement policy: replica will be placed into data node with lowest blocking probability to dynamically maintain overall load balancing. After determining how many replicas the system should maintain at least to satisfy availability requirement, we shall explain how to place these replicas efficiently to maximize performance and load balancing. Blocking probability describe the probability of data node to reject a request. Page 12 of 19

Framework of cost-effective dynamic replication management in HDFS Request to create a file with <Availability, Block Number> 1 Client Name Node Bm … B2 B1 Return replication policy <Bi, Replication factor, DataNode list> 3 4 Flush and replicate blocks to selected Datanodes Calculate the replication factor and Search the Datanode B+Tree to obtain Datanode list. 2 Data Nodes Besides default replication management, HDFS provides flexible API for us to extend and implement efficient replication management. Replication Pipelining Framework of cost-effective dynamic replication management in HDFS Page 13 of 19

4. Evaluation Setup Our test platform is built on a cluster with one name node and twenty data nodes of commodity computer The operating system is Red Hat AS4.4 with kernel 2.6.20. Hadoop version is 0.16.1 and java version is 1.6.0. AUSPEX file system trace is used A synthesizer is developed to create workloads with different characteristics, such as data sets of different sizes, varying data rates, and different popularities. These characteristics reflect the differences among various workloads to the cloud storage cluster. Page 14 of 19

4. Evaluation Cost effective Availability Initially, one replica per object. CDRM only maintain minimal replicas to satisfy availability. Higher failure rate, more replica required. Dynamic replication with Data node failure rate of 0.1 and 0.2 , Aexpect=0.8 Page 15 of 19

Effect of popularity and access arrival rate, 20 data nodes 4. Evaluation Performance CDRM vs. HDFS default Replication Management (HDRM) under different popularity and workload intensity. Performance of CDRM is much better than that of HDRM when popularity is small. CDRM outperform HDRM under heavy workload. popularity is a ratio between the size of the most popular data and the size of the total data set. A large popularity indicates the data are more uniformly accessed. A small popularity means accesses concentrate in a small amount of data. Effect of popularity and access arrival rate, 20 data nodes Page 16 of 19

System utilization among data nodes, popularity=10%, λ=0.6 4. Evaluation Load Balance The figure shows the difference of system utilization of each data node comparing to the average system utilization of the cluster. CDRM can dynamically distribute workload among whole cluster. System utilization among data nodes, popularity=10%, λ=0.6 Page 17 of 19

Current replication management policies 5. Conclusion Current replication management policies CDRM Data is same Data is different Storage node are same Storage nodes are different Same replica number for all data Different replica number for different data Static placement Dynamic placement High Cost Cost effective Poor load balance Good balance Low performance High performance Page 18 of 19

Thanks & Question For more questions, please contact Dr. Qingsong Wei by email: WEI_Qingsong@dsi.a-star.edu.sg Page 19 of 19