An Architecture for Distributed High Performance Video Processing in the Cloud 作者 :Pereira, R.; Azambuja, M.; Breitman, K.; Endler, M. 出處 :2010 IEEE 3rd.

Slides:



Advertisements
Similar presentations
LIBRA: Lightweight Data Skew Mitigation in MapReduce
Advertisements

Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
SLA-Oriented Resource Provisioning for Cloud Computing
Building Cloud-ready Video Transcoding System for Content Delivery Networks(CDNs) Zhenyun Zhuang and Chun Guo Speaker: 饒展榕.
EHarmony in Cloud Subtitle Brian Ko. eHarmony Online subscription-based matchmaking service Available in United States, Canada, Australia and United Kingdom.
Esma Yildirim Department of Computer Engineering Fatih University Istanbul, Turkey DATACLOUD 2013.
C-Store: Data Management in the Cloud Jianlin Feng School of Software SUN YAT-SEN UNIVERSITY Jun 5, 2009.
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
Copyright 2009 FUJITSU TECHNOLOGY SOLUTIONS PRIMERGY Servers and Windows Server® 2008 R2 Benefit from an efficient, high performance and flexible platform.
Authors: Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox Publish: HPDC'10, June 20–25, 2010, Chicago, Illinois, USA ACM Speaker: Jia Bao Lin.
Locality-Aware Request Distribution in Cluster-based Network Servers 1. Introduction and Motivation --- Why have this idea? 2. Strategies --- How to implement?
MapReduce in the Clouds for Science CloudCom 2010 Nov 30 – Dec 3, 2010 Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox {tgunarat, taklwu,
SPRING 2011 CLOUD COMPUTING Cloud Computing San José State University Computer Architecture (CS 147) Professor Sin-Min Lee Presentation by Vladimir Serdyukov.
Final Project: Video Transcoding on Cloud Environments Queenie Wong CMPT 880.
Copyright © 2012 Cleversafe, Inc. All rights reserved. 1 Combining the Power of Hadoop with Object-Based Dispersed Storage.
Load Test Planning Especially with HP LoadRunner >>>>>>>>>>>>>>>>>>>>>>
U.S. Department of the Interior U.S. Geological Survey David V. Hill, Information Dynamics, Contractor to USGS/EROS 12/08/2011 Satellite Image Processing.
SYSTEM ANALYSIS AND DESIGN
A User Experience-based Cloud Service Redeployment Mechanism KANG Yu.
Ch 4. The Evolution of Analytic Scalability
The Pursuit for Efficient S/C Design The Stanford Small Sat Challenge: –Learn system engineering processes –Design, build, test, and fly a CubeSat project.
Applying Twister to Scientific Applications CloudCom 2010 Indianapolis, Indiana, USA Nov 30 – Dec 3, 2010.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Cloud MapReduce : a MapReduce Implementation on top of a Cloud Operating System Speaker : 童耀民 MA1G Authors: Huan Liu, Dan Orban Accenture.
資訊工程系智慧型系統實驗室 iLab 南台科技大學 1 Optimizing Cloud MapReduce for Processing Stream Data using Pipelining 出處 : 2011 UKSim 5th European Symposium on Computer Modeling.
An Architecture for Video Surveillance Service based on P2P and Cloud Computing Yu-Sheng Wu, Yue-Shan Chang, Tong-Ying Juang, Jing-Shyang Yen speaker:
TEMPORAL DATA AND REAL- TIME ALGORITHMS CHAPTER 4 – GROUP 3.
Cloud Computing 1. Outline  Introduction  Evolution  Cloud architecture  Map reduce operation  Platform 2.
Authors: Jiann-Liang Chenz, Szu-Lin Wuy,Yang-Fang Li, Pei-Jia Yang,Yanuarius Teofilus Larosa th International Wireless Communications and Mobile.
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
天文信息技术联合实验室 New Progress On Astronomical Cross-Match Research Zhao Qing.
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
Cloud Distributed Computing Platform 2 Content of this lecture is primarily from the book “Hadoop, The Definite Guide 2/e)
An Autonomic Framework in Cloud Environment Jiedan Zhu Advisor: Prof. Gagan Agrawal.
Extreme-scale computing systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward exa-scale computing.
Optimizing Cloud MapReduce for Processing Stream Data using Pipelining 作者 :Rutvik Karve , Devendra Dahiphale , Amit Chhajer 報告 : 饒展榕.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Graduate Student Department Of CSE 1.
March 16 & 21, Csci 2111: Data and File Structures Week 9, Lectures 1 & 2 Indexed Sequential File Access and Prefix B+ Trees.
An Architecture for Distributed High Performance Video Processing in the Cloud Speaker : 吳靖緯 MA0G IEEE 3rd International Conference.
MARISSA: MApReduce Implementation for Streaming Science Applications 作者 : Fadika, Z. ; Hartog, J. ; Govindaraju, M. ; Ramakrishnan, L. ; Gunter, D. ; Canon,
Papers on Storage Systems 1) Purlieus: Locality-aware Resource Allocation for MapReduce in a Cloud, SC ) Making Cloud Intermediate Data Fault-Tolerant,
Eneryg Efficiency for MapReduce Workloads: An Indepth Study Boliang Feng Renmin University of China Dec 19.
Performance Evaluation of Image Conversion Module Based on MapReduce for Transcoding and Transmoding in SMCCSE Speaker : 吳靖緯 MA0G IEEE.
Optimizing Cloud MapReduce for Processing Stream Data using Pipelining 2011 UKSim 5th European Symposium on Computer Modeling and Simulation Speker : Hong-Ji.
The Alternative Larry Moore. 5 Nodes and Variant Input File Sizes Hadoop Alternative.
Computing Scientometrics in Large-Scale Academic Search Engines with MapReduce Leonidas Akritidis Panayiotis Bozanis Department of Computer & Communication.
A P2P-Based Architecture for Secure Software Delivery Using Volunteer Assistance Purvi Shah, Jehan-François Pâris, Jeffrey Morgan and John Schettino IEEE.
Mining Document Collections to Facilitate Accurate Approximate Entity Matching Presented By Harshda Vabale.
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
CS525: Big Data Analytics MapReduce Computing Paradigm & Apache Hadoop Open Source Fall 2013 Elke A. Rundensteiner 1.
Big traffic data processing framework for intelligent monitoring and recording systems 學生 : 賴弘偉 教授 : 許毅然 作者 : Yingjie Xia a, JinlongChen a,b,n, XindaiLu.
Massive Semantic Web data compression with MapReduce Jacopo Urbani, Jason Maassen, Henri Bal Vrije Universiteit, Amsterdam HPDC ( High Performance Distributed.
A N I N - MEMORY F RAMEWORK FOR E XTENDED M AP R EDUCE 2011 Third IEEE International Conference on Coud Computing Technology and Science.
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 4: Index Construction Related to Chapter 4:
Page 1 A Platform for Scalable One-pass Analytics using MapReduce Boduo Li, E. Mazur, Y. Diao, A. McGregor, P. Shenoy SIGMOD 2011 IDS Fall Seminar 2011.
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-1.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
INTRODUCTION TO GRID & CLOUD COMPUTING U. Jhashuva 1 Asst. Professor Dept. of CSE.
Big Data is a Big Deal!.
MapReduce Compiler RHadoop
Sushant Ahuja, Cassio Cristovao, Sameep Mohta
Distributed Cache Technology in Cloud Computing and its Application in the GIS Software Wang Qi Zhu Yitong Peng Cheng
Introduction to Distributed Platforms
Big Data Processing in Cloud Computing Environments
Ch 4. The Evolution of Analytic Scalability
Database System Architectures
MapReduce: Simplified Data Processing on Large Clusters
Presentation transcript:

An Architecture for Distributed High Performance Video Processing in the Cloud 作者 :Pereira, R.; Azambuja, M.; Breitman, K.; Endler, M. 出處 :2010 IEEE 3rd International Conference on Cloud Computing 報告者 : 黃群凱

Outline  INTRODUCTION  BACKGROUND AND PROBLEM STATEMENT  DISTRIBUTED VIDEO COMPRESSION  DISCUSSION  CONCLUSION

INTRODUCTION(1/2)  As computer systems evolve, the volume of data to be processed increases significantly, either as a consequence of the expanding amount of information available, or due to the possibility to perform highly complex operations that were not feasible in the past.  In this scenario, the efficient use of available computational resources is key, and creates a demand for systems that can optimize the use of resources in relation tothe amount of data to be processed.

INTRODUCTION(2/2)  In these cases, there are moments when there is very low demand and the resources are almost idle while at other moments, there is processing demand that exceeds the resource capacity, and which may cause undesirable delays.  In this light, the ability to build adaptive systems, capable of using on demand resources provided by Cloud Computing [16][17][18], is very interesting.

BACKGROUND AND PROBLEM STATEMENT(1/5)  The Map-Reduce paradigm [2], for example, is a framework for processing huge datasets of certain kinds of distributable problems using a large number of computers(nodes), collectively referred to as a cluster.

BACKGROUND AND PROBLEM STATEMENT(2/5)  It consists of an initial Map stage, where a master node takes the input, chops it into smaller or sub-problems, and distributes the parts to worker nodes, which process the information.  Following there is the Reduce stage, where the master node collects the answers to all the sub-problems and combines them to produce the job output.

BACKGROUND AND PROBLEM STATEMENT(3/5)

BACKGROUND AND PROBLEM STATEMENT(4/5)  In cases where there is a seasonal computation demand. the use of public Clouds, for information processing and storage, is emerging as an interesting alternative  With a public Cloud, one can quickly make provision for the resources required to perform a particular task, and pay only for the computational resources effectively used.

BACKGROUND AND PROBLEM STATEMENT(5/5)  A good example where Map-Reduce could be generalized, is the compression of high definition video files, which requires intensive information processing.  In this compression process, streams of audio and video are processed with different algorithms, and there is a great correlation between subsequent video frames, especially when there is temporal compression.

DISTRIBUTED VIDEO COMPRESSION(1/10)  Video applications require some form of data compression to facilitate storage and transmission. Digital video compression is one of the main issues in digital video encoding, enabling efficient distribution and interchange of visual information.

DISTRIBUTED VIDEO COMPRESSION(2/10)

DISTRIBUTED VIDEO COMPRESSION(3/10)  In order to speed up encoding times, there are basically two solutions. The first one is to augment the investment in encoding hardware infrastructure, to be used in full capacity only at peak times. The second solution is to try and optimize the use of available resources. The ideal scenario is to optimize resources by distributing the tasks among them evenly.  The synchronization between audio and video can be greatly affected, since the frame size of each one may not be equal.

DISTRIBUTED VIDEO COMPRESSION(4/10)  The Split Step

DISTRIBUTED VIDEO COMPRESSION(5/10)  The Process Step Once video is fragmented, the chunks generated should be distributed among the nodes to be processed. In the specific case of video compression, this process aims at reducing the size of the video file by eliminating redundancies. There are several open source tools for video compression, among the most popular, ffmpeg[12][19] and mencoder[13], which are compatible with various implementations of audio and video codecs.

DISTRIBUTED VIDEO COMPRESSION(6/10)  The Process Step  At the end of the processing step, we have all the compressed chunks, as well as the audio stream. To obtain the desired output, we must merge all fragments, thus reconstructing the original content.

DISTRIBUTED VIDEO COMPRESSION(7/10)  The Merge Step The first phase of the merge step is to join the chunks of processed video, which can be accomplished easily by ordering the fragments and rewriting the container. After the split, process and merge steps, implemented using the proposed architecture, we created a fully parallel and distributed video compression process, where the different pieces of content can be processed simultaneously in a cluster or, alternatively, using resources in the Cloud.

DISTRIBUTED VIDEO COMPRESSION(8/10)  Performance Tests To deploy the infrastructure, we chose the instance type m1.small for all instances of servers, which has the following characteristics, according to Amazon:  1.7 GB memory  1 EC2 Compute Unit (1 virtual core with 1 EC2Compute Unit)  160 GB instance storage (150 GB plus 10 GB rootpartition)  I/O Performance: Moderate

DISTRIBUTED VIDEO COMPRESSION(9/10)  Performance Tests

DISTRIBUTED VIDEO COMPRESSION(10/10)  Performance Tests

DISTRIBUTED VIDEO COMPRESSION(1/6)  Scalable and fault tolerant architectures that support parallel processing of large volumes of data in Cloud environments, are becoming increasingly necessary to enable flexible, robust and efficient processing of large volumes of data.  If we analyze Map-Reduce in its essence, we note that process optimization is achieved by distributing tasks among available computing resources.

DISTRIBUTED VIDEO COMPRESSION(2/6)  Fault Tolerance The typical Map-Reduce implementation provides a single master node, responsible for the scheduling tasks to worker nodes (responsible for doing the processing). The Split&Merge architecture tackles this problem by coupling a service to the master node, that periodically checks the conditions of its workers. This ensures that the master node, which controls the entire distribution of tasks, is always able to identify whether a node is healthy or not.

DISTRIBUTED VIDEO COMPRESSION(3/6)  Private Cluster Deployment One of the main features of a private cluster is the control over the environment and the ability to customize it. It is possible to have an accurate idea of the computational power of each component, and identify existing bottlenecks, which greatly facilitates the processes of deployment, management and maintenance of existing resources.

DISTRIBUTED VIDEO COMPRESSION(4/6)  Private Cluster Deployment

DISTRIBUTED VIDEO COMPRESSION(5/6)  Public Cloud Deployment In cases where there is a floating demand or services, or when sudden changes to the environment dictate the need for additional resources, the use of public Cloud Computing platforms to launch applications developed using the Split&Merge architecture becomes extremely interesting. If there is no demand, all workers can be turned off, on the fly, by the master node. Cost of operation is thus minimal. Even masters may be disconnected, and re-connected, manually, which makes the total cost of operation in idle situations very low.

DISTRIBUTED VIDEO COMPRESSION(6/6)  C. Public Cloud Deployment

CONCLUSION(1/2)  The emergence and evolution of services in the Cloud, which allows computing resources to be used on demand, increases flexibility and makes the processing of large datasets more accessible and scalable,especially when we have seasonal demands.  A specific case that was presented was video compression, and, for this type of processing, we propose amore efficient architecture and that can be deployed both inprivate clusters, as in the Cloud

CONCLUSION(2/2)  Future work Another issue we would like to investigate is the incorporation of autonomic computing mechanisms to help anticipate and identify faults,and the implementation of efficient prevention and recovery mechanisms. That will contribute to making the present solution more robust and scalable.