VIPIN VIJAYAN 11/11/03 A Performance Analysis of Two Distributed Computing Abstractions.

Slides:



Advertisements
Similar presentations
MINJAE HWANG THAWAN KOOBURAT CS758 CLASS PROJECT FALL 2009 Extending Task-based Programming Model beyond Shared-memory Systems.
Advertisements

A NOVEL APPROACH TO SOLVING LARGE-SCALE LINEAR SYSTEMS Ken Habgood, Itamar Arel Department of Electrical Engineering & Computer Science GABRIEL CRAMER.
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
SLA-Oriented Resource Provisioning for Cloud Computing
CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
Locality-Aware Dynamic VM Reconfiguration on MapReduce Clouds Jongse Park, Daewoo Lee, Bokyeong Kim, Jaehyuk Huh, Seungryoul Maeng.
1/1/ /e/e eindhoven university of technology Microprocessor Design Course 5Z008 Dr.ir. A.C. (Ad) Verschueren Eindhoven University of Technology Section.
LOAD BALANCING IN A CENTRALIZED DISTRIBUTED SYSTEM BY ANILA JAGANNATHAM ELENA HARRIS.
Efficient Autoscaling in the Cloud using Predictive Models for Workload Forecasting Roy, N., A. Dubey, and A. Gokhale 4th IEEE International Conference.
Lincoln University Canterbury New Zealand Evaluating the Parallel Performance of a Heterogeneous System Elizabeth Post Hendrik Goosen formerly of Department.
Computer Organization and Architecture 18 th March, 2008.
CSCE 212 Chapter 4: Assessing and Understanding Performance Instructor: Jason D. Bakos.
Performance Evaluation of Load Sharing Policies on a Beowulf Cluster James Nichols Marc Lemaire Advisor: Mark Claypool.
An Introduction to Grid Computing Research at Notre Dame Prof. Douglas Thain University of Notre Dame
1 Introduction to Load Balancing: l Definition of Distributed systems. Collection of independent loosely coupled computing resources. l Load Balancing.
Parallel Computation in Biological Sequence Analysis Xue Wu CMSC 838 Presentation.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
Advanced Topics: MapReduce ECE 454 Computer Systems Programming Topics: Reductions Implemented in Distributed Frameworks Distributed Key-Value Stores Hadoop.
PMIT-6102 Advanced Database Systems
Cloud MapReduce : a MapReduce Implementation on top of a Cloud Operating System Speaker : 童耀民 MA1G Authors: Huan Liu, Dan Orban Accenture.
High Throughput Computing with Condor at Purdue XSEDE ECSS Monthly Symposium Condor.
Predictive Runtime Code Scheduling for Heterogeneous Architectures 1.
Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Technology Education Lecture 5 Operating Systems.
 What is an operating system? What is an operating system?  Where does the OS fit in? Where does the OS fit in?  Services provided by an OS Services.
Performance Evaluation of Parallel Processing. Why Performance?
Performance Tuning on Multicore Systems for Feature Matching within Image Collections Xiaoxin Tang*, Steven Mills, David Eyers, Zhiyi Huang, Kai-Cheung.
Grid Computing I CONDOR.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
Computer Science and Engineering Parallel and Distributed Processing CSE 8380 March 01, 2005 Session 14.
임규찬. 1. Abstract 2. Introduction 3. Design Goals 4. Sample-Based Scheduling for Parallel Jobs 5. Implements.
CDA 3101 Fall 2013 Introduction to Computer Organization Computer Performance 28 August 2013.
Carrying Your Environment With You or Virtual Machine Migration Abstraction for Research Computing.
1 Multiprocessor and Real-Time Scheduling Chapter 10 Real-Time scheduling will be covered in SYSC3303.
StreamX10: A Stream Programming Framework on X10 Haitao Wei School of Computer Science at Huazhong University of Sci&Tech.
The Owner Share scheduler for a distributed system 2009 International Conference on Parallel Processing Workshops Reporter: 李長霖.
Distributed Framework for Automatic Facial Mark Detection Graduate Operating Systems-CSE60641 Nisha Srinivas and Tao Xu Department of Computer Science.
April 26, CSE8380 Parallel and Distributed Processing Presentation Hong Yue Department of Computer Science & Engineering Southern Methodist University.
Frontiers in Massive Data Analysis Chapter 3.  Difficult to include data from multiple sources  Each organization develops a unique way of representing.
BOF: Megajobs Gracie: Grid Resource Virtualization and Customization Infrastructure How to execute hundreds of thousands tasks concurrently on distributed.
Page 1 Process Migration & Allocation Paul Krzyzanowski Distributed Systems Except as otherwise noted, the content of this.
Review of Condor,SGE,LSF,PBS
Integrating Computing Resources on Multiple Grid-enabled Job Scheduling Systems Through a Grid RPC System Yoshihiro Nakajima, Mitsuhisa Sato, Yoshiaki.
Motivation: Sorting is among the fundamental problems of computer science. Sorting of different datasets is present in most applications, ranging from.
GEM: A Framework for Developing Shared- Memory Parallel GEnomic Applications on Memory Constrained Architectures Mucahid Kutlu Gagan Agrawal Department.
Design Issues of Prefetching Strategies for Heterogeneous Software DSM Author :Ssu-Hsuan Lu, Chien-Lung Chou, Kuang-Jui Wang, Hsiao-Hsi Wang, and Kuan-Ching.
Big traffic data processing framework for intelligent monitoring and recording systems 學生 : 賴弘偉 教授 : 許毅然 作者 : Yingjie Xia a, JinlongChen a,b,n, XindaiLu.
Advanced Computer Networks Lecture 1 - Parallelization 1.
A N I N - MEMORY F RAMEWORK FOR E XTENDED M AP R EDUCE 2011 Third IEEE International Conference on Coud Computing Technology and Science.
A System Performance Model Distributed Process Scheduling.
Futures, Scheduling, and Work Distribution Speaker: Eliran Shmila Based on chapter 16 from the book “The art of multiprocessor programming” by Maurice.
Chapter 8 System Management Semester 2. Objectives  Evaluating an operating system  Cooperation among components  The role of memory, processor,
SEC 2: EVALUATING ABSOLUTE VALUES & SIMPLIFYING EXPRESSIONS Learning Target: 1.I will simplify complex algebraic expressions by combining like terms, adding.
Christopher Brown and Kevin Hammond School of Computer Science, University of St. Andrews July 2010 Ever Decreasing Circles: Implementing a skeleton for.
2004 Queue Scheduling and Advance Reservations with COSY Junwei Cao Falk Zimmermann C&C Research Laboratories NEC Europe Ltd.
COMP7330/7336 Advanced Parallel and Distributed Computing MapReduce - Introduction Dr. Xiao Qin Auburn University
Measuring Performance II and Logic Design
Fundamental Operations Scalability and Speedup
Web Server Load Balancing/Scheduling
lecture 5: CPU Scheduling
Web Server Load Balancing/Scheduling
Condor – A Hunter of Idle Workstation
Scaling Up Scientific Workflows with Makeflow
Where are being used the OS?
Introduction to Parallelism.
PA an Coordinated Memory Caching for Parallel Jobs
CSCE 212 Chapter 4: Assessing and Understanding Performance
MapReduce Simplied Data Processing on Large Clusters
Chapter 6: CPU Scheduling
Basic Grid Projects – Condor (Part I)
CPU SCHEDULING.
Presentation transcript:

VIPIN VIJAYAN 11/11/03 A Performance Analysis of Two Distributed Computing Abstractions

Outline Introduction Problem AllPairs Work Queue & Condor Evaluation Preliminary Results

Introduction In a biometric matching experiment, a query image is compared against a target image to compute a match score. F is the matching function. M ij = F(A i,B j ) BiBi AiAi F(A 1,B 1 ) Similarity matrix, M

Problem Problems encountered while computing the similarity matrix: O(n 2 ) increase in computational time where n is the number of images. Increase in computational time due to the size of the images and complexity of the matching function. Proposed solution:  a distributed computing abstraction such as AllPairs.

AllPairs: a distributed computing abstraction As an abstraction:  It simplifies the calculation of a similarity matrix:  We simply have to provide the matching function, and the two sets of images. Because of the abstraction:  We can take advantage of the particular abstraction to speed up computation.  Ex: Taking advantage of sub-matrices.

AllPairs: a second look Worker 1 (Busy) Worker 2 (Busy) Worker 3 (Busy) Worker 4 (Busy) Waiting Active Here we upload 4+3 = 7 files to a single worker. As opposed to uploading 4*3 = 12 files to different workers. So the main advantage of splitting it up into sub-matrices is decrease in uploaded data.

Work Queue & Condor Condor: A framework which can be used to manage workload on a cluster of computers.  It tries to make use of cycles from idle computers.  Any Condor process can be stopped by the user of the machine. Work Queue: A system which manages Condor, which works with machines of different processing power and scheduling policies and is fault-tolerant.  A stopped process, for example, is restarted automatically by Work Queue.

Evaluation Metrics Speed-up:  Ratio of execution time using 1 processor to execution time using p processors. Scale-up:  Ratio to execution time using 1 processor to execution time using p processor where problem size is scaled by p.  Here we define problem size as the number of images we use. Overhead:  Overall time subtracted by core computation time.  This will mainly be from I/O operations and communication time between master and workers.

Speed-up

Scale-up

Overhead w.r.t problem size

Overhead w.r.t number of workers

Questions?