Highly Distributed Parallel Computing Neil Skrypuch COSC 3P93 3/21/2007.

Slides:



Advertisements
Similar presentations
Data Communications and Networking
Advertisements

Buffers & Spoolers J L Martin Think about it… All I/O is relatively slow. For most of us, input by typing is painfully slow. From the CPUs point.
Distributed Processing, Client/Server and Clusters
Chapter 1 Computer, Internet, Web, and Basics
Dinker Batra CLUSTERING Categories of Clusters. Dinker Batra Introduction A computer cluster is a group of linked computers, working together closely.
Chapter 9 Designing Systems for Diverse Environments.
8.
James Tam Computer Networks You will learn what is a network, how they work and what are the different types.
Symantec De-Duplication Solutions Complete Protection for your Information Driven Enterprise Richard Hobkirk Sr. Pre-Sales Consultant.
Scaling Distributed Machine Learning with the BASED ON THE PAPER AND PRESENTATION: SCALING DISTRIBUTED MACHINE LEARNING WITH THE PARAMETER SERVER – GOOGLE,
James Tam Computer Networks You will learn what is a network, how they work and what are the different types.
CS10 The Beauty and Joy of Computing Lecture #19 Distributed Computing Researchers at Indiana U used data mining techniques to uncover evidence.
Atomistic Protein Folding Simulations on the Submillisecond Timescale Using Worldwide Distributed Computing Qing Lu CMSC 838 Presentation.
Introduction What is Parallel Algorithms? Why Parallel Algorithms? Evolution and Convergence of Parallel Algorithms Fundamental Design Issues.
©Silberschatz, Korth and Sudarshan18.1Database System Concepts Centralized Systems Run on a single computer system and do not interact with other computer.
Data-centric computing with Netezza Architecture DISC reading group September 24, 2007.
Improving Data Access in P2P Systems Karl Aberer and Magdalena Punceva Swiss Federal Institute of Technology Manfred Hauswirth and Roman Schmidt Technical.
SM3121 Software Technology Mark Green School of Creative Media.
Distributed MapReduce Team B Presented by: Christian Bryan Matthew Dailey Greg Opperman Nate Piper Brett Ponsler Samuel Song Alex Ostapenko Keilin Bickar.
Tiered architectures 1 to N tiers. 2 An architectural history of computing 1 tier architecture – monolithic Information Systems – Presentation / frontend,
Network Topologies.
Ch 4. The Evolution of Analytic Scalability
1 Reasons for parallelization Can we make GA faster? One of the most promising choices is to use parallel implementations. The reasons for parallelization.
CLUSTER COMPUTING Prepared by: Kalpesh Sindha (ITSNS)
Distributed Systems Early Examples. Projects NOW – a Network Of Workstations University of California, Berkely Terminated about 1997 after demonstrating.
Day 10 Hardware Fault Tolerance RAID. High availability All servers should be on UPSs –2 Types Smart UPS –Serial cable connects from UPS to computer.
Computer System Architectures Computer System Software
Calculating Discrete Logarithms John Hawley Nicolette Nicolosi Ryan Rivard.
What Is A Network A network is a group of computers interconnected with communication lines which allows users to share information and resources.
1 Lecture 20: Parallel and Distributed Systems n Classification of parallel/distributed architectures n SMPs n Distributed systems n Clusters.
Multi-media Computers and Computer Networks. Questions ? Media is used for ………………… Multimedia computer is capable of integrating ………………………………….. OCR stands.
 Introduction to Operating System Introduction to Operating System  Types Of An Operating System Types Of An Operating System  Single User Single User.
Performance Concepts Mark A. Magumba. Introduction Research done on 1058 correspondents in 2006 found that 75% OF them would not return to a website that.
DCE (distributed computing environment) DCE (distributed computing environment)
Cloud Computing Characteristics A service provided by large internet-based specialised data centres that offers storage, processing and computer resources.
Data Warehousing 1 Lecture-24 Need for Speed: Parallelism Virtual University of Pakistan Ahsan Abdullah Assoc. Prof. & Head Center for Agro-Informatics.
Introduction  Client/Server technology is seen by many as the solution to the difficulty of linking together the various departments of corporation.
Operating System Concepts Chapter One: Introduction What is an operating system? Simple Batch Systems Multiprogramming Systems Time-Sharing Systems Personal-Computer.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
April 26, CSE8380 Parallel and Distributed Processing Presentation Hong Yue Department of Computer Science & Engineering Southern Methodist University.
G063 - Distributed Databases. Learning Objectives: By the end of this topic you should be able to: explain how databases may be stored in more than one.
Parallel Processing Steve Terpe CS 147. Overview What is Parallel Processing What is Parallel Processing Parallel Processing in Nature Parallel Processing.
Mr C Johnston ICT Teacher BTEC IT Unit 05 - Lesson 03 Network Topologies.
MapReduce and GFS. Introduction r To understand Google’s file system let us look at the sort of processing that needs to be done r We will look at MapReduce.
Presenters: Rezan Amiri Sahar Delroshan
1 CMPE 511 HIGH PERFORMANCE COMPUTING CLUSTERS Dilek Demirel İşçi.
PARALLEL COMPUTING overview What is Parallel Computing? Traditionally, software has been written for serial computation: To be run on a single computer.
CLUSTER COMPUTING TECHNOLOGY BY-1.SACHIN YADAV 2.MADHAV SHINDE SECTION-3.
CENTRALISED AND CLIENT / SERVER DBMS. Topics To Be Discussed………………………. (A) Centralized DBMS (i) IntroductionIntroduction (ii) AdvantagesAdvantages (ii)
Distributed Information Systems. Motivation ● To understand the problems that Web services try to solve it is helpful to understand how distributed information.
Copyright © 2011 Curt Hill MIMD Multiple Instructions Multiple Data.
Multimedia & Mobile Communications Lab.
SHAPE OF A NETWORK COPYRIGHT BTS TOPOLOGY The way the computers are cabled together Four different layouts Logical topology describes the way data travels.
The concept of RAID in Databases By Junaid Ali Siddiqui.
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
Cracking the DES Encryption
PARALLELIZATION OF ARTIFICIAL NEURAL NETWORKS Joe Bradish CS5802 Fall 2015.
| nectar.org.au NECTAR TRAINING Module 4 From PC To Cloud or HPC.
Outline Why this subject? What is High Performance Computing?
Data Communications and Networks Chapter 9 – Distributed Systems ICT-BVF8.1- Data Communications and Network Trainer: Dr. Abbes Sebihi.
Chapter 7 Channel organization. Group members  Bilal Ahmed  Mehmal javed  Faisal khan janjua  Harris bashir.
What Is A Network A network is a group of computers interconnected with communication lines which allows users to share information and resources.
Cluster computing. 1.What is cluster computing? 2.Need of cluster computing. 3.Architecture 4.Applications of cluster computing 5.Advantages of cluster.
Background Computer System Architectures Computer System Software.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Grid Computing Colton Lewis.
Distributed Computing with SAGE Yi Qiang
Chapter 16: Distributed System Structures
Cooperative Caching, Simplified
Distributed computing deals with hardware
Presentation transcript:

Highly Distributed Parallel Computing Neil Skrypuch COSC 3P93 3/21/2007

Overview a network of computers all working towards a similar goal  network consists of many nodes, few servers  nodes perform computing and send results to a server  servers distribute jobs  node machines do not communicate with eachother

Pros

Relatively Simple don't need to worry about special interconnections don't need to worry about cluster booting

Non-Homogeneous Network can work across different computer architectures, OSes, etc computers can be of varying speeds doesn't require the fastest or most expensive computers computers can be distributed anywhere in the world

Infrastructure infrastructure for HDPC already exists almost everywhere  anyone with a network of computers is already ready for HDPC lots of programs already exist that take advantage of HDPC

Expansion expansion is painless there are no special constraints on the “shape” of the network not fast enough yet? keep adding more computers until it is

Resilience to Failure it doesn't matter if one or more nodes die only the reliability of the central server(s) matter

Cons

Suitability not all problems are suited to HDPC highly communication bound problems are a poor fit for HDPC

Server Dependence central server dependence is a double edged sword if the central server becomes unavailable, everything grinds to a halt

Network (In)security how to verify if a client should be allowed to join the network? protecting data sent over the network verifying integrity and authenticity of data sent over the network

Network (Un)reliability nodes temporarily losing connectivity may make them temporarily useless

Dealing With the Issues

Server Dependence the central server need not be a single server server itself may be clustered  countless ways to cluster servers

Clustering With a Database allow nodes to talk directly to the database cluster the database over multiple servers  multi-master replication  single master replication  lots more...

Server Hierarchy multiple tiers of servers may also be used could be considered recursive HDPC very similar to the tree architecture of supercomputers

Lost Nodes define a maximum amount of time to wait for a node's response use redundancy  assume some nodes will always be lost  send duplicate jobs to multiple nodes simultaneously

Network (In)security not as big of an issue as one might think encryption and public key infrastructures mitigate most confidentiality and authenticity concerns redundancy is useful for both reliability and security

Work Buffering taking larger portions of work at a time temporary connectivity issues pose less of a problem this way  a node can continue working without talking to a central server for longer

Where is HDPC Useful?

Combinatorics search enumeration generation

Cryptography brute force cipher cracking gives a glimpse of the future, in terms of what the average person will be able to crack

Artificial Intelligence genetic algorithms genetic programming alpha-beta search

Graphics ray tracing animation fractal generation and calculation

Simulation weather and climate modeling particle physics

Guidelines for Suitability most problems involving a large search tree are well suited to HDPC anything that can be broken down into smaller, self-contained, chunks is a good candidate for HDPC

How Well Does HDPC Work?

~200,000 non-dedicated nodes 240 TFLOPS approximately 40 central servers, unknown speeds

~200,000 non-dedicated nodes 288 TFLOPS 10 central servers, all relatively modest

Blue Gene/L currently the fastest supercomputer not HDPC 65,536 dedicated nodes 280 TFLOPS cost about $100,000,000 US

HDPC Works Well typical speedup is close to linear cost is substantially less than a comparable supercomputer nodes can also be general purpose computers

Why Does HDPC Work Well?

Infrastructure Reuse in general, new hardware investments are not necessary  creating new infrastructure is expensive and time consuming it's easy to justify using things you already have for additional purposes there are tons of idle CPUs at any given time, why not use them?

Low Barrier to Entry anyone with a couple of networked computers can start experimenting

Painlessly Scalable smooth curve upwards for both cost and performance

Simpler to Program doesn't require as much “thinking in parallel” in comparison to other approaches  thinking in parallel is hard and fundamentally different than thinking serially pushes the heavy lifting onto the database instead of the application programmer

Commodity Hardware is Fast a typical desktop machine today is more powerful than a supercomputer from 15 years ago  and costs orders of magnitude less  and outputs much less heat  and takes up much less space  and consumes much less power

The Future supercomputers will become faster HDPC will become even faster than supercomputers  as both number of computers and speed increases both supercomputers and HDPC will fill their own separate niche

Questions and Discussion

References