COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.

Slides:



Advertisements
Similar presentations
Multiple Processor Systems
Advertisements

Multiple Processor Systems
Distributed Systems CS
COS 461 Fall 1997 Workstation Clusters u replace big mainframe machines with a group of small cheap machines u get performance of big machines on the cost-curve.
Protocols and software for exploiting Myrinet clusters Congduc Pham and the main contributors P. Geoffray, L. Prylli, B. Tourancheau, R. Westrelin.
2. Computer Clusters for Scalable Parallel Computing
Beowulf Supercomputer System Lee, Jung won CS843.
Types of Parallel Computers
CSCI-455/522 Introduction to High Performance Computing Lecture 2.
Cache Coherent Distributed Shared Memory. Motivations Small processor count –SMP machines –Single shared memory with multiple processors interconnected.
Distributed Processing, Client/Server, and Clusters
History of Distributed Systems Joseph Cordina
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
Multiple Processor Systems Chapter Multiprocessors 8.2 Multicomputers 8.3 Distributed systems.
A Comparative Study of Network Protocols & Interconnect for Cluster Computing Performance Evaluation of Fast Ethernet, Gigabit Ethernet and Myrinet.
Comparative Study of Beowulf Clusters and Windows 2000 Clusters By Seshendranath Pitcha.
Operating Systems CS208. What is Operating System? It is a program. It is the first piece of software to run after the system boots. It coordinates the.
Cluster Computing Slides by: Kale Law. Cluster Computing Definition Uses Advantages Design Types of Clusters Connection Types Physical Cluster Interconnects.
1 Computer Science, University of Warwick Architecture Classifications A taxonomy of parallel architectures: in 1972, Flynn categorised HPC architectures.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Chapter 2 Computer Clusters Lecture 2.1 Overview.
Design and Implementation of a Single System Image Operating System for High Performance Computing on Clusters Christine MORIN PARIS project-team, IRISA/INRIA.
CLUSTER COMPUTING Prepared by: Kalpesh Sindha (ITSNS)
Distributed Systems Early Examples. Projects NOW – a Network Of Workstations University of California, Berkely Terminated about 1997 after demonstrating.
Parallel Computing The Bad News –Hardware is not getting faster fast enough –Too many architectures –Existing architectures are too specific –Programs.
1 In Summary Need more computing power Improve the operating speed of processors & other components constrained by the speed of light, thermodynamic laws,
KUAS.EE Parallel Computing at a Glance. KUAS.EE History Parallel Computing.
Computer System Architectures Computer System Software
Lecture 13 Warehouse Computing Prof. Xiaoyao Liang 2015/6/10
Lappeenranta University of Technology / JP CT30A7001 Concurrent and Parallel Computing Introduction to concurrent and parallel computing.
CLUSTER COMPUTING STIMI K.O. ROLL NO:53 MCA B-5. INTRODUCTION  A computer cluster is a group of tightly coupled computers that work together closely.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Beowulf Cluster Jon Green Jay Hutchinson Scott Hussey Mentor: Hongchi Shi.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
Multiple Processor Systems Chapter Multiprocessors 8.2 Multicomputers 8.3 Distributed systems.
Message Passing Computing 1 iCSC2015,Helvi Hartmann, FIAS Message Passing Computing Lecture 1 High Performance Computing Helvi Hartmann FIAS Inverted CERN.
Multiple Processor Systems. Multiprocessor Systems Continuous need for faster computers –shared memory model ( access nsec) –message passing multiprocessor.
1 CMPE 511 HIGH PERFORMANCE COMPUTING CLUSTERS Dilek Demirel İşçi.
Copyright ©2003 Digitask Consultants Inc., All rights reserved Cluster Concepts Digitask Seminar November 29, 1999 Digitask Consultants, Inc.
PARALLEL COMPUTING overview What is Parallel Computing? Traditionally, software has been written for serial computation: To be run on a single computer.
CLUSTER COMPUTING TECHNOLOGY BY-1.SACHIN YADAV 2.MADHAV SHINDE SECTION-3.
Copyright © 2011 Curt Hill MIMD Multiple Instructions Multiple Data.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
Distributed Programming CA107 Topics in Computing Series Martin Crane Karl Podesta.
Distributed Computing Systems CSCI 6900/4900. Review Distributed system –A collection of independent computers that appears to its users as a single coherent.
By Chi-Chang Chen.  Cluster computing is a technique of linking two or more computers into a network (usually through a local area network) in order.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
3/12/2013Computer Engg, IIT(BHU)1 INTRODUCTION-2.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Performance Evaluation of Parallel Algorithms on a Computational Grid Environment Simona Blandino 1, Salvatore Cavalieri 2 1 Consorzio COMETA, 2 Faculty.
Background Computer System Architectures Computer System Software.
Computer Science and Engineering Parallel and Distributed Processing CSE 8380 April 28, 2005 Session 29.
Introduction Goal: connecting multiple computers to get higher performance – Multiprocessors – Scalability, availability, power efficiency Job-level (process-level)
SYSTEM MODELS FOR ADVANCED COMPUTING Jhashuva. U 1 Asst. Prof CSE
COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques Dr. Xiao Qin Auburn University
Constructing a system with multiple computers or processors 1 ITCS 4/5145 Parallel Programming, UNC-Charlotte, B. Wilkinson. Jan 13, 2016.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Berkeley Cluster Projects
CLUSTER COMPUTING Presented By, Navaneeth.C.Mouly 1AY05IS037
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Distributed Systems CS
Types of Parallel Computers
Presentation transcript:

COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC

COMP381 by M. Hamdi 2 Some Conclusions About Multiprocessors Small size multiprocessors ( < 10 processors) –Use shared memory with shared bus –Not expensive –Commercially available, and highly used as small servers Medium size multiprocessors ( < 64 processors) –Use shared memory with crossbar switch –Commercially available –Used as high-end servers and computing engines Large size multiprocessors ( > 64 processors) –Use distributed memory with custom-made interconnection network (e.g., 3D Torus) –Very powerful computing machines –Extremely expensive

COMP381 by M. Hamdi 3 Shared Pool of Computing Resources: Processors, Memory, Disks Interconnect Guarantee at least one PC to many individuals (when active) Deliver large % of collective resources to few individuals at any one time Clusters: Network of PCs

COMP381 by M. Hamdi 4 Clusters Dedicated resources

COMP381 by M. Hamdi 5 Scalability Vs. Cost SMP Super Server Departmental Server Personal System Cluster of PCs MPP

COMP381 by M. Hamdi 6 Motivations of using Clusters over Specialized Parallel Computers Individual PCs are becoming increasingly powerful Communication bandwidth between PCs is increasing and latency is decreasing (Gigabit Ethernet, Myrinet) PC clusters are easier to integrate into existing networks Typical low user utilization of PCs (<10%) Development tools for workstations and PCs are mature PC clusters are a cheap and readily available Clusters can be easily grown

COMP381 by M. Hamdi 7 Cluster Architecture Sequential Applications Parallel Applications Parallel Programming Environment Cluster Middleware (Single System Image and Availability Infrastructure) Cluster Interconnection Network/Switch PC/Workstation Network Interface Hardware Communications Software PC/Workstation Network Interface Hardware Communications Software PC/Workstation Network Interface Hardware Communications Software PC/Workstation Network Interface Hardware Communications Software Sequential Applications Parallel Applications

COMP381 by M. Hamdi 8 How Can we Benefit From Clusters?  Given a certain user application Phase 1 –If the application can be run fast enough on a single PC, there is no need to do anything else –Otherwise go to Phase 2 Phase 2 –Try to put the whole application on the DRAM to avoid going to the disk. –If that is not possible, use the DRAM of the other idle workstations –Network DRAM is 5 to 10 times faster than local disk

COMP381 by M. Hamdi 9 Remote Memory Paging Background –Application’s working sets have increased dramatically –Applications require more memory than a single workstation can provide. Solution –Inserts the Network DRAM in the memory hierarchy between local memory and the disk –Swaps the page to remote memory Cache Main Memory Disk Cache Main Memory Disk Network RAM

COMP381 by M. Hamdi 10 How Can we Benefit From Clusters?  In this case, the DRAM of the networked PCs behave like a huge cache system for the disk  Otherwise go to Phase MB 512 MB + Disk All DRAM Networked DRAM Problem size Time

COMP381 by M. Hamdi 11 How Can we Benefit From Clusters? Phase 3 –If the network DRAM is not fast enough, then try using all the disks in the network in parallel for reading and writing data and program code (e.g., RAID) to speedup the I/O –Otherwise go to Phase 4 File Cache P File Cache P File Cache P File Cache P File Cache P File Cache P File Cache P File Cache P Communication Network Network RAID striping Local Cache Cluster Caching

COMP381 by M. Hamdi 12 How Can we Benefit From Clusters? Phase 4 –Execute the program on a multiple number of workstations (PCs) at the same time – Parallel processing Tools –There are many tools that do all these phases in a transparent way (except parallelizing the program) as well as load- balancing and scheduling. Beowulf (CalTech and NASA) - USA Condor - Wisconsin State University, USA MPI (MPI Forum, MPICH is one of the popular implementations) NOW (Network of Workstations) - Berkeley, USA PVM - Oak Ridge National Lab./UTK/Emory, USA

COMP381 by M. Hamdi 13 What network should be used? Fast Ethernet Gigabit Ethernet Myrinet Latency ~120  s ~7  s Bandwidth~100Mbps peak ~1Gbps peak~1.98Gbps real

COMP381 by M. Hamdi Top500 List Clusters are the fastest growing category of supercomputers in the TOP500 List. –406 clusters (81%) in November 2007 list –130 clusters (23%) in the June 2003 list –80 clusters (16%) in the June 2002 list –33 clusters (6.6%) in the June 2001 list 4% of the supercomputers in the November 2007 TOP500 list use Myrinet technology! 54% of the supercomputers in the November 2007 TOP500 list Gigabit Ethernet technology!