Chapter 2 Computer Clusters Lecture 2.1 Overview.

Slides:



Advertisements
Similar presentations
2. Computer Clusters for Scalable Parallel Computing
Advertisements

Beowulf Supercomputer System Lee, Jung won CS843.
Distributed Systems 1 Topics  What is a Distributed System?  Why Distributed Systems?  Examples of Distributed Systems  Distributed System Requirements.
Dinker Batra CLUSTERING Categories of Clusters. Dinker Batra Introduction A computer cluster is a group of linked computers, working together closely.
CREAM-CE status and evolution plans Paolo Andreetto, Sara Bertocco, Alvise Dorigo, Eric Frizziero, Alessio Gianelle, Massimo Sgaravatto, Lisa Zangrando.
Lincoln University Canterbury New Zealand Evaluating the Parallel Performance of a Heterogeneous System Elizabeth Post Hendrik Goosen formerly of Department.
Distributed Processing, Client/Server, and Clusters
Single System Image Clustering. Source ex.pl?node_id=38692&lastnode_id=131
1 CIS450/IMSE450/ECE478 Operating Systems Winter 2003 Professor Jinhua Guo.
Cluster Computing Slides by: Kale Law. Cluster Computing Definition Uses Advantages Design Types of Clusters Connection Types Physical Cluster Interconnects.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
Is Windows Right for High-Availability Enterprise Applications? Dan Kusnetzky, Vice President System Software Research IDC.
Design and Implementation of a Single System Image Operating System for High Performance Computing on Clusters Christine MORIN PARIS project-team, IRISA/INRIA.
Ch 4. The Evolution of Analytic Scalability
CLUSTER COMPUTING Prepared by: Kalpesh Sindha (ITSNS)
Distributed Systems Early Examples. Projects NOW – a Network Of Workstations University of California, Berkely Terminated about 1997 after demonstrating.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
1 In Summary Need more computing power Improve the operating speed of processors & other components constrained by the speed of light, thermodynamic laws,
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 7 2/23/2015.
LINUX/UNIX WORKSTATIONS Franklin Montenegro Carlos Sierra.
1 Lecture 20: Parallel and Distributed Systems n Classification of parallel/distributed architectures n SMPs n Distributed systems n Clusters.
N. GSU Slide 1 Chapter 02 Cloud Computing Systems N. Xiong Georgia State University.
DISTRIBUTED COMPUTING
UNIT - 1Topic - 2 C OMPUTING E NVIRONMENTS. What is Computing Environment? Computing Environment explains how a collection of computers will process and.
Seaborg Cerise Wuthrich CMPS Seaborg  Manufactured by IBM  Distributed Memory Parallel Supercomputer  Based on IBM’s SP RS/6000 Architecture.
CLUSTER COMPUTING STIMI K.O. ROLL NO:53 MCA B-5. INTRODUCTION  A computer cluster is a group of tightly coupled computers that work together closely.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming in C with MPI and OpenMP Michael J. Quinn.
Cluster Workstations. Recently the distinction between parallel and distributed computers has become blurred with the advent of the network of workstations.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
1 Multiprocessor and Real-Time Scheduling Chapter 10 Real-Time scheduling will be covered in SYSC3303.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
N. GSU Slide 1 Chapter 05 Clustered Systems for Massive Parallelism N. Xiong Georgia State University.
Ch System Models for Distributed and Cloud Computing Classification of Massive systems (Table 1.2) Clusters of Cooperative Computers 
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
1 CMPE 511 HIGH PERFORMANCE COMPUTING CLUSTERS Dilek Demirel İşçi.
PARALLEL COMPUTING overview What is Parallel Computing? Traditionally, software has been written for serial computation: To be run on a single computer.
CLUSTER COMPUTING TECHNOLOGY BY-1.SACHIN YADAV 2.MADHAV SHINDE SECTION-3.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
Computing Environment The computing environment rapidly evolving ‑ you need to know not only the methods, but also How and when to apply them, Which computers.
CCSM Performance, Successes and Challenges Tony Craig NCAR RIST Meeting March 12-14, 2002 Boulder, Colorado, USA.
7. Grid Computing Systems and Resource Management
Distributed Computing Systems CSCI 6900/4900. Review Distributed system –A collection of independent computers that appears to its users as a single coherent.
By Chi-Chang Chen.  Cluster computing is a technique of linking two or more computers into a network (usually through a local area network) in order.
Server HW CSIS 4490 n-Tier Client/Server Dr. Hoganson Server Hardware Mission-critical –High reliability –redundancy Massive storage (disk) –RAID for redundancy.
Outline Why this subject? What is High Performance Computing?
Data Communications and Networks Chapter 9 – Distributed Systems ICT-BVF8.1- Data Communications and Network Trainer: Dr. Abbes Sebihi.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
1 MSCS 237 Introduction to Distributed Systems. 2 Outline Distributed computing (DC) Distributed Systems (DS) Motivation Architecture of a DS Advantages.
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-2.
The History of Clustering. What is computer clustering? Computer clustings is when a group of computers are linked together operating as one, sharing.
Cluster computing. 1.What is cluster computing? 2.Need of cluster computing. 3.Architecture 4.Applications of cluster computing 5.Advantages of cluster.
Computer System Evolution. Yesterday’s Computers filled Rooms IBM Selective Sequence Electroinic Calculator, 1948.
SYSTEM MODELS FOR ADVANCED COMPUTING Jhashuva. U 1 Asst. Prof CSE
Chapter 16 Client/Server Computing Dave Bremer Otago Polytechnic, N.Z. ©2008, Prentice Hall Operating Systems: Internals and Design Principles, 6/E William.
Page : 1 SC2004 Pittsburgh, November 12, 2004 DEISA : integrating HPC infrastructures in Europe DEISA : integrating HPC infrastructures in Europe Victor.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
Clouds , Grids and Clusters
Advanced Topics in Concurrency and Reactive Programming: Case Study – Google Cluster Majeed Kassis.
Grid Computing.
Grid Computing Colton Lewis.
Chapter 16: Distributed System Structures
IT Infrastructure: Hardware and Software
Chapter 17: Database System Architectures
Chapter 17 Parallel Processing
Ch 4. The Evolution of Analytic Scalability
CLUSTER COMPUTING.
IT Infrastructure: Hardware and Software
Subject Name: Operating System Concepts Subject Number:
Presentation transcript:

Chapter 2 Computer Clusters Lecture 2.1 Overview

Outline Cluster definition and significance Cluster development trend Cluster family classification Analysis of the Top 500 Supercomputers

Definition and Significance Definition – A Collection of interconnected stand-alone computer which can work together collectively and cooperatively as a single integrated computer resource pool – Clustering explores massive parallelism at the job level and achieves high availability though stand- alone operations

Significance – Of the top 500 supercomputers reported in 2010, 85% were computer clusters or MPPs built with homogeneous nodes. – Computer clusters have laid the foundation for today’s supercomputers, computational grids and Internet clouds build over data centers

Cluster Development Trends Computer clustering started with the linking of large mainframe computers such as the IBM Sysplex and the SCI Origin Subsequently, the clustering trend moved toward the networking of many minicomputers, such as Dec’s VMS cluster. In the early 1990s, the next move was to build UNIX- based workstation clusters represented by the Berkeley NOW (network of workstations) and IBM SP2 AIX- based server cluster Beyond 2000, we see the trend moving to the clustering x86 PC engines.

Milestone Cluster Systems – Clustering has been a hot research challenge in computer architecture. Fast communication, job scheduling, SSI HA – The following tables lists some milestone cluster research projects and commercial cluster products, each of which has contributed some unique features.

Cluster Family Classification Based on application demand, computer clusters are divided into three classes: – Computer Clusters – High-Availability Clusters – Load-balancing Clusters

– Compute Clusters These are clusters designed mainly for collective computation over a single large job, such as large scale simulation and modeling. The cluster typically shares a dedicated network, and the nodes are mostly homogeneous and tightly coupled. This type of clusters is also known as a Beowulf cluster. Beowulf cluster typically runs Linux, or other Unix-like operating systems such as BSD, and contain shared libraries. Such as Message-Passing Interface (MPI), the Beowulf cluster behaves more like a single (super)computer.

– High-Availability clusters HA (high-availability) clusters are designed to be fault- tolerant and achieve HA of services. HA clusters operate with many redundant nodes to sustain faults or failures. They are designed to avoid all single points of failure. Many commercial HA clusters are available for various operating systems.

– Load-balancing clusters For this type of clusters, requests initiated from the user are distributed to all node computers, which results in a balanced workload among different machines. Middleware is needed to achieve dynamic load balancing by job or process migration among all the cluster nodes.

Analysis of the Top 500 Supercomputers Every six months, the world’s Top 500 supercomputers are evaluated by running the Linpack Benchmark program over very large data sets. The ranking varies from year to year, similar to a competition. We will analyze the trend of top 500 supercomputers over time from the following perspectives: – Architectural Evolution – Speed Improvement over time – Operating System Trends

Architectural Evolution

Speed Improvement Over Time

Operating System Trends in the Top 500 – The five most popular operating systems have more than a 10 percent share among the Top 500 Computers in November – 410 supercomputers are using Linux with a total processor count exceeding 4.5 million. This constitutes 82 percent of the systems adopting Linux. – The IBM AIX/OS is in second place with 17 systems (a 3.4 percent share) and more than 94,288 processors. – Third place is represented by the combined use of the SLES10 with the SGI ProPack5, with 15 systems (3 percent) over 135,200 processors. – Fourth place goes to the CNK/SLES9 used by 14 systems (2.8 percent) over 1.13 million processors. – Finally, the CNL/OS was used in 10 systems (2 percent) over 178,577 processors. – The remaining 34 systems applied 13 other operating systems with a total share of only 6.8 percent.

The Top Five Systems in 2010

Image from

Image from

Image from aibob.blogspot.com

Image from petaflop-supercomputer-tsubame-2-0-to-be-launched-this-autumn

Top