Evolution of High Performance Cluster Architectures David E. Culler NPACI 2001 All Hands Meeting.

Slides:



Advertisements
Similar presentations
Hardware & the Machine room Week 5 – Lecture 1. What is behind the wall plug for your workstation? Today we will look at the platform on which our Information.
Advertisements

Technology Drivers Traditional HPC application drivers – OS noise, resource monitoring and management, memory footprint – Complexity of resources to be.
System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
Beowulf Supercomputer System Lee, Jung won CS843.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
History of Distributed Systems Joseph Cordina
Module 8: Concepts of a Network Load Balancing Cluster
NPACI Panel on Clusters David E. Culler Computer Science Division University of California, Berkeley
UCB Millennium and the Vineyard Cluster Architecture Phil Buonadonna University of California, Berkeley
1 Improving Web Servers performance Objectives:  Scalable Web server System  Locally distributed architectures  Cluster-based Web systems  Distributed.
Developing a Cluster Strategy for NPACI All Hands Meeting Panel Feb 11, 2000 David E. Culler Computer Science Division University of California, Berkeley.
Finale’ cs294-8 Design of Deeply Networked Systems Spring 2000 David Culler & Randy Katz U.C. Berkeley
IPPS 981 Berkeley FY98 Resource Working Group David E. Culler Computer Science Division U.C. Berkeley
Cluster Computing Slides by: Kale Law. Cluster Computing Definition Uses Advantages Design Types of Clusters Connection Types Physical Cluster Interconnects.
A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares, Alexander Loukissas, Amin Vahdat Presented by Gregory Peaker and Tyler Maclean.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
Storage area network and System area network (SAN)
Module – 7 network-attached storage (NAS)
Chapter 2 Computer Clusters Lecture 2.1 Overview.
Vision/Benefits/Introduction Randy Armstrong (OPC Foundation)
CLUSTER COMPUTING Prepared by: Kalpesh Sindha (ITSNS)
SANPoint Foundation Suite HA Robert Soderbery Sr. Director, Product Management VERITAS Software Corporation.
LECTURE 9 CT1303 LAN. LAN DEVICES Network: Nodes: Service units: PC Interface processing Modules: it doesn’t generate data, but just it process it and.
UC Berkeley 1 The Datacenter is the Computer David Patterson Director, RAD Lab January, 2007.
1/29/2002 CS Distributed Systems 1 Infiniband Architecture Aniruddha Bohra.
Common Devices Used In Computer Networks
+ CS 325: CS Hardware and Software Organization and Architecture Cloud Architectures.
Chord & CFS Presenter: Gang ZhouNov. 11th, University of Virginia.
CLUSTER COMPUTING STIMI K.O. ROLL NO:53 MCA B-5. INTRODUCTION  A computer cluster is a group of tightly coupled computers that work together closely.
Trends In Network Industry - Exploring Possibilities for IPAC Network Steven Lo.
Chapter 1. Introduction. By Sanghyun Ahn, Deot. Of Computer Science and Statistics, University of Seoul A Brief Networking History §Internet – started.
1 Web Server Administration Chapter 2 Preparing For Server Installation.
NETWORKING COMPONENTS AN OVERVIEW OF COMMONLY USED HARDWARE Christopher Johnson LTEC 4550.
March 9, 2015 San Jose Compute Engineering Workshop.
Application Block Diagram III. SOFTWARE PLATFORM Figure above shows a network protocol stack for a computer that connects to an Ethernet network and.
Distributed Information Systems. Motivation ● To understand the problems that Web services try to solve it is helpful to understand how distributed information.
Orange Coast College Business Division Computer Science Department CS 116- Computer Architecture Multiprocessors.
An Architecture and Prototype Implementation for TCP/IP Hardware Support Mirko Benz Dresden University of Technology, Germany TERENA 2001.
Cluster Software Overview
1 Secure Peer-to-Peer File Sharing Frans Kaashoek, David Karger, Robert Morris, Ion Stoica, Hari Balakrishnan MIT Laboratory.
Computer Science and Engineering Copyright by Hesham El-Rewini Advanced Computer Architecture CSE 8383 April 11, 2006 Session 23.
LRPC Firefly RPC, Lightweight RPC, Winsock Direct and VIA.
HADOOP DISTRIBUTED FILE SYSTEM HDFS Reliability Based on “The Hadoop Distributed File System” K. Shvachko et al., MSST 2010 Michael Tsitrin 26/05/13.
ProActive Infrastructure Eric Brewer, David Culler, Anthony Joseph, Randy Katz Computer Science Division U.C. Berkeley ninja.cs.berkeley.edu Active Networks.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
SYSTEM MODELS FOR ADVANCED COMPUTING Jhashuva. U 1 Asst. Prof CSE
Network Processing Systems Design
Parallel Virtual File System (PVFS) a.k.a. OrangeFS
Enhancements for Voltaire’s InfiniBand simulator
Video Security Design Workshop:
MAC Address Tables on Connected Switches
Berkeley Cluster Projects
Infiniband Architecture
Planning and Troubleshooting Routing and Switching
Parallel Computers Definition: “A parallel computer is a collection of processing elements that cooperate and communicate to solve large problems fast.”
Introduction to Networks
U.C. Berkeley Millennium Project
Storage Virtualization
Module – 7 network-attached storage (NAS)
Storage Networking Protocols
Web Server Administration
Storage area network and System area network (SAN)
Constructing a system with multiple computers or processors
System Models and Networking Chapter 2,3
Towards Distributed Test-Lab for Planetary-Scale Services
Types of Parallel Computers
Cluster Computers.
Presentation transcript:

Evolution of High Performance Cluster Architectures David E. Culler NPACI 2001 All Hands Meeting

Much has changed since “NOW” NOW0 HP+medusa FDDI NOW1 SS+ATM/Myrinet NOW 110 UltraSparc +Myrinet inktomi.berkeley.edu

Millennium Cluster Editions

The Basic Argument performance cost of engineering lag –miss the 2x per 18 months –=> rapid assembly of leading edge HW and SW building blocks –=> availability through fault masking, not inherent reliability emergence of the “killer switch” opportunities for innovation –move data between as fast as within machine –protected user-level communication –large-scale management –fault isolation –novel applications

Clusters Took Off scalable internet services –only way to match growth rate changing supercomputer market web hosting

Engineering the Building Block argument came full circle in ~98 wide-array of 3U, 2U, 1U rack-mounted servers –thermals and mechanicals –processing per square-foot –110 AC routing a mixed blessing –component OS & drivers became the early entry to the market

Emergence of the Killer Switch ATM, Fiberchannel, FDDI “died” ServerNet bumps along IBM, SGI do the proprietary thing little Myrinet just keeps going –quite nice at this stage SAN standards shootout –NGIO + FutureIO => Infiniband –specs entire stack from phy to api »nod to IPv6 –big, complex, deeply integrated, DBC Gigabit EtherNet steamroller... –limited by TCP/IP stack, NIC, and cost

Opportunities for Innovation

Unexpected Breakthru: layer-7 switches fell out of modern switch design –process packets in chunks vast # of simultaneous connections many line-speed packet filters per port can be made redundant => multi-gigabit cluster “front end” –virtualize IP address of services –move service within cluster –replicate it, distribute it  high-level xforms  fail-over,  load management Layer-7Switch Network Switch

e-Science any useful app should be a service

Protected User-level messaging  Virtual Interface Architecture (VIA) emerged  primitive & complex relative to academic prototypes  industrial compromise  went dormant  Incorporated in Infiniband  big one to watch  Potential breakthrough  user-level TCP, UDP with IP NIC  storage over IP

Management workstation -> PC transition a step back –boot image distribution, OS distribution –network troubleshoot and service multicast proved a powerful tool emerging health monitoring and control –HW level –service level –OS level still a problem

Rootstock Local Rootstock Server Internet Rootstock Server Local Rootstock Server UC Berkeley

Ganglia and REXEC rexec d vexecd (Policy A) rexec Cluster IP Multicast Channel %rexec –n 2 –r 3 indexer minimum $ vexecd (Policy B) Node ANode BNode CNode D “Nodes AB” Also: bWatch BPROC: Beowulf Distributed Process Space VA Linux Systems: VACM, VA Cluster Manager

Network Storage state-of-practice still NFS + local copies local disk replica management lacking NFS doesn’t scale –major source of naive user frustration limited structured parallel access SAN movement only changing the device interface Need cluster content distribution, caching, parallel access and network striping see: GPFS, CFS, PVFS, HPSS, GFS,PPFS,CXFS, HAMFS,Petal, NASD...

Distributed Persistent Data Structure Alternative Service DDS lib Storage “brick” Service DDS lib Service DDS lib Storage “brick” Storage “brick” Storage “brick” Storage “brick” Storage “brick” System Area Network Clustered Service Distr Hash table API Single-node durable hash table Redundant low latency high xput network

Scalable Throughput

“Performance Available” Storage D A D A D A D A Static Parallel Aggregation D A D A D A D A Distributed Queue Adaptive Parallel Aggregation

Application Software very little movement towards harnessing architectural potential application as service –process stream of requests (not shell or batch) –grow & shrink on demand –replication for availability »data and functionality –tremendous internal bandwidth outer-level optimizations, not algorithmic

Time is NOW  f inish the system area network  tackle the cluster I/O problem  come together around management tools  get serious about application services