March, 2003 SOS 7 Jim Harrell Unlimited Scale Inc.

Slides:



Advertisements
Similar presentations
18 July 2001 Work In Progress – Not for Publication 2001 ITRS Test Chapter ITRS Test ITWG Mike Rodgers Don Edenfeld.
Advertisements

Technology Drivers Traditional HPC application drivers – OS noise, resource monitoring and management, memory footprint – Complexity of resources to be.
System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
Introductions to Parallel Programming Using OpenMP
Clusters Part 1 - Definition of and motivation for clusters Lars Lundberg The slides in this presentation cover Part 1 (Chapters 1-4) in Pfister’s book.
2. Computer Clusters for Scalable Parallel Computing
Beowulf Supercomputer System Lee, Jung won CS843.
Types of Parallel Computers
The Stanford Directory Architecture for Shared Memory (DASH)* Presented by: Michael Bauer ECE 259/CPS 221 Spring Semester 2008 Dr. Lebeck * Based on “The.
©HCCS & IBM® 2008 Stephen Linkin1 Mainframe Hardware Systems And High Availability Stephen S. Linkin Houston Community College © HCCS and IBM 2008.
Introduction to Operating Systems CS-2301 B-term Introduction to Operating Systems CS-2301, System Programming for Non-majors (Slides include materials.
1. Overview  Introduction  Motivations  Multikernel Model  Implementation – The Barrelfish  Performance Testing  Conclusion 2.
Data Analytics and Dynamic Languages Lee E. Edlefsen, Ph.D. VP of Engineering 1.
Some Thoughts on Technology and Strategies for Petaflops.
1 BGL Photo (system) BlueGene/L IBM Journal of Research and Development, Vol. 49, No. 2-3.
Computational Astrophysics: Methodology 1.Identify astrophysical problem 2.Write down corresponding equations 3.Identify numerical algorithm 4.Find a computer.
1 Breakout thoughts (compiled with N. Carter): Where will RAMP be in 3-5 Years (What is RAMP, where is it going?) Is it still RAMP if it is mapping onto.
1 Computer Science, University of Warwick Architecture Classifications A taxonomy of parallel architectures: in 1972, Flynn categorised HPC architectures.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Stack Management Each process/thread has two stacks  Kernel stack  User stack Stack pointer changes when exiting/entering the kernel Q: Why is this necessary?
ASHIMA KALRA.  COMPUTER NETWORK  Local Area Network (LAN) Local Area Network (LAN)  Metropolitan Area Network(MAN) Metropolitan Area Network(MAN) 
Programming for High Performance Computers John M. Levesque Director Cray’s Supercomputing Center Of Excellence.
CLUSTER COMPUTING Prepared by: Kalpesh Sindha (ITSNS)
1 In Summary Need more computing power Improve the operating speed of processors & other components constrained by the speed of light, thermodynamic laws,
Yavor Todorov. Introduction How it works OS level checkpointing Application level checkpointing CPR for parallel programing CPR functionality References.
1 Developing Native Device for MPJ Express Advisor: Dr. Aamir Shafi Co-advisor: Ms Samin Khaliq.
Operating System Review September 10, 2012Introduction to Computer Security ©2004 Matt Bishop Slide #1-1.
Why do so many chips fail? Ira Chayut, Verification Architect (opinions are my own and do not necessarily represent the opinion of my employer)
Small File File Systems USC Jim Pepin. Level Setting  Small files are ‘normal’ for lots of people Metadata substitute (lots of image data are done this.
CLUSTER COMPUTING STIMI K.O. ROLL NO:53 MCA B-5. INTRODUCTION  A computer cluster is a group of tightly coupled computers that work together closely.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Chapter 2 Computer Clusters Lecture 2.2 Computer Cluster Architectures.
Multi-core Programming Introduction Topics. Topics General Ideas Moore’s Law Amdahl's Law Processes and Threads Concurrency vs. Parallelism.
A Virtual Machine Monitor for Utilizing Non-dedicated Clusters Kenji Kaneda Yoshihiro Oyama Akinori Yonezawa (University of Tokyo)
Cluster Workstations. Recently the distinction between parallel and distributed computers has become blurred with the advent of the network of workstations.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Frank Casilio Computer Engineering May 15, 1997 Multithreaded Processors.
HPC User Forum Back End Compiler Panel SiCortex Perspective Kevin Harris Compiler Manager April 2009.
Large Scale Parallel File System and Cluster Management ICT, CAS.
Issues in Accelerator Control Bob Dalesio, December 23, 2002.
Copyright © 2011 Curt Hill MIMD Multiple Instructions Multiple Data.
Experts in numerical algorithms and HPC services Compiler Requirements and Directions Rob Meyer September 10, 2009.
Distributed Programming CA107 Topics in Computing Series Martin Crane Karl Podesta.
Simics: A Full System Simulation Platform Synopsis by Jen Miller 19 March 2004.
Outline Why this subject? What is High Performance Computing?
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
HOW A COMPUTER PROCESSES DATA. What is hardware? Hardware is the electric, electronic and mechanical equipment that makes up a computer What is software?
Spring EE 437 Lillevik 437s06-l22 University of Portland School of Engineering Advanced Computer Architecture Lecture 22 Distributed computer Interconnection.
Tackling I/O Issues 1 David Race 16 March 2010.
Background Computer System Architectures Computer System Software.
Computer Science and Engineering Parallel and Distributed Processing CSE 8380 April 28, 2005 Session 29.
SYSTEM MODELS FOR ADVANCED COMPUTING Jhashuva. U 1 Asst. Prof CSE
Chapter 16 Client/Server Computing Dave Bremer Otago Polytechnic, N.Z. ©2008, Prentice Hall Operating Systems: Internals and Design Principles, 6/E William.
Constructing a system with multiple computers or processors 1 ITCS 4/5145 Parallel Programming, UNC-Charlotte, B. Wilkinson. Jan 13, 2016.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Productive Performance Tools for Heterogeneous Parallel Computing
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
The Mach System Sri Ramkrishna.
What is Fibre Channel? What is Fibre Channel? Introduction
Distributed Shared Memory
CLUSTER COMPUTING.
Characteristics of Reconfigurable Hardware
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
The performance requirements for DSP applications continue to grow and the traditional solutions do not adequately address this new challenge Paradigm.
Alternative Processor Panel Results 2008
MPJ: A Java-based Parallel Computing System
Chapter-1 Computer is an advanced electronic device that takes raw data as an input from the user and processes it under the control of a set of instructions.
Types of Parallel Computers
CS 295: Modern Systems Organizing Storage Devices
Presentation transcript:

March, 2003 SOS 7 Jim Harrell Unlimited Scale Inc.

The Good The Bad The Future

Good things are unimportant. Bad things should be fixed. The Future is what you make of it. Ron asked for Opinions…

Introduction Everything begins and ends with the hardware and applications System Software is just glue – it is nothing by itself –My background System Software projects all the hardware capabilities to applications in a easy to use well understood fashion –System Services – HSM, etc. must be out of the way… –Systems have to be up to be used –One system is better than 10 thousand

Good Processors are cheap and fast –Peak speeds are like gas mileage projections –Hot – and poorly cooled –Complex – additions not always useful –Driven out specialized processors and software –Fewer choices – used to be more processors than companies… Memory is cheap –Larger memory is traded off for expensive and fast Interconnects exist –Just implemented as I/O devices I/O channel speeds are as fast as HPC systems 10 years ago –More than 10x cheaper –Striping data can extend across nodes

The Good Unix is the OS of choice –20+ years of dominating delivered systems in HPTC –Software system research is irrelevant – (Rob Pike) –Linux is Unix Recognizable at source level Only Unix for IA32 Java is programming MPPs are gone –Commodity pressure on engineering costs –Late generation MPPs had single process space and single filename space – this should be salvaged We have clusters –This is good – expect small smps for the foreseeable future Multi-threading is hard and all components have to be threaded..

The Bad Operating System development and checkout should not be done in public –New systems and larger systems always provide surprises –Onsite integration of what ought to be compatible equipment is often hard to do.. –Traditionally large systems need to be shipped and require some work onsite

The Bad I/O and filesystems don’t work well enough –MPP model of data delivery requires high speed interconnect –Years of MPPs have reduced the expectations for I/O performance Expectations appear to be for large sequential I/O –Clusters have to share data using software – this becomes a difficult filesystem issue –Need global filename spaces and reasonable performance Cluster filesystems may be only part of the answer. Quorum is a Latin word. Latin is a dead language.

The Bad Complexity … –The level of complexity is too high for applications MPI has a high learning curve and looks like a protocol –Specialization and growth can create distance -

The Future Heterogeneous processors – in a single system –Use processors for the work suited to the capabilities –Apply specialized cpus, PIMM, FPGA, etc. to application work –Support with commodity processors Interconnects built as interconnects - not I/O devices Distributed system software instead of cluster service software –Single systems – single process space, global resource management, and single filename space systems CoArray Fortran, upc, etc. – move towards integrating parallelism in programming environments Cycle of commodity dominance has not peaked – but will