Slide 1 Computers for the Post-PC Era Aaron Brown, Jim Beck, Kimberly Keeton, Rich Martin, David Oppenheimer, Randi Thomas, John Kubiatowicz, Kathy Yelick,

Slides:



Advertisements
Similar presentations
IBM Software Group ® Integrated Server and Virtual Storage Management an IT Optimization Infrastructure Solution from IBM Small and Medium Business Software.
Advertisements

Chapter Six Networking Hardware.
1 Magnetic Disks 1956: IBM (RAMAC) first disk drive 5 Mb – Mb/in $/year 9 Kb/sec 1980: SEAGATE first 5.25’’ disk drive 5 Mb – 1.96 Mb/in2 625.
RAID Technology. Use Arrays of Small Disks? 14” 10”5.25”3.5” Disk Array: 1 disk design Conventional: 4 disk designs Low End High End Katz and Patterson.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
1 Recap (RAID and Storage Architectures). 2 RAID To increase the availability and the performance (bandwidth) of a storage system, instead of a single.
Reliability Week 11 - Lecture 2. What do we mean by reliability? Correctness – system/application does what it has to do correctly. Availability – Be.
Slide 1 Patterson’s Projects, People, Impact Reduced Instruction Set Computer (RISC) –What: simplified instructions to exploit VLSI: ‘80-’84 –With:
Slide 1 Computers for the Post-PC Era John Kubiatowicz, Kathy Yelick, and David Patterson IBM Visit.
Teraserver Darrel Sharpe Matt Todd Rob Neff Mentor: Dr. Palaniappan.
Slide 1 Adaptive Compilers and Runtime Systems Kathy Yelick U.C. Berkeley.
Slide 1 ISTORE: System Support for Introspective Storage Appliances Aaron Brown, David Oppenheimer, and David Patterson Computer Science Division University.
Slide 1 ISTORE: Introspective Storage for Data-Intensive Network Services Aaron Brown, David Oppenheimer, Kimberly Keeton, Randi Thomas, Noah Treuhaft,
Slide 1 ISTORE: An Introspective Storage Architecture for Network Service Applications Aaron Brown, David Oppenheimer, Kimberly Keeton, Randi Thomas, Jim.
Slide 1 Computers for the Post-PC Era Aaron Brown, Jim Beck, Kimberly Keeton, Rich Martin, David Oppenheimer, Randi Thomas, John Kubiatowicz, Kathy Yelick,
By Mr. Abdalla A. Shaame 1. What is Computer An electronic device that stores, retrieves, and processes data, and can be programmed with instructions.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Secondary Storage Unit 013: Systems Architecture Workbook: Secondary Storage 1G.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Chapter 8 Input/Output. Busses l Group of electrical conductors suitable for carrying computer signals from one location to another l Each conductor in.
… when you will open a computer We hope you will not look like …
Chapter 2 Chapter 2: Planning for Server Hardware.
Baydel Founded in 1972 Headquarters: Surrey, England North American Headquarters: San Jose, CA Engineering Driven Organization Specialize in Computer Storage.
SCSI Richard Goldman April 2000
Server Hardware Chapter 22 Release 22/10/2010Jetking Infotrain Ltd.
Guide to Linux Installation and Administration, 2e 1 Chapter 9 Preparing for Emergencies.
CPU (CENTRAL PROCESSING UNIT): processor chip (computer’s brain) found on the motherboard.
1 Fault Tolerance in the Nonstop Cyclone System By Scott Chan Robert Jardine Presented by Phuc Nguyen.
The Basic Input/Output System Unit objectives: Access the BIOS setup utility, change hardware configuration values, and research BIOS updates Explain the.
Eng. Mohammed Timraz Electronics & Communication Engineer University of Palestine Faculty of Engineering and Urban planning Software Engineering Department.
1 CS503: Operating Systems Spring 2014 Dongyan Xu Department of Computer Science Purdue University.
DAC-FF The Ultimate Fibre-to-Fibre Channel External RAID Controller Solution for High Performance Servers, Clusters, and Storage Area Networks (SAN)
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
1 Web Server Administration Chapter 2 Preparing For Server Installation.
Lecture No 11 Storage Devices
What is a BIOS? * basic input/output system (BIOS), also known as the System BIOS * The BIOS software is built into the PC on a non-volatile ROM and is.
1 Selecting LAN server (Week 3, Monday 9/8/2003) © Abdou Illia, Fall 2003.
Computer Architecture Part IV-B: I/O Buses. Chipsets Intelligent bus controller chips found on the motherboard Enable higher speeds on one or more buses.
Large Scale Parallel File System and Cluster Management ICT, CAS.
Slide 1 ISTORE: Introspective Storage for Data-Intensive Network Services Aaron Brown, David Oppenheimer, Kimberly Keeton, Randi Thomas, John Kubiatowicz,
EEL 5708 Cluster computers. Case study: Google Lotzi Bölöni.
Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Technology Education Copyright © 2006 by The McGraw-Hill Companies,
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
Hard Disk Storage  A hard disk drive contains rigid, disk- shaped platters, usually constructed of aluminum or glass §Also known as a Fixed Disk §Grown.
Slide 1 IRAM and ISTORE Projects Aaron Brown, Jim Beck, Rich Fromm, Joe Gebis, Kimberly Keeton, Christoforos Kozyrakis, David Martin, Morley Mao, Rich.
Slide 1 What Happens Before A Disk Fails? Randi Thomas, Nisha Talagala
 The End to the Means › (According to IBM ) › 03.ibm.com/innovation/us/thesmartercity/in dex_flash.html?cmp=blank&cm=v&csr=chap ter_edu&cr=youtube&ct=usbrv111&cn=agus.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
By Harshal Ghule Guided by Mrs. Anita Mahajan G.H.Raisoni Institute Of Engineering And Technology.
1 High-availability and disaster recovery  Dependability concepts:  fault-tolerance, high-availability  High-availability classification  Types of.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Server Upgrade HA/DR Integration
Computer Hard Drive.
Operating Systems (CS 340 D)
Berkeley Cluster: Zoom Project
Introduction to Networks
TYPES OFF OPERATING SYSTEM
IRAM and ISTORE Projects
An Overview of the Computer System
Computers for the Post-PC Era
Design Unit 26 Design a small or home office network
Web Server Administration
Cost Effective Network Storage Solutions
Cluster Computers.
Presentation transcript:

Slide 1 Computers for the Post-PC Era Aaron Brown, Jim Beck, Kimberly Keeton, Rich Martin, David Oppenheimer, Randi Thomas, John Kubiatowicz, Kathy Yelick, and David Patterson Industrial Relations Conference

Slide 2 Outline Motivation and Background: Berkeley’s Past ISTORE Goals Hardware Architecture Software Architecture Discussion and Feedback

Slide 3 Motivation Next generation fixes problems of last gen. 1960s: batch processing + slow turnaround  Timesharing –15-20 years of performance improvement, cost reduction (minicomputers, semiconductor memory) 1980s: Time sharing + inconsistent response times  Workstations/Personal Computers –15-20 years of performance improvement, cost reduction (microprocessors, DRAM memory, disk) 2000s: PCs + difficulty of use/high cost of ownership  ???

Slide 4 Perspective on Post-PC Era PostPC Era Divides built on two technologies: 1) Mobile Consumer Electronic Devices –e.g., successor to PDA, Cell phone –Prior talks in this session –See Posters on Ninja, Iceberg, IRAM (12-1:30) and Post PC session 1:30-3:30 in 306 Soda 2) Infrastructure to Support such Devices –e.g., successor to Big Fat Web Servers, Databases –This talk and Posters on ISTORE (12-1:30)

Slide 5 Background for ISTORE: RAID-I RAID-I (1989) –consisted of a Sun 4/280 workstation with 128 MB of DRAM, four dual- string SCSI controllers, inch SCSI disks and specialized disk striping software

Slide 6 Background for ISTORE: RAID-II RAID-II (1993) –A network attached storage device. 2 outer racks contained 144 disks (3.5” IBM 320 MB SCSI) & power supplies. Center rack in 3 parts: top chassis holds VME disk controller boards, center chassis contains custom crossbar switch and HIPPI network (1Gb/s) interface boards; bottom chassis contains the Sun 4/280 workstation.

Slide 7 Background: Tertiary Disk Tertiary Disk (1997) –cluster of 20 PCs hosting ” IBM disks (8.4 GB) in 7 7’x19” racks, or 3 TB. The 200MHz, 96 MB P6 PCs run FreeBSD and a switched 100Mb/s Ethernet connects the hosts. Also 4 UPS units. –Hosts world’s largest art database:72,000 images in cooperation with San Francisco Fine Arts Museum: Try

Slide 8 Tertiary Disk HW Failure Experience Reliability of hardware components (20 months) 7 IBM SCSI disk failures (out of 364, or 2%) 6 IDE (internal) disk failures (out of 20, or 30%) 1 SCSI controller failure (out of 44, or 2%) 1 SCSI Cable (out of 39, or 3%) 1 Ethernet card failure (out of 20, or 5%) 1 Ethernet switch (out of 2, or 50%) 3 enclosure power supplies (out of 92, or 3%) 1 short power outage (covered by UPS) Did not match expectations: SCSI disks more reliable than cables!

Slide 9 Saw 2 Error Messages per Day SCSI Error Messages: –Time Outs: Response: a BUS RESET command –Parity: Cause of an aborted request Data Disk Error Messages: –Hardware Error: The command unsuccessfully terminated due to a non-recoverable HW failure. –Medium Error: The operation was unsuccessful due to a flaw in the medium (try reassigning sectors) –Recovered Error: The last command completed with the help of some error recovery at the target –Not Ready: The drive cannot be accessed

Slide 10 SCSI Time Outs+Recovered Errors (m0) SCSI Bus 0

Slide 11 Zoom In: Disk Recovered Errors SCSI Bus 0

Slide 12 Can we predict a disk failure? Yes, we can look for Recovered Error messages  on : –There were 433 Recovered Error Messages –These messages lasted for slightly over an hour between: 12:43 and 14:10 On : Disk 5 on m0 was “fired”, i.e. it appeared to operator it was about to fail, so it was swapped

Slide 13 SCSI Time Outs + Hardware Failures (m11) SCSI Bus 0

Slide 14 Can we predict a disk failure? Yes, look for Hardware Error messages –These messages lasted for 8 days between: » and –On disk 9 there were: »1763 Hardware Error Messages, and »297 SCSI Timed Out Messages On : Disk 9 on SCSI Bus 0 of m11 was “fired”, i.e. appeared it was about to fail, so it was swapped

Slide 15 SCSI Bus 2 Parity Errors (m2)

Slide 16 Can We Predict Other Kinds of Failures? Yes, the flurry of parity errors on m2 occurred between: – and , as well as – and On –m2 had a bad enclosure  cables or connections defective –The enclosure was then replaced

Slide 17 Lessons from Tertiary Disk Project Maintenance is hard on current systems –Hard to know what is going on, who is to blame Everything can break –Its not what you expect in advance –Follow rule of no single point of failure Nothing fails fast –Eventually behaves bad enough that operator fires poor performer, but it doesn’t quit Many failures may be predicted

Slide 18 Outline Motivation and Background: Berkeley’s Past ISTORE Goals Hardware Architecture Software Architecture Discussion and Feedback

Slide 19 Storage Priorities: Research v. Users Current Research Priorities 1) Performance 1’) Cost 3) Scalability 4) Availability 5) Maintainability Current Server Customer Priorities 1) Availability 2) Maintainability 3) Scalability 4) Performance 5) Cost (From Sun marketing presentation, 2/99)

Slide 20 Intelligent Storage Project Goals ISTORE: a hardware/software architecture for building scaleable, self-maintaining storage –An introspective system: it monitors itself and acts on its observations Self-maintenance: does not rely on administrators to configure, monitor, or tune system

Slide 21 Self-maintenance Failure management –devices must fail fast without interrupting service –predict failures and initiate replacement –failures  immediate human intervention System upgrades and scaling –new hardware automatically incorporated without interruption –new devices immediately improve performance or repair failures Performance management –system must adapt to changes in workload or access patterns

Slide 22 ISTORE-I Hardware ISTORE uses “intelligent” hardware Intelligent Chassis: scaleable, redundant, fast network + UPS Device CPU, memory, NI Intelligent Disk “Brick”: a disk, plus a fast embedded CPU, memory, and redundant network interfaces

Slide 23 ISTORE-I: Summer 99? Intelligent disk –Portable PC Hardware: Pentium II, DRAM –Low Profile SCSI Disk (9 to 18 GB) –4 100-Mbit/s Ethernet links per Idisk –Placed inside Half-height canister –Monitor Processor/path to power off components? Intelligent Chassis –64 IDisks: 8 enclosures, 8 IDisks/enclosure »64 x 4 or 256 Ethernet ports –2 levels of Ethernet switches: 14 small, 2 large »Small: Mbit/s Gbit; Large: 25 1-Gbit –Enclosure sensing, UPS, redundant PS, fans,...

Slide 24 ISTORE Hardware Vision System-on-a-chip enables computer, memory, redundant network interfaces without increasing size of disk canister Disk enclosure includes (redundant) 1st-level switches as well as redundant power supplies, fans Rack includes 2nd-level switches, UPS

Slide 25 ISTORE-I Software Plan Modify Database (e.g., Predator) to send log to mirrored Idisk –Since 1 processor per disk, continuously replay the log on mirrored system Insert faults in original Idisk to get fail over Add monitoring, maintenance, fault insertion Run **ix OS –By running Linix binaries, can get multiple OS with same API: Linix, Free BSD Unix,... –Increase genetic base of OS software to reduce chances of simulatenous software bugs –Periodic reboot to “refresh” system

Slide 26 Benefits of ISTORE Decentralized processing (shared-nothing) –system can withstand partial failure Monitor their own “health,” test themselves, manage failures, collect application-specified performance data, and execute applications –fault insertion to test availability –provides the foundation for self-maintenance and self-tuning Plug & play, hot-swappable bricks ease configuration, scaling –hardware maybe specialized by selecting an collection of devices: DRAMs, WAN/LAN interfaces

Slide 27 Other (Potential) Benefits of ISTORE Scalability: add processing power, memory, network bandwidth as add disks Smaller footprint vs. traditional server/disk Less power –embedded processors vs. servers –spin down idle disks? For decision-support or web-service applications, potentially better performance than traditional servers

Slide 28 Related Work ISTORE adds several recent research efforts –Active Disks, NASD (UCSB, CMU) –Network service appliances (NetApp, Snap!, Qube,...) –High availability systems (Compaq/Tandem,...) –Adaptive systems (HP AutoRAID, M/S AutoAdmin, M/S Millennium) –Plug-and-play system construction (Jini, PC Plug&Play,...)

Slide 29 Interested in Participating? Project just getting formed Contact us if you’re interested: Thanks for support: DARPA Thanks for advice/inspiration: Dave Anderson (Seagate), Greg Papadopolous (Sun), Mike Ziegler (HP)

Slide 30 Backup Slides

Slide 31 ISTORE Cluster? disks / enclosure 12 enclosures / rack = disks/rack Cluster of PCs? 2 disks / PC 10 PCs /rack = 20 disks/rack Reliability? Ease of Repair? System Admin.? Cost only plus?

Slide 32 ISTORE and IRAM ISTORE relies on intelligent devices IRAM is an easy way to add intelligence to a device –embedded, low-power CPU meets size and power constraints –integrated DRAM reduces chip count –fast network interface (serial lines) meets connectivity needs Initial ISTORE prototype won’t use IRAM –will use collection of commodity components that approximate IRAM functionality, not size/power