Download presentation
Presentation is loading. Please wait.
1
Clustering & Fibre Channel Next wave in PC Computing
2
2 SP041098.ppt Trends in Clustering Today clusters are a niche Unix market. But, Microsoft will bring clusters to the masses Microsoft has announced NT clusters SCO has announced UnixWare clusters Sun has announced Solaris / Intel clusters Novell has announced Wolf Mountain clusters In 1998, 2M Intel servers will ship 100K in clusters In 2001, 3M Intel servers will ship 1M in clusters (IDC’s forecast) Clusters will be a huge market and RAID is essential to clusters Fibre Channel and storage area networks are ideal for use in clusters.
3
3 SP041098.ppt Why is the IT Market Moving To Clustering Growth In On-Line Applications E-Commence On-Line Transaction Processing Web Servers Real Time Manufacturing Server Must Be On Line 24 Hours a Day, 7 Days A Week, 365 Days a Year New Levels Of Fault Tolerance Performance Must Scale “Without Our Data, Our Business Is Dead”
4
4 SP041098.ppt What Are Clusters? What Are Clusters? Group of independent servers which Function as a single system Appear to users as a single system And are managed as a single system Clusters are “virtual servers”
5
5 SP041098.ppt Clusters Improve System Availability Clusters Enable Application Scaling Clusters Simplify System Management Clusters Are A Superior Server Solution Why Are Clusters Important?
6
6 SP041098.ppt When a networked server fails, the service it provided is down. When a clustered server fails, the service it provided “failover” and downtime is avoided. Mail Server Internet Server Networked Servers Clustered Servers Mail & Internet Clusters Improve System Availability
7
7 SP041098.ppt Clusters Enable Application Scaling With networked SMP servers, application scaling is limited to a single server With clusters, applications scale across multiple SMP servers (typically up to 16 servers) E-MailVoice mail Oracle Voice mail Oracle E-Mail
8
8 SP041098.ppt Clusters Simplify System Management Clusters present a Single System Image; the cluster looks like a single server to management applications Hence, clusters reduce system management costs Three Management Domains One Management Domain
9
9 SP041098.ppt An Analogy to RAID RAID Makes Disks Fault Tolerant Clusters make servers fault tolerant RAID Increases I/O Performance Clusters increase compute performance RAID (GAM) Makes Disks Easier to Manage Clusters make servers easier to manage RAID
10
10 SP041098.ppt Two Flavors of Clusters High Availability Clusters Microsoft's Wolfpack 1 Compaq’s Recovery Server Load Balancing Clusters (a.k.a. Parallel Application Clusters) Microsoft’s Wolfpack 2 Digital’s VAXClusters Note: Load balancing clusters are a superset of high availability clusters.
11
11 SP041098.ppt Available Clustering Software Failover Clusters Microsoft Clustered Servers (Enterprise Server 4.0) Compaq’s Recovery Server NCR Lifekeeper Digital NT Clusters UnixWare 7 Reliant NSI DoubleTake failover in Wide Area Networks Novell’s Vinca Corp. Standby Server Load Balancing Clusters Microsoft’s Clustered Servers (NT 5.0) Digital’s VAXClusters IBM’s HACMP Oracle’s Parallel Server Novell’s Orion
12
12 SP041098.ppt Failover Example Two node clusters (node = server) During normal operations, both servers do useful work Failover When a node fails, applications “failover” to the surviving node and it assumes the workload of both nodes MailWeb Mail & Web
13
13 SP041098.ppt Load Balancing Example Failover (cont’d) Multi-node clusters (two or more nodes) Load balancing clusters typically runs a single application, e.g. database, distributed across all nodes Cluster capacity is increased by adding nodes (but like SMP servers, scaling is less than linear) 3,000 TPM3,600 TPM
14
14 SP041098.ppt Load Balancing Example Load Balancing Example (cont’d) Cluster rebalance the workload when a node dies If different apps are running on each server, they failover to the least busy server or as directed by preset failover policies
15
15 SP041098.ppt Sharing Storage In Clusters #1. “Shared Nothing” Model Microsoft’s Wolfpack Cluster #2. “Shared Disk” Model VAXClusters Two Models For Storage
16
16 SP041098.ppt What is Fibre Channel? Fibre channel is a high performance multiple protocol data transfer technology. Fibre channels primary task is to transport data extremely fast with the least possible delay. Fibre channel is a serial interconnect standard that allows servers, storage devices, and workstation users to share large amounts of data quickly. Fibre channel gives networks superior data transfer speeds, flexible topology, and flexible upper-level protocols. Fibre channel easily handles both networking and peripheral I/O communication over a single channel.
17
17 SP041098.ppt Key Features of Fibre Channel Data Transfer Rates up to 100 MB/sec per connection Three Fibre channel connection schemes Point-to-Point: Provides a single connection between two servers or a server and its RAID storage. Switched Fabric: Using a Fibre channel switch this connection allows each server or RAID storage to be connected point-to-point to a switch. This method allows for the construction of massive data storage and server networks. Loop: Connects up to 126 servers, RAID systems, or other storage devices in a loop topology.
18
18 SP041098.ppt Key Features of Fibre Channel Long Distance Connections: Up to 10 kilometers High reliability connections providing high data integrity Low Overhead Multiple media types support: Copper: lower cost not as reliable as fibre optic cable distances up to 20 meters between nodes Fibre optic: higher cost more reliable than copper distance up to 10,000 meters between nodes Allows for creation of Storage Area Networks
19
19 SP041098.ppt Benfits of Fibre Channel Runs multiple protocols over a single network or loop Ideal solution for clustering of servers Enables multiple servers to share storage and RAID systems Scales from small numbers of peripherals attached short distances apart, to large numbers attached many kilometers apart.
20
20 SP041098.ppt Benfits of Fibre Channel Benfits of Fibre Channel (cont’d) Delivers speeds that are 2.5 to 250 times faster than existing communication and I/O interfaces Overcomes today's network performance limitations regarding error detection and recovery Provided low-cost and reliable performance for distance connections Offers flexible protocols and topologies to best leverage existing technology investments
21
21 SP041098.ppt Cluster Interconnect This is about how servers are tied together and how disks are physically connected to the cluster.
22
22 SP041098.ppt Clustered servers always have a client network interconnect, typically Ethernet, to talk to users, and at least on cluster interconnect to talk to other nodes and to disks. Cluster Interconnect Client Network HBA Cluster Interconnect HBA RAID
23
23 SP041098.ppt Or they can have Two Cluster Interconnects One for nodes to talk to each other: Heartbeat Interconnect typically Ethernet And one for nodes to talk to disks Shared Disk Interconnect typically SCSI or Fibre channel Cluster Interconnect NIC Cluster Interconnect RAID Shared Disk Interconnect HBANICHBA
24
24 SP041098.ppt Each node has: Ethernet NIC -- Heartbeat Private system disks (HBA) PCI-based RAID controller - SCSI or Fibre, such as Mylex’s eXtremeRAID™ Nodes share access to data disks but do not share data NT Cluster with Host-Based RAID Array Shared Disk Interconnect “Heartbeat” Interconnect HBA NIC RAID
25
25 SP041098.ppt NT Cluster w/ SCSI External RAID Array Each node has: Ethernet NIC -- Heartbeat Multi-channel HBA’s connect boot disk and external array Shared external RAID controller on the SCSI Shared Disk Interconnect Shared Disk Interconnect “Heartbeat” Interconnect HBA NIC RAID
26
26 SP041098.ppt NT Cluster w/ Fibre External RAID Array Fibre Channel Host to Ultra2 SCSI LVD RAID system Fibre Channel Host to Fibre Channel RAID system Shared Disk Interconnect “Heartbeat” Interconnect HBA NIC RAID
27
27 SP041098.ppt Internal vs External RAID in Clustering Internal RAID Lower cost solution Higher performance in read-intensive applications Proven TPC-C performance enhances cluster performance External RAID Higher performance in write-intensive application Write-back cache is turned-off in PCI-RAID controllers Higher connectivity Attach more disk drives Greater footprint flexibility Until PCI-RAID implements fibre
28
28 SP041098.ppt Active / Active Fibre Channel Host - Ultra2 LVD RAID Controller in a Clustered Loop HBA SX Single FC Array Interconnect FC HBA
29
29 SP041098.ppt SF (or FL) Active / Active Duplex HBA Dual FC Array Interconnect FC HBA FC Disk Interconnect FC HBA SF Fibre Channel Host - Ultra2 LVD RAID Controller in a Clustered Environment
30
30 SP041098.ppt Fibre Host HBA Single FC Array Interconnect FC HBA FF Active / Active Duplex
31
31 SP041098.ppt FC - FC in Cluster HBA Dual FC Array Interconnect FC HBA FF
32
32 SP041098.ppt Mylex - The RAID Clustering Experts eXtremeRAID™ 1100 AcceleRAID™ 250 DAC SF DAC FL DAC FF
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.