CS252 Graduate Computer Architecture Lecture 15 Multiprocessor Networks March 12 th, 2012 John Kubiatowicz Electrical Engineering and Computer Sciences.

Slides:



Advertisements
Similar presentations
Comparison Of Network On Chip Topologies Ahmet Salih BÜYÜKKAYHAN Fall.
Advertisements

Presentation of Designing Efficient Irregular Networks for Heterogeneous Systems-on-Chip by Christian Neeb and Norbert Wehn and Workload Driven Synthesis.
Flattened Butterfly: A Cost-Efficient Topology for High-Radix Networks ______________________________ John Kim, William J. Dally &Dennis Abts Presented.
1 Lecture 12: Interconnection Networks Topics: dimension/arity, routing, deadlock, flow control.
1 Interconnection Networks Direct Indirect Shared Memory Distributed Memory (Message passing)
1 Lecture 23: Interconnection Networks Topics: communication latency, centralized and decentralized switches (Appendix E)
CS252 Graduate Computer Architecture Lecture 14 Multiprocessor Networks March 10 th, 2010 John Kubiatowicz Electrical Engineering and Computer Sciences.
EECC756 - Shaaban #1 lec # 9 Spring Network Definitions A network is a graph V = {switches and nodes} connected by communication channels.
CS 258 Parallel Computer Architecture Lecture 4 Network Topology and Routing February 4, 2008 Prof John D. Kubiatowicz
CS252 Graduate Computer Architecture Lecture 21 Multiprocessor Networks (con’t) John Kubiatowicz Electrical Engineering and Computer Sciences University.
CS 258 Parallel Computer Architecture Lecture 5 Routing February 6, 2008 Prof John D. Kubiatowicz
Generic Multiprocessor Architecture
CS252 Graduate Computer Architecture Lecture 15 Multiprocessor Networks (con’t) March 15 th, 2010 John Kubiatowicz Electrical Engineering and Computer.
ECE669 L25: Final Exam Review May 6, 2004 ECE 669 Parallel Computer Architecture Lecture 25 Final Exam Review.
1 Lecture 24: Interconnection Networks Topics: communication latency, centralized and decentralized switches (Sections 8.1 – 8.5)
CS 258 Parallel Computer Architecture Lecture 3 Introduction to Scalable Interconnection Network Design January 30, 2008 Prof John D. Kubiatowicz
Issues in System-Level Direct Networks Jason D. Bakos.
EECS 570: Fall rev1 1 Chapter 10: Scalable Interconnection Networks.
EECC756 - Shaaban #1 lec # 9 Spring Generic Multiprocessor Architecture Generic Multiprocessor Architecture Node: processor(s), memory system,
Interconnection Network Topology Design Trade-offs
1 Lecture 24: Interconnection Networks Topics: topologies, routing, deadlocks, flow control.
1 Lecture 25: Interconnection Networks Topics: communication latency, centralized and decentralized switches, routing, deadlocks (Appendix E) Review session,
CS252/Patterson Lec /28/01 CS162 Computer Architecture Lecture 16: Multiprocessor 2: Directory Protocol, Interconnection Networks.
1 Static Interconnection Networks CEG 4131 Computer Architecture III Miodrag Bolic.
EECC756 - Shaaban #1 lec # 10 Spring Generic Multiprocessor Architecture Generic Multiprocessor Architecture Node: processor(s), memory system,
ECE669 L16: Interconnection Topology March 30, 2004 ECE 669 Parallel Computer Architecture Lecture 16 Interconnection Topology.
CS252 Graduate Computer Architecture Lecture 14 Multiprocessor Networks March 9 th, 2011 John Kubiatowicz Electrical Engineering and Computer Sciences.
John Kubiatowicz Electrical Engineering and Computer Sciences
Storage area network and System area network (SAN)
Switching, routing, and flow control in interconnection networks.
Interconnect Network Topologies
CS252 Graduate Computer Architecture Lecture 15 Multiprocessor Networks March 14 th, 2011 John Kubiatowicz Electrical Engineering and Computer Sciences.
Interconnection Networks. Applications of Interconnection Nets Interconnection networks are used everywhere! ◦ Supercomputers – connecting the processors.
1 The Turn Model for Adaptive Routing. 2 Summary Introduction to Direct Networks. Deadlocks in Wormhole Routing. System Model. Partially Adaptive Routing.
Interconnect Networks
On-Chip Networks and Testing
CS252 Graduate Computer Architecture Lecture 14 Multiprocessor Networks March 7 th, 2012 John Kubiatowicz Electrical Engineering and Computer Sciences.
Winter 2006 ENGR 9861 – High Performance Computer Architecture March 2006 Interconnection Networks.
1 Lecture 7: Interconnection Network Part I: Basic Definitions Part II: Message Passing Multicomputers.
1 Scalable Interconnection Networks. 2 Scalable, High Performance Network At Core of Parallel Computer Architecture Requirements and trade-offs at many.
Dynamic Interconnect Lecture 5. COEN Multistage Network--Omega Network Motivation: simulate crossbar network but with fewer links Components: –N.
Multiprocessor Interconnection Networks Todd C. Mowry CS 740 November 3, 2000 Topics Network design issues Network Topology.
ECE669 L21: Routing April 15, 2004 ECE 669 Parallel Computer Architecture Lecture 21 Routing.
Anshul Kumar, CSE IITD CSL718 : Multiprocessors Interconnection Mechanisms Performance Models 20 th April, 2006.
InterConnection Network Topologies to Minimize graph diameter: Low Diameter Regular graphs and Physical Wire Length Constrained networks Nilesh Choudhury.
Anshul Kumar, CSE IITD ECE729 : Advanced Computer Architecture Lecture 27, 28: Interconnection Mechanisms In Multiprocessors 29 th, 31 st March, 2010.
Interconnect Networks Basics. Generic parallel/distributed system architecture On-chip interconnects (manycore processor) Off-chip interconnects (clusters.
Super computers Parallel Processing
Networks: Routing, Deadlock, Flow Control, Switch Design, Case Studies Alvin R. Lebeck CPS 220.
Topology How the components are connected. Properties Diameter Nodal degree Bisection bandwidth A good topology: small diameter, small nodal degree, large.
1 Lecture 24: Interconnection Networks Topics: communication latency, centralized and decentralized switches, routing, deadlocks (Appendix F)
1 Lecture 14: Interconnection Networks Topics: dimension vs. arity, deadlock.
Interconnect Networks
Lecture 23: Interconnection Networks
Multiprocessor Interconnection Networks Todd C
Prof John D. Kubiatowicz
Static and Dynamic Networks
Interconnection Network Routing, Topology Design Trade-offs
John Kubiatowicz Electrical Engineering and Computer Sciences
Interconnection Network Design Contd.
Introduction to Scalable Interconnection Network Design
Switching, routing, and flow control in interconnection networks
Lecture 14: Interconnection Networks
Interconnection Network Design Lecture 14
Introduction to Scalable Interconnection Networks
Interconnection Network Design
Interconnection Networks Contd.
CS 6290 Many-core & Interconnect
Networks: Routing and Design
Presentation transcript:

CS252 Graduate Computer Architecture Lecture 15 Multiprocessor Networks March 12 th, 2012 John Kubiatowicz Electrical Engineering and Computer Sciences University of California, Berkeley

3/12/ cs252-S12, Lecture15 What characterizes a network? Topology (what) –physical interconnection structure of the network graph –direct: node connected to every switch –indirect: nodes connected to specific subset of switches Routing Algorithm(which) –restricts the set of paths that msgs may follow –many algorithms with different properties »deadlock avoidance? Switching Strategy(how) –how data in a msg traverses a route –circuit switching vs. packet switching Flow Control Mechanism(when) –when a msg or portions of it traverse a route –what happens when traffic is encountered?

3/12/ cs252-S12, Lecture15 Formalism network is a graph V = {switches and nodes} connected by communication channels C  V  V Channel has width w and signaling rate f =  –channel bandwidth b = wf –phit (physical unit) data transferred per cycle –flit - basic unit of flow-control Number of input (output) channels is switch degree Sequence of switches and links followed by a message is a route Think streets and intersections

3/12/ cs252-S12, Lecture15 Links and Channels transmitter converts stream of digital symbols into signal that is driven down the link receiver converts it back –tran/rcv share physical protocol trans + link + rcv form Channel for digital info flow between switches link-level protocol segments stream of symbols into larger units: packets or messages (framing) node-level protocol embeds commands for dest communication assist within packet Transmitter...ABC123 => Receiver...QR67 =>

3/12/ cs252-S12, Lecture15 Clock Synchronization? Receiver must be synchronized to transmitter –To know when to latch data Fully Synchronous –Same clock and phase: Isochronous –Same clock, different phase: Mesochronous »High-speed serial links work this way »Use of encoding (8B/10B) to ensure sufficient high-frequency component for clock recovery Fully Asynchronous –No clock: Request/Ack signals –Different clock: Need some sort of clock recovery? Data Req Ack Transmitter Asserts Data t0 t1 t2 t3 t4 t5

3/12/ cs252-S12, Lecture15 Topological Properties Routing Distance - number of links on route Diameter - maximum routing distance Average Distance A network is partitioned by a set of links if their removal disconnects the graph

3/12/ cs252-S12, Lecture15 Interconnection Topologies Class of networks scaling with N Logical Properties: –distance, degree Physical properties –length, width Fully connected network –diameter = 1 –degree = N –cost? »bus => O(N), but BW is O(1) - actually worse »crossbar => O(N 2 ) for BW O(N) VLSI technology determines switch degree

3/12/ cs252-S12, Lecture15 Example: Linear Arrays and Rings Linear Array –Diameter? –Average Distance? –Bisection bandwidth? –Route A -> B given by relative address R = B-A Torus? Examples: FDDI, SCI, FiberChannel Arbitrated Loop, KSR1

3/12/ cs252-S12, Lecture15 Example: Multidimensional Meshes and Tori n-dimensional array –N = k n-1 X...X k O nodes –described by n-vector of coordinates (i n-1,..., i O ) n-dimensional k-ary mesh: N = k n –k = n  N –described by n-vector of radix k coordinate n-dimensional k-ary torus (or k-ary n-cube)? 2D Grid 3D Cube 2D Torus

3/12/ cs252-S12, Lecture15 On Chip: Embeddings in two dimensions Embed multiple logical dimension in one physical dimension using long wires When embedding higher-dimension in lower one, either some wires longer than others, or all wires long 6 x 3 x 2

3/12/ cs252-S12, Lecture15 Trees Diameter and ave distance logarithmic –k-ary tree, height n = log k N –address specified n-vector of radix k coordinates describing path down from root Fixed degree Route up to common ancestor and down –R = B xor A –let i be position of most significant 1 in R, route up i+1 levels –down in direction given by low i+1 bits of B H-tree space is O(N) with O(  N) long wires Bisection BW?

3/12/ cs252-S12, Lecture15 Fat-Trees Fatter links (really more of them) as you go up, so bisection BW scales with N

3/12/ cs252-S12, Lecture15 Butterflies Tree with lots of roots! N log N (actually N/2 x logN) Exactly one route from any source to any dest R = A xor B, at level i use ‘straight’ edge if r i =0, otherwise cross edge Bisection N/2 vs N (n-1)/n (for n-cube) 16 node butterfly building block

3/12/ cs252-S12, Lecture15 k-ary n-cubes vs k-ary n-flies degree n vs degree k N switches vs N log N switches diminishing BW per node vs constant requires localityvs little benefit to locality Can you route all permutations?

3/12/ cs252-S12, Lecture15 Benes network and Fat Tree Back-to-back butterfly can route all permutations What if you just pick a random mid point?

3/12/ cs252-S12, Lecture15 Hypercubes Also called binary n-cubes. # of nodes = N = 2 n. O(logN) Hops Good bisection BW Complexity –Out degree is n = logN correct dimensions in order –with random comm. 2 ports per processor 0-D1-D2-D3-D 4-D 5-D !

3/12/ cs252-S12, Lecture15 Some Properties Routing –relative distance: R = (b n-1 - a n-1,..., b 0 - a 0 ) –traverse ri = b i - a i hops in each dimension –dimension-order routing? Adaptive routing? Average DistanceWire Length? –n x 2k/3 for mesh –nk/2 for cube Degree? Bisection bandwidth?Partitioning? –k n-1 bidirectional links Physical layout? –2D in O(N) spaceShort wires –higher dimension?

3/12/ cs252-S12, Lecture15 The Routing problem: Local decisions Routing at each hop: Pick next output port!

3/12/ cs252-S12, Lecture15 How do you build a crossbar?

3/12/ cs252-S12, Lecture15 Input buffered switch Independent routing logic per input –FSM Scheduler logic arbitrates each output –priority, FIFO, random Head-of-line blocking problem –Message at head of queue blocks messages behind it

3/12/ cs252-S12, Lecture15 Output Buffered Switch How would you build a shared pool?

3/12/ cs252-S12, Lecture15 Properties of Routing Algorithms Routing algorithm: –R: N x N -> C, which at each switch maps the destination node n d to the next channel on the route –which of the possible paths are used as routes? –how is the next hop determined? »arithmetic »source-based port select »table driven »general computation Deterministic –route determined by (source, dest), not intermediate state (i.e. traffic) Adaptive –route influenced by traffic along the way Minimal –only selects shortest paths Deadlock free –no traffic pattern can lead to a situation where packets are deadlocked and never move forward

3/12/ cs252-S12, Lecture15 Example: Simple Routing Mechanism need to select output port for each input packet –in a few cycles Simple arithmetic in regular topologies –ex:  x,  y routing in a grid »west (-x)  x < 0 »east (+x)  x > 0 »south (-y)  x = 0,  y < 0 »north (+y)  x = 0,  y > 0 »processor  x = 0,  y = 0 Reduce relative address of each dimension in order –Dimension-order routing in k-ary d-cubes –e-cube routing in n-cube

3/12/ cs252-S12, Lecture15 Communication Performance Typical Packet includes data + encapsulation bytes –Unfragmented packet size S = S data +S encapsulation Routing Time: –Time(S) s-d = overhead + routing delay + channel occupancy + contention delay –Channel occupancy = S/b = (S data + S encapsulation )/b –Routing delay in cycles (  »Time to get head of packet to next hop –Contention?

3/12/ cs252-S12, Lecture15 Store&Forward vs Cut-Through Routing Time:h(S/b +  ) vsS/b + h  OR(cycles): h(S/w +  ) vsS/w + h  what if message is fragmented? wormhole vs virtual cut-through

3/12/ cs252-S12, Lecture15 Contention Two packets trying to use the same link at same time –limited buffering –drop? Most parallel mach. networks block in place –link-level flow control –tree saturation Closed system - offered load depends on delivered –Source Squelching

3/12/ cs252-S12, Lecture15 Bandwidth What affects local bandwidth? –packet density:b x S data /S –routing delay:b x S data /(S + w  ) –contention »endpoints »within the network Aggregate bandwidth –bisection bandwidth »sum of bandwidth of smallest set of links that partition the network –total bandwidth of all the channels: Cb –suppose N hosts issue packet every M cycles with ave dist »each msg occupies h channels for l = S/w cycles each »C/N channels available per node »link utilization for store-and-forward:  = (hl/M channel cycles/node)/(C/N) = Nhl/MC < 1! »link utilization for wormhole routing?

3/12/ cs252-S12, Lecture15 Saturation

3/12/ cs252-S12, Lecture15 How Many Dimensions? n = 2 or n = 3 –Short wires, easy to build –Many hops, low bisection bandwidth –Requires traffic locality n >= 4 –Harder to build, more wires, longer average length –Fewer hops, better bisection bandwidth –Can handle non-local traffic k-ary n-cubes provide a consistent framework for comparison –N = k n –scale dimension (n) or nodes per dimension (k) –assume cut-through

3/12/ cs252-S12, Lecture15 Traditional Scaling: Latency scaling with N Assumes equal channel width –independent of node count or dimension –dominated by average distance

3/12/ cs252-S12, Lecture15 Average Distance but, equal channel width is not equal cost! Higher dimension => more channels ave dist = n(k-1)/2

3/12/ cs252-S12, Lecture15 Dally Paper: In the 3D world For N nodes, bisection area is O(N 2/3 ) For large N, bisection bandwidth is limited to O(N 2/3 ) –Bill Dally, IEEE TPDS, [Dal90a] –For fixed bisection bandwidth, low-dimensional k-ary n- cubes are better (otherwise higher is better) –i.e., a few short fat wires are better than many long thin wires –What about many long fat wires?

3/12/ cs252-S12, Lecture15 Dally paper (con’t) Equal Bisection,W=1 for hypercube  W= ½k Three wire models: –Constant delay, independent of length –Logarithmic delay with length (exponential driver tree) –Linear delay (speed of light/optimal repeaters) Logarithmic Delay Linear Delay

3/12/ cs252-S12, Lecture15 Equal cost in k-ary n-cubes Equal number of nodes? Equal number of pins/wires? Equal bisection bandwidth? Equal area? Equal wire length? What do we know? switch degree: ndiameter = n(k-1) total links = Nn pins per node = 2wn bisection = k n-1 = N/k links in each directions 2Nw/k wires cross the middle

3/12/ cs252-S12, Lecture15 Latency for Equal Width Channels total links(N) = Nn

3/12/ cs252-S12, Lecture15 Latency with Equal Pin Count Baseline n=2, has w = 32 (128 wires per node) fix 2nw pins => w(n) = 64/n distance up with n, but channel time down

3/12/ cs252-S12, Lecture15 Latency with Equal Bisection Width N-node hypercube has N bisection links 2d torus has 2N 1/2 Fixed bisection  w(n) = N 1/n / 2 = k/2 1 M nodes, n=2 has w=512!

3/12/ cs252-S12, Lecture15 Larger Routing Delay (w/ equal pin) Dally’s conclusions strongly influenced by assumption of small routing delay –Here, Routing delay  =20

3/12/ cs252-S12, Lecture15 Saturation Fatter links shorten queuing delays

3/12/ cs252-S12, Lecture15 Discuss of paper: Virtual Channel Flow Control Basic Idea: Use of virtual channels to reduce contention –Provided a model of k-ary, n-flies –Also provided simulation Tradeoff: Better to split buffers into virtual channels –Example (constant total storage for 2-ary 8-fly):

3/12/ cs252-S12, Lecture15 When are virtual channels allocated? Two separate processes: –Virtual channel allocation –Switch/connection allocation Virtual Channel Allocation –Choose route and free output virtual channel –Really means: Source of link tracks channels at destination Switch Allocation –For incoming virtual channel, negotiate switch on outgoing pin Hardware efficient design For crossbar

3/12/ cs252-S12, Lecture15 Problem: Low-dimensional networks have high k –Consequence: may have to travel many hops in single dimension –Routing latency can dominate long-distance traffic patterns Solution: Provide one or more “express” links –Like express trains, express elevators, etc »Delay linear with distance, lower constant »Closer to “speed of light” in medium »Lower power, since no router cost –“Express Cubes: Improving performance of k-ary n-cube interconnection networks,” Bill Dally 1991 Another Idea: route with pass transistors through links Reducing routing delay: Express Cubes

3/12/ cs252-S12, Lecture15 Summary Network Topologies: Fair metrics of comparison –Equal cost: area, bisection bandwidth, etc Routing Algorithms restrict set of routes within the topology –simple mechanism selects turn at each hop –arithmetic, selection, lookup Virtual Channels –Adds complexity to router –Can be used for performance –Can be used for deadlock avoidance TopologyDegreeDiameterAve DistBisectionD (D P=1024 1D Array2N-1N / 31huge 1D Ring2N/2N/42 2D Mesh42 (N 1/2 - 1)2/3 N 1/2 N 1/2 63 (21) 2D Torus4N 1/2 1/2 N 1/2 2N 1/2 32 (16) k-ary n-cube2nnk/2nk/4nk/415 Hypercuben =log Nnn/2N/210 (5)