Penn ESE680-002 Spring2007 -- DeHon 1 ESE680-002 (ESE534): Computer Organization Day 22: April 4, 2007 Interconnect 7: Time Multiplexed Interconnect.

Slides:



Advertisements
Similar presentations
Jaringan Komputer Lanjut Packet Switching Network.
Advertisements

What is Flow Control ? Flow Control determines how a network resources, such as channel bandwidth, buffer capacity and control state are allocated to packet.
Parallel System Performance CS 524 – High-Performance Computing.
Caltech CS184a Fall DeHon1 CS184a: Computer Architecture (Structures and Organization) Day17: November 20, 2000 Time Multiplexing.
Point-to-Point Network Switching Advanced Computer Networks.
High Performance Router Architectures for Network- based Computing By Dr. Timothy Mark Pinkston University of South California Computer Engineering Division.
Penn ESE Spring DeHon 1 ESE (ESE534): Computer Organization Day 21: April 2, 2007 Time Multiplexing.
Packet-Switched vs. Time-Multiplexed FPGA Overlay Networks Kapre et. al RC Reading Group – 3/29/2006 Presenter: Ilya Tabakh.
Shivkumar Kalyanaraman Rensselaer Polytechnic Institute 1 Review of Networking Concepts (Part 2) Shivkumar Kalyanaraman Rensselaer Polytechnic Institute.
Penn ESE534 Spring DeHon 1 ESE534: Computer Organization Day 24: April 21, 2010 Interconnect 6: Dynamically Switched Interconnect.
Caltech CS184a Fall DeHon1 CS184a: Computer Architecture (Structures and Organization) Day8: October 18, 2000 Computing Elements 1: LUTs.
Penn ESE Fall DeHon 1 ESE (ESE534): Computer Organization Day 19: March 26, 2007 Retime 1: Transformations.
Penn ESE Spring DeHon 1 ESE (ESE534): Computer Organization Day 13: February 26, 2007 Interconnect 1: Requirements.
Penn ESE Spring DeHon 1 ESE (ESE534): Computer Organization Day 11: February 14, 2007 Compute 1: LUTs.
Penn ESE535 Spring DeHon 1 ESE535: Electronic Design Automation Day 8: February 11, 2009 Dataflow.
Issues in System-Level Direct Networks Jason D. Bakos.
Parallel System Performance CS 524 – High-Performance Computing.
Penn ESE Spring DeHon 1 ESE (ESE534): Computer Organization Day 23: April 9, 2007 Control.
Point-to-Point Network Switching Computer Networks Computer Networks Spring 2012 Spring 2012.
Penn ESE Spring DeHon 1 FUTURE Timing seemed good However, only student to give feedback marked confusing (2 of 5 on clarity) and too fast.
Storage area network and System area network (SAN)
WAN Technologies.
Performance and Power Efficient On-Chip Communication Using Adaptive Virtual Point-to-Point Connections M. Modarressi, H. Sarbazi-Azad, and A. Tavakkol.
Switching, routing, and flow control in interconnection networks.
Lecture 1, 1Spring 2003, COM1337/3501Computer Communication Networks Rajmohan Rajaraman COM1337/3501 Textbook: Computer Networks: A Systems Approach, L.
Switching Techniques Student: Blidaru Catalina Elena.
Data Communications and Networking
Caltech CS184 Spring DeHon 1 CS184b: Computer Architecture (Abstractions and Optimizations) Day 20: May 23, 2003 Interconnect.
Communication Networks
Caltech CS184 Winter DeHon 1 CS184a: Computer Architecture (Structure and Organization) Day 13: February 4, 2005 Interconnect 1: Requirements.
On-Chip Networks and Testing
CALTECH cs184c Spring DeHon CS184c: Computer Architecture [Parallel and Multithreaded] Day 15: May 29, 2001 Interconnect.
High-Performance Networks for Dataflow Architectures Pravin Bhat Andrew Putnam.
Caltech CS184 Winter DeHon 1 CS184a: Computer Architecture (Structure and Organization) Day 15: February 12, 2003 Interconnect 5: Meshes.
ESE Spring DeHon 1 ESE534: Computer Organization Day 19: April 7, 2014 Interconnect 5: Meshes.
Networks for Distributed Systems n network types n Connection-oriented and connectionless communication n switching technologies l circuit l packet.
Chapter 2 – X.25, Frame Relay & ATM. Switched Network Stations are not connected together necessarily by a single link Stations are typically far apart.
High-Level Interconnect Architectures for FPGAs An investigation into network-based interconnect systems for existing and future FPGA architectures Nick.
High-Level Interconnect Architectures for FPGAs Nick Barrow-Williams.
Circuit & Packet Switching. ► Two ways of achieving the same goal. ► The transfer of data across networks. ► Both methods have advantages and disadvantages.
Switching breaks up large collision domains into smaller ones Collision domain is a network segment with two or more devices sharing the same Introduction.
ECE 4450:427/527 - Computer Networks Spring 2015 Dr. Nghi Tran Department of Electrical & Computer Engineering Lecture 2: Overview of Computer Network.
Computer Networks with Internet Technology William Stallings
1 Network Properties (Ownership, Service Paradigm, Measures of Performance)
CS 8501 Networks-on-Chip (NoCs) Lukasz Szafaryn 15 FEB 10.
Multiplexing FDM & TDM. Multiplexing When two communicating nodes are connected through a media, it generally happens that bandwidth of media is several.
COP 5611 Operating Systems Spring 2010 Dan C. Marinescu Office: HEC 439 B Office hours: M-Wd 2:00-3:00 PM.
Penn ESE534 Spring DeHon 1 ESE534: Computer Organization Day 25: April 28, 2014 Interconnect 7: Dynamically Switched Interconnect.
Sem1 - Module 8 Ethernet Switching. Shared media environments Shared media environment: –Occurs when multiple hosts have access to the same medium. –For.
BZUPAGES.COM Presentation On SWITCHING TECHNIQUE Presented To; Sir Taimoor Presented By; Beenish Jahangir 07_04 Uzma Noreen 07_08 Tayyaba Jahangir 07_33.
Final Chapter Packet-Switching and Circuit Switching 7.3. Statistical Multiplexing and Packet Switching: Datagrams and Virtual Circuits 4. 4 Time Division.
Penn ESE534 Spring DeHon 1 ESE534: Computer Organization Day 18: April 2, 2014 Interconnect 4: Switching.
Lecture # 03 Switching Course Instructor: Engr. Sana Ziafat.
Ch 8. Switching. Switch  Devices that interconnected with each other  Connecting all nodes (like mesh network) is not cost-effective  Some topology.
1 IEX8175 RF Electronics Avo Ots telekommunikatsiooni õppetool, TTÜ raadio- ja sidetehnika inst.
WAN Transmission Media
CSE 413: Computer Network Circuit Switching and Packet Switching Networks Md. Kamrul Hasan
1 Switching and Forwarding Sections Connecting More Than Two Hosts Multi-access link: Ethernet, wireless –Single physical link, shared by multiple.
Penn ESE534 Spring DeHon 1 ESE534 Computer Organization Day 9: February 13, 2012 Interconnect Introduction.
Lecture 17: Dynamic Reconfiguration I November 10, 2004 ECE 697F Reconfigurable Computing Lecture 17 Dynamic Reconfiguration I Acknowledgement: Andre DeHon.
Caltech CS184 Winter DeHon 1 CS184a: Computer Architecture (Structure and Organization) Day 11: January 31, 2005 Compute 1: LUTs.
Computer Communication & Networks Lecture # 03 Circuit Switching, Packet Switching Nadeem Majeed Choudhary
Univ. of TehranIntroduction to Computer Network1 An Introduction to Computer Networks University of Tehran Dept. of EE and Computer Engineering By: Dr.
WAN Technologies. 2 Large Spans and Wide Area Networks MAN networks: Have not been commercially successful.
ESE534: Computer Organization
ESE534: Computer Organization
ESE532: System-on-a-Chip Architecture
Azeddien M. Sllame, Amani Hasan Abdelkader
ESE534: Computer Organization
CS184a: Computer Architecture (Structure and Organization)
Presentation transcript:

Penn ESE Spring DeHon 1 ESE (ESE534): Computer Organization Day 22: April 4, 2007 Interconnect 7: Time Multiplexed Interconnect

Penn ESE Spring DeHon 2 Previously Multicontext computation Interconnect Topology Configured Interconnect –Lock down route between source and sink

Penn ESE Spring DeHon 3 Today Interconnect Style –Static Time Multiplexed –Dynamic Packet Switched Online/local vs. Offline/global

Penn ESE Spring DeHon 4 Motivation Holding a physical interconnect link for a logic path may be wasteful –Data Rate Logical Link < Physical Link E.g. –Time multiplexed logic Logic only appears one cycle –Multirate Some data may come at lower rates –Data dependent communication Only need to send fraction of time

Penn ESE Spring DeHon 5 Issues/Axes Throughput of Communication relative to data rate of media –Single point-to-point link consume media BW? –Can share links between multiple comm streams? –What is the sharing factor? Binding time/Predictability of Interconnect –Pre-fab –Before communication then use for long time –Cycle-by-cycle Network latency vs. persistence of communication –Comm link persistence

Penn ESE Spring DeHon 6 Axes Sharefactor (Media Rate/App. Rate) PredictabilityNet Latency Persistence

Penn ESE Spring DeHon 7 Hardwired Direct, fixed wire between two points E.g. Conventional gate-array, std. cell Efficient when: –know communication a priori fixed or limited function systems high load of fixed communication –often control in general-purpose systems –links carry high throughput traffic continually between fixed points

Penn ESE Spring DeHon 8 Configurable Offline, lock down persistent route. E.g. FPGAs Efficient when: –link carries high throughput traffic (loaded usefully near capacity) –traffic patterns change on timescale >> data transmission

Penn ESE Spring DeHon 9 Axes Sharefactor (Media Rate/App. Rate) PredictabilityNet Latency Persistence Config urable

Penn ESE Spring DeHon 10 Time-Switched Statically scheduled, wire/switch sharing E.g. TDMA, NuMesh, TSFPGA Efficient when: –thruput per channel < thruput capacity of wires and switches –traffic patterns change on timescale >> data transmission

Penn ESE Spring DeHon 11 Axes Sharefactor (Media Rate/App. Rate) Predictability Time Mux

Penn ESE Spring DeHon 12 Self-Route, Circuit-Switched Dynamic arbitration/allocation, lock down routes E.g. METRO/RN1, old telephone net Efficient when: –instantaneous communication bandwidth is high (consume channel) –lifetime of comm. > delay through network –communication pattern unpredictable –rapid connection setup important

Penn ESE Spring DeHon 13 Axes Net Latency Persistence Circuit Switch Phone; Videoconf; Cable Sharefactor (Media Rate/App. Rate) Predictability Circuit Switch

Penn ESE Spring DeHon 14 Self-Route, Store-and- Forward, Packet Switched Dynamic arbitration, packetized data Get entire packet before sending to next node E.g. nCube, early Internet routers Efficient when: –lifetime of comm < delay through net –communication pattern unpredictable –can provide buffer/consumption guarantees –packets small

Penn ESE Spring DeHon 15 Store-and-Forward

Penn ESE Spring DeHon 16 Self-Route, Virtual Cut Through Dynamic arbitration, packetized data Start forwarding to next node as soon as have header Don’t pay full latency of storing packet Keep space to buffer entire packet if necessary Efficient when: –lifetime of comm < delay through net –communication pattern unpredictable –can provide buffer/consumption guarantees –packets small

Penn ESE Spring DeHon 17 Virtual Cut Through Three words from same packet

Penn ESE Spring DeHon 18 Self-Route, Wormhole Packet-Switched Dynamic arbitration, packetized data E.g. Caltech MRC, Modern Internet Routers Efficient when: –lifetime of comm < delay through net –communication pattern unpredictable –can provide buffer/consumption guarantees –message > buffer length allow variable (? Long) sized messages

Penn ESE Spring DeHon 19 Wormhole Single Packet spread through net when not stalled

Penn ESE Spring DeHon 20 Wormhole Single Packet spread through net when stalled.

Penn ESE Spring DeHon 21 Axes Sharefactor (Media Rate/App. Rate) PredictabilityNet Latency Persistence Packet Switch Circuit Switch Phone; Videoconf; Cable IP Packet SMS message Config urable Packet Switch Time Mux Circuit Switch

Penn ESE Spring DeHon 22 PS vs. TM Following from Kapre et al. / FCCM 2006

Penn ESE Spring DeHon 23 Time-Multiplexed (TM) Message paths are computed offline prior to execution –Based on a workload specified already +Quality of route potentially better due to global view during routing  Need to store routing decisions in memory in hardware + Faster, simpler switches –Need to spend time computing routes offline –Need to known traffic pattern beforehand

Penn ESE Spring DeHon 24 Intuitive Tradeoff (TM) Benefit of Time-Multiplexing? –Minimum end-to-end latency –No added decision latency at runtime –Offline route  high quality route  use wires efficiently Cost of Time-Multiplexing? –Route task must be static Cannot exploit low activity –Need memory bit per switch per time step Lots of memory if need large number of time steps…

Penn ESE Spring DeHon 25 Packet-Switched (PS) Messages are dynamically routed on the network –Based on address in the packet –Complex switches (queues and address decode logic) –Known issues with deadlock/livelock/load- distribution (complex routing algorithms, poor route quality) +No prior knowledge of routes required,

Penn ESE Spring DeHon 26 Intuitive Tradeoff (PS) Benefit of Packet Switching? –No area proportional to time steps –Route only active connections –Avoids slow, off-line routing Cost of Packet Switching? –Online decision making Maybe won’t use wires as well Potentially slower routing? –Slower clock or more clocks across net –Data will be blocked in network Adds latency Requires packet queues (area)

Penn ESE Spring DeHon 27 Local Online vs. Global Offline

Penn ESE Spring DeHon 28 Butterfly Fat Trees (BFTs) Familiar from Day 15, 16 Similar phenomena with other topologies Directional version

Penn ESE Spring DeHon 29 BFT Terminology T = t-switch π = pi-switch p = Rent Parameter (defines sequence of T and π switches) c = PE IO Ports (parallel BFT planes)

Penn ESE Spring DeHon 30 PS Hardware Primitives T-Switch π -Switch

Penn ESE Spring DeHon 31 TM Hardware Primitives

Penn ESE Spring DeHon 32 Complete Network

Penn ESE Spring DeHon 33 Analysis PS v/s TM for same topologies –Quantify inherent benefit of TM PS v/s TM for same area –Understand area tradeoffs (PEs v/s Interconnect) PS v/s TM for dynamic traffic –PS routes limited traffic, TM has to route all traffic

Penn ESE Spring DeHon 34 Iso-PEs

Penn ESE Spring DeHon 35 Iso-PEs PS vs. TM ratio at same PE counts –Small number of PEs little difference Dominated by serialization (self-messages) Not stressing the network –Larger PE counts TM ~60% better TM uses global congestion knowledge while scheduling

Penn ESE Spring DeHon 36 Area Effects Based on FPGA overlay model i.e. build PS or TM on top of FPGA

Penn ESE Spring DeHon 37 Area Analysis Evaluate PS and TM for multiple BFTs –Tradeoff Logic Area for Interconnect –Fixed Area of 130K slices p=0, BFT => 128 PS PEs => 1476 cycles p=0.5, BFT => 64 PS PEs => 943 cycles Extract best topologies for PS and TM at each area point –BFT of different p best at different area points Compare performance achieved at these bests at each area point

Penn ESE Spring DeHon 38 PS Iso Area: Topology Selection

Penn ESE Spring DeHon 39 TM Iso Area

Penn ESE Spring DeHon 40 Iso Area

Penn ESE Spring DeHon 41 Iso Area

Penn ESE Spring DeHon 42 Iso Area Iso-PEs = TM 1~2x better With Area –PS 2x better at small areas –TM 4-5x better at large areas –PS catches up at the end Iso-Area = TM ~5x better

Penn ESE Spring DeHon 43 Activity Factors Activity = Fraction of traffic to be routed TM needs to route all PS can route fraction Variable activity queries in ConceptNet –Simple queries ~1% edges –Complex queries ~40% edges

Penn ESE Spring DeHon 44 Activity Factors

Penn ESE Spring DeHon 45 Crossover could be less

Penn ESE Spring DeHon 46 Lessons Latency –PS could achieve same clock rate –But took more cycles –Didn’t matter for this workload Quality of Route –PS could be 60% worse Area –PS larger, despite all the TM instrs –Big factor –May be “technology” dependent Need to review for custom model Will be smaller relative factor for custom

Penn ESE Spring DeHon 47 Admin Homework 8 Final Exercise

Penn ESE Spring DeHon 48 Big Ideas [MSB Ideas] Different interconnect switching styles based on –Relative throughput, predictability, persistence, latency Low throughput  time share interconnect High predictability  –Efficiency for offline/global solutions Low predictability  dynamic