Serial Network SDRAM ENEE 759H Spring 2003. Introduction SDRAM system drawbacks  No parallelism for memory accesses  Multitude of pins for address/command/data.

Slides:



Advertisements
Similar presentations
Chapter 7: System Buses Dr Mohamed Menacer Taibah University
Advertisements

3D Graphics Content Over OCP Martti Venell Sr. Verification Engineer Bitboys.
Main MemoryCS510 Computer ArchitecturesLecture Lecture 15 Main Memory.
Chapter 5 Internal Memory
Computer Organization and Architecture
Multi-Level Caches Vittorio Zaccaria. Preview What you have seen: Data organization, Associativity, Cache size Policies -- how to manage the data once.
AMD OPTERON ARCHITECTURE Omar Aragon Abdel Salam Sayyad This presentation is missing the references used.
Anshul Kumar, CSE IITD CSL718 : Main Memory 6th Mar, 2006.
COEN 180 DRAM. Dynamic Random Access Memory Dynamic: Periodically refresh information in a bit cell. Else it is lost. Small footprint: transistor + capacitor.
Main Mem.. CSE 471 Autumn 011 Main Memory The last level in the cache – main memory hierarchy is the main memory made of DRAM chips DRAM parameters (memory.
CS.305 Computer Architecture Memory: Structures Adapted from Computer Organization and Design, Patterson & Hennessy, © 2005, and from slides kindly made.
Lecture 12: DRAM Basics Today: DRAM terminology and basics, energy innovations.
DRAM. Any read or write cycle starts with the falling edge of the RAS signal. –As a result the address applied in the address lines will be latched.
1 Lecture 15: DRAM Design Today: DRAM basics, DRAM innovations (Section 5.3)
1 COMP 206: Computer Architecture and Implementation Montek Singh Wed., Nov. 19, 2003 Topic: Main Memory (DRAM) Organization.
1 Evgeny Bolotin – ClubNet Nov 2003 Network on Chip (NoC) Evgeny Bolotin Supervisors: Israel Cidon, Ran Ginosar and Avinoam Kolodny ClubNet - November.
1 COMP 206: Computer Architecture and Implementation Montek Singh Mon., Nov. 18, 2002 Topic: Main Memory (DRAM) Organization – contd.
8-5 DRAM ICs High storage capacity Low cost Dominate high-capacity memory application Need “refresh” (main difference between DRAM and SRAM) -- dynamic.
Network-on-Chip: Communication Synthesis Department of Computer Science Texas A&M University.
Overview Booth’s Algorithm revisited Computer Internal Memory Cache memory.
CSIT 301 (Blum)1 Memory. CSIT 301 (Blum)2 Types of DRAM Asynchronous –The processor timing and the memory timing (refreshing schedule) were independent.
Lecture 12 Today’s topics –CPU basics Registers ALU Control Unit –The bus –Clocks –Input/output subsystem 1.
5.1 Semiconductor main memory  Organization The basic element of a semiconductor memory is the memory cell. Semiconductor memory cells properties: 1.
HyperTransport™ Technology I/O Link Presentation by Mike Jonas.
ECE 526 – Network Processing Systems Design Network Processor Architecture and Scalability Chapter 13,14: D. E. Comer.
CPE232 Memory Hierarchy1 CPE 232 Computer Organization Spring 2006 Memory Hierarchy Dr. Gheith Abandah [Adapted from the slides of Professor Mary Irwin.
CSIE30300 Computer Architecture Unit 07: Main Memory Hsin-Chou Chi [Adapted from material by and
Survey of Existing Memory Devices Renee Gayle M. Chua.
Lessons Learned The Hard Way: FPGA  PCB Integration Challenges Dave Brady & Bruce Riggins.
Chapter 5 Internal Memory. Semiconductor Memory Types.
Buffer-On-Board Memory System 1 Name: Aurangozeb ISCA 2012.
Main Memory CS448.
Computer Architecture System Interface Units Iolanthe II approaches Coromandel Harbour.
University of Tehran 1 Interface Design DRAM Modules Omid Fatemi
CS-350 TERM PROJECT COMPUTER BUSES By : AJIT UMRANI.
Memory Cell Operation.
Modern DRAM Memory Architectures Sam Miller Tam Chantem Jon Lucas CprE 585 Fall 2003.
Dynamic Random Access Memory (DRAM) CS 350 Computer Organization Spring 2004 Aaron Bowman Scott Jones Darrell Hall.
CS/EE 5810 CS/EE 6810 F00: 1 Main Memory. CS/EE 5810 CS/EE 6810 F00: 2 Main Memory Bottom Rung of the Memory Hierarchy 3 important issues –capacity »BellÕs.
Semiconductor Memory Types
The Evolution of Dynamic Random Access Memory (DRAM) CS 350 Computer Organization and Architecture Spring 2002 Section 1 Nicole Chung Brian C. Hoffman.
COMP541 Memories II: DRAMs
1 Adapted from UC Berkeley CS252 S01 Lecture 18: Reducing Cache Hit Time and Main Memory Design Virtucal Cache, pipelined cache, cache summary, main memory.
Simultaneous Multi-Layer Access Improving 3D-Stacked Memory Bandwidth at Low Cost Donghyuk Lee, Saugata Ghose, Gennady Pekhimenko, Samira Khan, Onur Mutlu.
Contemporary DRAM memories and optimization of their usage Nebojša Milenković and Vladimir Stanković, Faculty of Electronic Engineering, Niš.
Chapter 5 Internal Memory. contents  Semiconductor main memory - organisation - organisation - DRAM and SRAM - DRAM and SRAM - types of ROM - types of.
Computer Architecture Chapter (5): Internal Memory
“With 1 MB RAM, we had a memory capacity which will NEVER be fully utilized” - Bill Gates.
Univ. of TehranIntroduction to Computer Network1 An Introduction to Computer Networks University of Tehran Dept. of EE and Computer Engineering By: Dr.
COMP541 Memories II: DRAMs
Chapter 5 Internal Memory
William Stallings Computer Organization and Architecture 7th Edition
HyperTransport™ Technology I/O Link
COMP541 Memories II: DRAMs
William Stallings Computer Organization and Architecture 7th Edition
William Stallings Computer Organization and Architecture 8th Edition
Lecture: Memory, Multiprocessors
Israel Cidon, Ran Ginosar and Avinoam Kolodny
William Stallings Computer Organization and Architecture 8th Edition
Direct Rambus DRAM (aka SyncLink DRAM)
DRAM Hwansoo Han.
William Stallings Computer Organization and Architecture 8th Edition
Presentation transcript:

Serial Network SDRAM ENEE 759H Spring 2003

Introduction SDRAM system drawbacks  No parallelism for memory accesses  Multitude of pins for address/command/data Overall Goals  Increase parallelism, reduce latency  Reduce pin count  Attempt to increase bandwidth

Motivation Poulton’s idea  Bi-directional serial links.  Theoretically high bandwidth!  Less pins required for same functionality!  Looks perfect! *Graphic from Poulton’s Signaling Tutorial

Evolution I Initial design  Split topology.  Effectively halve latency.  Complicated protocol and connection details.

Evolution II Initial design  Individual DRAM chips directly connected.  High overall bandwidth.  Inflexible, lower capacity for system. We need a better design!

The Next Step Want simple system interconnects Keep basic SDRAM chip structure intact Utilize the strengths of both parallel and serial connections Create a system that facilitates parallelism

System Overview Take a “step back”… Consider memory module interface. Consider inter-chip interface on module.

System Overview 1 logical channel, 4 physical channels 3.2 GHz point-to-point connections Each channel called “module” 5 pins/module on memory controller Intra-module connections: parallel External connections: high speed serial

Module Topology

System Details I

System Details II

System Details – Protocol I The Command Set CMDUSEOPADDR? NOPNo operation.000N ACTActivate a row; uses bank and row address.001Y READSelects bank/column, initiates read burst.010Y WRITESelect bank and column, initiate write burst.011Y PRECPrecharge; deactivate row in bank.100* AUTORAuto-refresh; enter refresh mode.101N XXXReserved110 XXXReserved111

System Details – Protocol II Packets  18 bit command/address  32 bit data packets COMMANDActivate this row and bank… COMMANDStart a READ burst at this column… *Operating at 3.2GHz, command packets take 5.62ns; data packets take 10ns (the same as SDRAM operating at 100 MHz).

Cubing I “Chip stacking” Developed by Irvine- Sensors Corp. Currently can stack two 256 Mbit chips. Smaller footprint/area! Much shorter connection wires! *Graphics from Irvine-Sensors Data Sheet

Cubing II – Serial Network Point-to-point star topology. Dedicated circuits - high speed serial lines. Departure from “traditional” bus concept.

System Access Protocol Consecutive access to same module  Similar timing as SDRAM. Bandwidth matched between parallel and serial. DIN/DOUT buffers - no additional timing constraints. *Graphic from Dr. Jacob and Dave Wang

System Access Protocol Independent, simultaneous access to separate modules.  No inter-module timing issues. *Graphic from Dr. Jacob and Dave Wang Conventional SDRAM:

Serial Network Advantages I Path length matching  No more heroic routing!  Star topology is symmetric. No clock mismatch issues…  Everyone is on time! *Graphic from Dr. Jacob and Dave Wang

Serial Network Advantages IIa No need for bus termination.  Point-to-point communication, terminated in module. *Graphic from Dr. Jacob and Dave Wang

Serial Network Advantages IIb Serial/P2P vs. RAMBUS multi-drop.  Faster signaling!  No ringing!  Clean timing.  Serial wins… RAMBUSted! *Graphic from Dr. Jacob and Dave Wang

System Simulation SimpleScalar  Single CPU, Single Thread  SNSDRAM(32 Meg x 8)  1 rank in every memory module  Channel width : 32 bits  One extra cycle of Transaction Queue Delay to model the parallel to serial conversion

Simulation Run I - Parallel Bus Channel Rank Per Channel Sim_Cycles

Simulation Run I - Serial Network Channel Rank Per Channel Sim_Cycles

Simulation I Cycles Chart

Simulation Run II – Parallel Bus Channel Rank Per Channel Sim_Cycles

Simulation Run II – Serial Network Channel Rank Per Channel Sim_Cycles

Simulation II Cycles Chart

Memory Mapping Basic SDRAM High Performance SDRAM Row IDRankBankHi Col IDChannel IDLo Col IDCol Size RankRow IDBankHi Col IDChannel IDLo Col IDCol Size

Analysis Cache line = 64 byte channel width Read after Read Multi-CPU Single CPU Multi-Thread

Summary I Recall…  SDRAM has complex interface, simple chips.  RDRAM has a simple interface, but very complex chips. SNSDRAM…  Blends these seemingly split philosophies!

Summary II Advantages  Smaller pin count on memory controller.  Independent memory modules facilitate parallelism.  Simulated performance improvement over similar SDRAM configurations.  Smaller system footprint with cubing technology.  Theoretically scalable.