Download presentation
Presentation is loading. Please wait.
Published byJames Mervin Gallagher Modified over 9 years ago
1
Memory Network: Enabling Technology for Scalable Near-Data Computing Gwangsun Kim, John Kim Korea Advanced Institute of Science and Technology Jung Ho Ahn Seoul National University Yongkee Kwon SK Hynix
2
Memory Network I/O port … Vault controller I/O port Intra-HMC Network Vault controller … Logic layer High-speed link DRAM layers Vault Hybrid Memory Cube (HMC) 2/10 “Near”-data processing with multiple memories? “Far”-data? Memory network enables scalable near-data computing. Data AData B “compute A+B” Data B
3
DIVA Processing-in-Memory (PIM) Chip Draper et al., “The architecture of the DIVA processing-in-memory chip”, ICS’02 For multimedia and irregular applications. Proposed memory network for PIM modules. Simple low-dimensional network (e.g., ring ) High packet hop count performance & energy inefficiency Advanced technology is available – high off-chip bandwidth 3/10
4
Memory Networks from Micron D. R. Resnick, “Memory Network Methods, Apparatus, and Systems,” US Patent Application Publication, US20100211721 A1, 2010. 2D Mesh topology 4/10 Local memories Network-attached memories
5
Memory Network Design Issues Difficult to leverage high-radix topology – Low-radix vs. high-radix topology – High-radix topology smaller network diameter – Limited # of ports in memory modules. Adaptive routing requirement – Can increase network cost – Depends on traffic pattern, memory mapping, etc. 5/10 Low-radix networks High-radix networks
6
Memory-centric Network Host-memory bandwidth still matters. – To support conventional applications while adopting NDP. – NDP involves communication with host processors. MCN Leverage the same network for NDP. Network … Processor-centric Network (PCN) (e.g., Intel QPI, AMD HyperTransport) … … CPU Network … … … CPU Memory-centric Network (MCN) [PACT’13] Memory BW Processor-to processor BW Flexible BW utilization Separate network required for NDP The same network can be used for NDP 6/10
7
Memory Network for Heterogeneous NDP NDP for not only CPU, but also for GPU. Unified memory network for multi-GPU systems [MICRO’14]. Extending the memory network for heterogeneous NDP. CPUGPU … … Unified Memory Network … … … … … …… … FPGA … … … 7/10
8
Hierarchical Network With intra-HMC network, the memory network is a hierarchical network. NDP requires additional processing elements at the logic layer. Need to support various types of traffic – Local (on-chip) traffic vs. global traffic – Conventional memory access traffic vs. NDP-induced traffic DRAM (stacked) Hybrid Memory Cube Vault controller On-chip channel I/O port Concentrated Mesh-based intra-HMC network [PACT’13] 8/10
9
Issues with Memory Network-based NDP Power management – Large number of channels possible in memory network – Power-gating, DVFS, and other circuit-level techniques. Data placement & migration – Optimal placement of shared data – Migration within memory network Consistency & coherence – Direct memory access by multiple processors – Heterogeneous processors 9/10
10
Summary Memory network can enable scalable near-data processing. Leveraging recent memory network researches – Memory-centric network [PACT’13] – Unified memory network [MICRO’14] Intra-HMC design considerations Further issues 10/10
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.