Electrical and Computer Engineering

Slides:



Advertisements
Similar presentations
1 Lecture 13: Cache and Virtual Memroy Review Cache optimization approaches, cache miss classification, Adapted from UCB CS252 S01.
Advertisements

Virtual Memory. The Limits of Physical Addressing CPU Memory A0-A31 D0-D31 “Physical addresses” of memory locations Data All programs share one address.
Appendix C: Review of Memory Hierarchy David Patterson Electrical Engineering and Computer Sciences University of California, Berkeley
CPE 731 Advance Computer Architecture Memory Hierarchy Review Dr. Gheith Abandah Adapted from the slides of Prof. David Patterson, University of California,
Computer Architecture Lec 4 – Memory Hierarchy Review.
Virtual Memory Adapted from lecture notes of Dr. Patterson and Dr. Kubiatowicz of UC Berkeley.
CIS429/529 Cache Basics 1 Caches °Why is caching needed? Technological development and Moore’s Law °Why are caches successful? Principle of locality °Three.
Virtual Memory Adapted from lecture notes of Dr. Patterson and Dr. Kubiatowicz of UC Berkeley and Rabi Mahapatra & Hank Walker.
Now, Review of Memory Hierarchy
1  1998 Morgan Kaufmann Publishers Chapter Seven Large and Fast: Exploiting Memory Hierarchy.
1 Chapter Seven Large and Fast: Exploiting Memory Hierarchy.
1 COMP 206: Computer Architecture and Implementation Montek Singh Mon, Oct 31, 2005 Topic: Memory Hierarchy Design (HP3 Ch. 5) (Caches, Main Memory and.
331 Lec20.1Fall :332:331 Computer Architecture and Assembly Language Fall 2003 Week 13 Basics of Cache [Adapted from Dave Patterson’s UCB CS152.
CIS °The Five Classic Components of a Computer °Today’s Topics: Memory Hierarchy Cache Basics Cache Exercise (Many of this topic’s slides were.
ECE 232 L27.Virtual.1 Adapted from Patterson 97 ©UCBCopyright 1998 Morgan Kaufmann Publishers ECE 232 Hardware Organization and Design Lecture 27 Virtual.
Review of Memory Hierarchy (Appendix C)
ENEE350 Ankur Srivastava University of Maryland, College Park Based on Slides from Mary Jane Irwin ( )
331 Lec20.1Spring :332:331 Computer Architecture and Assembly Language Spring 2005 Week 13 Basics of Cache [Adapted from Dave Patterson’s UCB CS152.
1 SRAM: –value is stored on a pair of inverting gates –very fast but takes up more space than DRAM (4 to 6 transistors) DRAM: –value is stored as a charge.
1 CSE SUNY New Paltz Chapter Seven Exploiting Memory Hierarchy.
Storage HierarchyCS510 Computer ArchitectureLecture Lecture 12 Storage Hierarchy.
Lecture 10 Memory Hierarchy and Cache Design Computer Architecture COE 501.
Memory Hierarchy Review. 2 Review from last lecture Quantify and summarize performance –Ratios, Geometric Mean, Multiplicative Standard Deviation F&P:
Memory Hierarchy. Since 1980, CPU has outpaced DRAM... CPU 60% per yr 2X in 1.5 yrs DRAM 9% per yr 2X in 10 yrs 10 DRAM CPU Performance (1/latency) 100.
Lecture 5 Review of Memory Hierarchy (Appendix C in textbook)
CSE 820 Advanced Computer Architecture Lec 4 – Memory Hierarchy Review Based on slides by David Patterson.
CSIE30300 Computer Architecture Unit 08: Cache Hsin-Chou Chi [Adapted from material by and
Appendix C Memory Hierarchy. Why care about memory hierarchies? Processor-Memory Performance Gap Growing Major source of stall cycles: memory accesses.
1  1998 Morgan Kaufmann Publishers Recap: Memory Hierarchy of a Modern Computer System By taking advantage of the principle of locality: –Present the.
EEL5708/Bölöni Lec 4.1 Fall 2004 September 10, 2004 Lotzi Bölöni EEL 5708 High Performance Computer Architecture Review: Memory Hierarchy.
Virtual Memory. Virtual Memory: Topics Why virtual memory? Virtual to physical address translation Page Table Translation Lookaside Buffer (TLB)
Review (1/2) °Caches are NOT mandatory: Processor performs arithmetic Memory stores data Caches simply make data transfers go faster °Each level of memory.
The Goal: illusion of large, fast, cheap memory Fact: Large memories are slow, fast memories are small How do we create a memory that is large, cheap and.
Computer Organization & Programming
1 Chapter Seven. 2 Users want large and fast memories! SRAM access times are ns at cost of $100 to $250 per Mbyte. DRAM access times are ns.
Multilevel Caches Microprocessors are getting faster and including a small high speed cache on the same chip.
CS.305 Computer Architecture Memory: Caches Adapted from Computer Organization and Design, Patterson & Hennessy, © 2005, and from slides kindly made available.
CPE232 Cache Introduction1 CPE 232 Computer Organization Spring 2006 Cache Introduction Dr. Gheith Abandah [Adapted from the slides of Professor Mary Irwin.
1 Chapter Seven CACHE MEMORY AND VIRTUAL MEMORY. 2 SRAM: –value is stored on a pair of inverting gates –very fast but takes up more space than DRAM (4.
1 Chapter Seven. 2 Users want large and fast memories! SRAM access times are ns at cost of $100 to $250 per Mbyte. DRAM access times are ns.
1  1998 Morgan Kaufmann Publishers Chapter Seven.
Summary of caches: The Principle of Locality: –Program likely to access a relatively small portion of the address space at any instant of time. Temporal.
1 Chapter Seven. 2 SRAM: –value is stored on a pair of inverting gates –very fast but takes up more space than DRAM (4 to 6 transistors) DRAM: –value.
CS 5513 Computer Architecture Lecture 4 – Memory Hierarchy Review.
HY425 – Αρχιτεκτονική Υπολογιστών Διάλεξη 04
CS161 – Design and Architecture of Computer
CMSC 611: Advanced Computer Architecture
Soner Onder Michigan Technological University
ECE232: Hardware Organization and Design
Yu-Lun Kuo Computer Sciences and Information Engineering
CS161 – Design and Architecture of Computer
The Goal: illusion of large, fast, cheap memory
Since 1980, CPU has outpaced DRAM ...
Virtual Memory Use main memory as a “cache” for secondary (disk) storage Managed jointly by CPU hardware and the operating system (OS) Programs share main.
Lec 3 – Memory Hierarchy Review
Sandeep K. S. Gupta Dept. CSE, ASU
David Patterson Electrical Engineering and Computer Sciences
Lecture 14 Virtual Memory and the Alpha Memory Hierarchy
CPSC 614 Computer Architecture Lec 6 – Memory Hierarchy Review
Rose Liu Electrical Engineering and Computer Sciences
CMSC 611: Advanced Computer Architecture
CS 5513 Computer Architecture Lecture 4 – Memory Hierarchy Review
5 Basic Cache Optimizations
CPE 631 Lecture 05: Cache Design
EE108B Review Session #6 Daxia Ge Friday February 23rd, 2007
CSC3050 – Computer Architecture
September 1, 2000 Prof. John Kubiatowicz
Cache Memory Rabi Mahapatra
CPE 631 Lecture 04: Review of the ABC of Caches
Presentation transcript:

Electrical and Computer Engineering EEL 5764: Graduate Computer Architecture Appendix C – Memory Hierarchy Review Ann Gordon-Ross Electrical and Computer Engineering University of Florida http://www.ann.ece.ufl.edu/ These slides are provided by: David Patterson Electrical Engineering and Computer Sciences, University of California, Berkeley Modifications/additions have been made from the originals CS252 S05

Since 1980, CPU has outpaced DRAM ... Q. How do architects address this gap? A. Put smaller, faster “cache” memories between CPU and DRAM. Create a “memory hierarchy”. Performance (1/latency) CPU 60% per yr 2X in 1.5 yrs 1000 CPU Gap grew 50% per year 100 DRAM 9% per yr 2X in 10 yrs 10 RamBus Solution to gap = caches DRAM 1980 1990 2000 Year CS252 S05

1977: DRAM faster than microprocessors Apple ][ (1977) Steve Wozniak Steve Jobs CPU: 1000 ns DRAM: 400 ns Old days…. No cache CPU slower than DRAM No worries CS252 S05

Levels of the Memory Hierarchy Upper Level Capacity Access Time Cost Staging Xfer Unit faster CPU Registers 100s Bytes <10s ns Registers Instr. Operands prog./compiler 1-8 bytes Cache K Bytes 10-100 ns 1-0.1 cents/bit Cache cache cntl 8-128 bytes Blocks Main Memory M Bytes 200ns- 500ns $.0001-.00001 cents /bit Memory OS 512-4K bytes Pages Disk G Bytes, 10 ms (10,000,000 ns) 10 - 10 cents/bit Memory hierarchy has been talked about forever Disk -5 -6 user/operator Mbytes Files Tape infinite sec-min 10 Larger Tape Lower Level -8 2/28/2019 CS252 S05

Memory Hierarchy: Apple iMac G5 1.6 GHz Managed by compiler Managed by hardware Managed by OS, hardware, application 07 Reg L1 Inst L1 Data L2 DRAM Disk Size 1K 64K 32K 512K 256M 80G Latency Cycles, Time 1, 0.6 ns 3, 1.9 ns 11, 6.9 ns 88, 55 ns 107, 12 ms Goal: Illusion of large, fast, cheap memory 20 years later, but this is a couple years old Create illusion of fast, large, cheap memory Let programs address a memory space that scales to the disk size, at a speed that is usually as fast as register access CS252 S05

iMac’s PowerPC 970: All caches on-chip L1 (64K Instruction) (1K) Registers 512K L2 Registers are large because they are dual ported L2 is denser so not so big Just bits, no control logic Memory 1/2 of chip Some with L3 cache L1 (32K Data) CS252 S05

The Principle of Locality Program access a relatively small portion of the address space at any instant of time. Two Different Types of Locality: Temporal Locality (Locality in Time): If an item is referenced, it will tend to be referenced again soon (e.g., loops, reuse) Spatial Locality (Locality in Space): If an item is referenced, items whose addresses are close by tend to be referenced soon (e.g., straightline code, array access) Last 15 years, HW relied on locality for speed Why caches stick around?? Cuz of locality, why? Could be the nature of how we program, but this may change one day It is a property of programs which is exploited in machine design. 2/28/2019 CS252 S05

Programs with locality cache well ... Bad locality behavior Temporal Locality Memory Address (one dot per access) Case for virtual memory, collected data at IBM Spatial Locality Donald J. Hatfield, Jeanette Gerald: Program Restructuring for Virtual Memory. IBM Systems Journal 10(3): 168-192 (1971) Time CS252 S05

Memory Hierarchy: Terminology Hit: data appears in some block in the upper level (example: Block X) Hit Rate: the fraction of memory access found in the upper level Hit Time: Time to access the upper level which consists of RAM access time + Time to determine hit/miss Miss: data needs to be retrieve from a block in the lower level (Block Y) Miss Rate = 1 - (Hit Rate) Miss Penalty: Time to replace a block in the upper level + Time to deliver the block the processor Hit Time << Miss Penalty (500 instructions) May be better to recalculate results instead of refetching Lower Level Memory Upper Level To Processor From Processor Blk X Blk Y 2/28/2019 CS252 S05

Cache Measures Hit rate: fraction found in that level So high that usually talk about Miss rate Miss rate fallacy: as MIPS to CPU performance, miss rate to average memory access time in memory Average memory-access time = Hit time + Miss rate x Miss penalty (ns or clocks) Miss penalty: time to replace a block from lower level, including time to replace in CPU access time: time to lower level = f(latency to lower level) transfer time: time to transfer block =f(BW between upper & lower levels) Average memory access time is a better measurement 2/28/2019 CS252 S05

4 Questions for Memory Hierarchy Q1: Where can a block be placed in the upper level? (Block placement) Q2: How is a block found if it is in the upper level? (Block identification) Q3: Which block should be replaced on a miss? (Block replacement) Q4: What happens on a write? (Write strategy) 2/28/2019 CS252 S05

Q1: Where can a block be placed in the upper level? Block 12 placed in 8 block cache: Fully associative, direct mapped, 2-way set associative S.A. Mapping = Block Number Modulo Number Sets Direct Mapped (12 mod 8) = 4 2-Way Assoc (12 mod 4) = 0 Full Mapped 01234567 01234567 01234567 Cache 1111111111222222222233 01234567890123456789012345678901 Memory 2/28/2019 CS252 S05

Q2: How is a block found if it is in the upper level? Tag on each block No need to check index or block offset Increasing associativity shrinks index, expands tag Block Offset Block Address Index Tag 2/28/2019 CS252 S05

Q3: Which block should be replaced on a miss? Easy for Direct Mapped Set Associative or Fully Associative: Random LRU (Least Recently Used) Assoc: 2-way 4-way 8-way Size LRU Ran LRU Ran LRU Ran 16 KB 5.2% 5.7% 4.7% 5.3% 4.4% 5.0% 64 KB 1.9% 2.0% 1.5% 1.7% 1.4% 1.5% 256 KB 1.15% 1.17% 1.13% 1.13% 1.12% 1.12% LRU - Becomes more complicated as associativity goes up - more hardware required Usually use random for high associativities Random and LRU perform close 2/28/2019 CS252 S05

Q4: What happens on a write? Write-Through Write-Back Policy Data written to cache block also written to lower-level memory Write data only to the cache Update lower level when a block falls out of the cache Debug Easy Hard Do read misses produce writes? No Yes Do repeated writes make it to lower level? What do you do on write miss? Write allocate or write no-allocate Additional option -- let writes to an un-cached address allocate a new cache line (“write-allocate”). CS252 S05

Write Buffers for Write-Through Caches Processor Cache Write Buffer Lower Level Memory Holds data awaiting write-through to lower level memory Q. Why a write buffer ? A. So CPU doesn’t stall Q. Why a buffer, why not just one register ? A. Bursts of writes are common. To reduce traffic and eliminate CPU wait time Q. Are Read After Write (RAW) hazards an issue for write buffer? A. Yes! Drain buffer before next read, or send read 1st after check write buffers. CS252 S05

5 Basic Cache Optimizations Reducing Miss Rate Larger Block size (compulsory misses) Larger Cache size (capacity misses) Higher Associativity (conflict misses) Reducing Miss Penalty Multilevel Caches Reducing hit time Giving Reads Priority over Writes E.g., Read complete before earlier writes in write buffer 2,3 can effect hit time 2/28/2019 CS252 S05

Outline Memory hierarchy Locality Cache design Virtual address spaces Page table layout TLB design options 2/28/2019 CS252 S05

The Limits of Physical Addressing “Physical addresses” of memory locations Data All programs share one address space: The physical address space A0-A31 A0-A31 CPU Memory D0-D31 D0-D31 Nothing to stop program from screwing up another program Programmers used to have to figure out how to fit program in memory and manually load parts in and out as needed Machine language programs must be aware of the machine organization No way to prevent a program from accessing any machine resource CS252 S05

Solution: Add a Layer of Indirection “Physical Addresses” Address Translation Virtual Physical “Virtual Addresses” A0-A31 A0-A31 CPU Memory D0-D31 D0-D31 Data User programs run in an standardized virtual address space Solve with indirection Each user how has their own address space Otherwise OS has to figure out when and who can access parts of memory Could change grades Provides protection Allow for memory to appear bigger Address Translation hardware managed by the operating system (OS) maps virtual address to physical memory Hardware supports “modern” OS features: Protection, Translation, Sharing CS252 S05

Three Advantages of Virtual Memory Translation: Program can be given consistent view of memory, even though physical memory is scrambled Makes multithreading reasonable (now used a lot!) Only the most important part of program (“Working Set”) must be in physical memory. Contiguous structures (like stacks) use only as much physical memory as necessary yet still grow later. Protection: Different processes protected from each other. Different pages can be given special behavior (Read Only, Invisible to user programs, etc). Kernel data protected from User programs Very important for protection from malicious programs Sharing: Can map same physical page to multiple users (“Shared memory”) Original goals - put the important part in memory, rest on disk Now there is more memory so main goals have changed 2/28/2019 CS252 S05

Page tables encode virtual address spaces A virtual address space is divided into blocks of memory called pages Physical Address Space Virtual frame A machine usually supports pages of a few sizes (MIPS R4000): frame frame frame Page based memory Mapping (page table) between virtual and physical address space A valid page table entry codes physical memory “frame” address for the page CS252 S05

Page tables encode virtual address spaces OS manages the page table for each ASID Physical Memory Space A valid page table entry codes physical memory “frame” address for the page A virtual address space is divided into blocks of memory called pages frame frame frame A machine usually supports pages of a few sizes (MIPS R4000): frame A page table is indexed by a virtual address CS252 S05

Details of Page Table Page Table Physical Memory Space Virtual Address Page Table index into page table Base Reg V Access Rights PA V page no. offset 12 table located in physical memory P page no. Physical Address frame frame frame frame virtual address This is where access rights are checked Page table maps virtual page numbers to physical frames (“PTE” = Page Table Entry) Virtual memory => treat memory  cache for disk 2/28/2019 CS252 S05

Page tables may not fit in memory! A table for 4KB pages for a 32-bit address space has 1M entries Each process needs its own address space! Two-level Page Tables P1 index P2 index Page Offset 31 12 11 21 22 32 bit virtual address Multi-level page tables to save space if page tables too big Top-level table wired in main memory Subset of 1024 second-level tables in main memory; rest are on disk or unallocated CS252 S05

VM and Disk: Page replacement policy ... Page Table 1 0 used dirty 0 1 1 1 0 0 Dirty bit: page written. Used bit: set to 1 on any reference Set of all pages in Memory Tail pointer: Clear the used bit in the page table Head pointer Place pages on free list if used bit is still clear. Schedule pages with dirty bit set to be written to disk. Freelist Free Pages Ring model, check pages Architect’s role: support setting dirty and used bits CS252 S05

MIPS Address Translation: How does it work? “Virtual Addresses” “Physical Addresses” A0-A31 Virtual Physical A0-A31 Translation Look-Aside Buffer (TLB) Translation Look-Aside Buffer (TLB) A small fully-associative cache of mappings from virtual to physical addresses CPU Memory D0-D31 D0-D31 Data What is the table of mappings that it caches? Now each memory access could take twice as long - page fault and fetch data TLB also contains protection bits for virtual address Fast common case: Virtual address is in TLB, process has permission to read/write it. CS252 S05