Caching Principles and Paging Performance

Slides:



Advertisements
Similar presentations
Paging 1 CS502 Spring 2006 Paging CS-502 Operating Systems.
Advertisements

CGS 3763 Operating Systems Concepts Spring 2013 Dan C. Marinescu Office: HEC 304 Office hours: M-Wd 11: :30 AM.
CSIE30300 Computer Architecture Unit 10: Virtual Memory Hsin-Chou Chi [Adapted from material by and
Virtual Memory Hardware Support
CMPT 300: Operating Systems I Dr. Mohamed Hefeeda
Outline, October 2CS-3013 A-term 20091Outline, October 2CS-3013 A-term Outline for October 2 Review:– Caches and Quantification of Cache Performance.
PagingCS-3013 A-term Paging CS-3013, Operating Systems A-term 2009 (Slides include materials from Modern Operating Systems, 3 rd ed., by Andrew Tanenbaum.
CSCE 212 Chapter 7 Memory Hierarchy Instructor: Jason D. Bakos.
1 Lecture 20 – Caching and Virtual Memory  2004 Morgan Kaufmann Publishers Lecture 20 Caches and Virtual Memory.
S.1 Review: The Memory Hierarchy Increasing distance from the processor in access time L1$ L2$ Main Memory Secondary Memory Processor (Relative) size of.
Recap. The Memory Hierarchy Increasing distance from the processor in access time L1$ L2$ Main Memory Secondary Memory Processor (Relative) size of the.
1 Chapter Seven Large and Fast: Exploiting Memory Hierarchy.
CS-3013 & CS-502 Summer 2006 Paging1 CS-3013 & CS-502 Summer 2006.
PagingCS-3013 C-term Paging CS-3013 Operating Systems C-term 2008 (Slides include materials from Operating System Concepts, 7 th ed., by Silbershatz,
1 SRAM: –value is stored on a pair of inverting gates –very fast but takes up more space than DRAM (4 to 6 transistors) DRAM: –value is stored as a charge.
Virtual Memory Topics Virtual Memory Access Page Table, TLB Programming for locality Memory Mountain Revisited.
Virtual Memory Management CS-3013 A-term Virtual Memory Management CS-3013, Operating Systems A-term 2009 (Slides include materials from Modern Operating.
1 CSE SUNY New Paltz Chapter Seven Exploiting Memory Hierarchy.
CMPE 421 Parallel Computer Architecture
CS 153 Design of Operating Systems Spring 2015 Lecture 17: Paging.
CSE431 L22 TLBs.1Irwin, PSU, 2005 CSE 431 Computer Architecture Fall 2005 Lecture 22. Virtual Memory Hardware Support Mary Jane Irwin (
0 High-Performance Computer Architecture Memory Organization Chapter 5 from Quantitative Architecture January 2006.
PagingCS-502 Fall Paging CS-502 Operating Systems Fall 2006 (Slides include materials from Operating System Concepts, 7 th ed., by Silbershatz, Galvin,
CS 149: Operating Systems March 3 Class Meeting Department of Computer Science San Jose State University Spring 2015 Instructor: Ron Mak
3-May-2006cse cache © DW Johnson and University of Washington1 Cache Memory CSE 410, Spring 2006 Computer Systems
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition Virtual Memory.
Caching Principles & Paging Performance CS-502 (EMC) Fall Caching Principles and Paging Performance CS-502, Operating Systems Fall 2009 (EMC) (Slides.
Paging (continued) & Caching CS-3013 A-term Paging (continued) & Caching CS-3013 Operating Systems A-term 2008 (Slides include materials from Modern.
Memory Architecture Chapter 5 in Hennessy & Patterson.
1 Virtual Memory. Cache memory: provides illusion of very high speed Virtual memory: provides illusion of very large size Main memory: reasonable cost,
1 Chapter Seven CACHE MEMORY AND VIRTUAL MEMORY. 2 SRAM: –value is stored on a pair of inverting gates –very fast but takes up more space than DRAM (4.
CS.305 Computer Architecture Memory: Virtual Adapted from Computer Organization and Design, Patterson & Hennessy, © 2005, and from slides kindly made available.
1  1998 Morgan Kaufmann Publishers Chapter Seven.
LECTURE 12 Virtual Memory. VIRTUAL MEMORY Just as a cache can provide fast, easy access to recently-used code and data, main memory acts as a “cache”
Memory Management – Page 1CSCI 4717 – Computer Architecture CSCI 4717/5717 Computer Architecture Topic: Memory Management *** Modified – look for Reading:
Memory Management memory hierarchy programs exhibit locality of reference - non-uniform reference patterns temporal locality - a program that references.
Cache Advanced Higher.
CMSC 611: Advanced Computer Architecture
Improving Memory Access The Cache and Virtual Memory
Memory Key Revision Points.
Memory and cache CPU Memory I/O.
The Goal: illusion of large, fast, cheap memory
CS352H: Computer Systems Architecture
Lecture 12 Virtual Memory.
Virtual Memory - Part II
Ramya Kandasamy CS 147 Section 3
How will execution time grow with SIZE?
CS 704 Advanced Computer Architecture
Chapter 8 Digital Design and Computer Architecture: ARM® Edition
Andy Wang Operating Systems COP 4610 / CGS 5765
Lecture 23: Cache, Memory, Virtual Memory
Lecture 22: Cache Hierarchies, Memory
Introduction to Memory Management
Lecture 29: Virtual Memory-Address Translation
Andy Wang Operating Systems COP 4610 / CGS 5765
CMSC 611: Advanced Computer Architecture
Morgan Kaufmann Publishers Memory Hierarchy: Virtual Memory
Practical Session 9, Memory
Memory Operation and Performance
Caching Principles and Paging Performance
Outline for October 9 Input and Output (continued)
Virtual Memory: Working Sets
Fundamentals of Computing: Computer Architecture
Cache Memory and Performance
Sarah Diesburg Operating Systems CS 3430
Andy Wang Operating Systems COP 4610 / CGS 5765
Virtual Memory.
Memory Management & Virtual Memory
CS-3013 Operating Systems A-term 2008
Sarah Diesburg Operating Systems COP 4610
Presentation transcript:

Caching Principles and Paging Performance CS-3013 Operating Systems Hugh C. Lauer (Slides include materials from Slides include materials from Modern Operating Systems, 3rd ed., by Andrew Tanenbaum and from Operating System Concepts, 7th ed., by Silbershatz, Galvin, & Gagne) CS-3013, A-Term 2011 Caching Principles and Paging Performance

Fundamental Observation Paging allows us to treat physical memory as a cache of virtual memory Therefore, all of the principles and issues of caches apply to paging … especially paging performance CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Definition — Cache A small subset of active items held in small, fast storage while most of the items remain in much larger, slower storage Includes a mechanism for bringing things into cache and throwing them out again! CS-3013, A-Term 2011 Caching Principles and Paging Performance

Note on Caches and Caching This topic is not adequately covered in Tanenbaum or most other Operating System textbooks. Silbershatz, Galvin, & Gagne discuss paging performance somewhat It is treated extensively in Computer Architecture textbooks Caching is a fundamental, cross-cutting concept with broad application to all areas of computing science and to many areas of system design, embedded systems, etc. CS-3013, A-Term 2011 Caching Principles and Paging Performance

Paging — Two Examples of Caches Paged Virtual Memory Very large, mostly stored on (slow) disk Small working set in (fast) RAM during execution Page tables Very large, mostly stored in (slow) RAM Small working set stored in (fast) TLB registers CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching is Ubiquitous in Computing Transaction processing Keep records of today’s departures in RAM or local storage while records of future flights are on remote database Program execution Keep the bytes near the current program counter in on-chip memory while rest of program is in RAM File management Keep disk maps of open files in RAM while retaining maps of all files on disk Game design Keep nearby environment in cache of each character … CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Caching issues This is a very important list! When to put something in the cache What to throw out to create cache space for new items How to keep cached item and stored item in sync after one or the other is updated How to keep multiple caches in sync across processors or machines Size of cache needed to be effective Size of cache items for efficiency … CS-3013, A-Term 2011 Caching Principles and Paging Performance

General Observation on Caching We create caches because There is not enough fast memory to hold everything we need Memory that is large enough is too slow Performance metric for all caches is EAT Effective Access Time Goal is to make overall performance close to cache memory performance By taking advantage of locality — temporal and spatial By burying a small number of accesses to slow memory under many, many accesses to fast memory CS-3013, A-Term 2011 Caching Principles and Paging Performance

Definition – Effective Access Time (EAT) The average access time to memory items, where some items are cached in fast storage and other items are not cached… …weighted by p, the fault rate 0 ≤ p < 1 EAT = (1-p) * (cache access time) + p * (non-cache access time) CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Goal of Caching To take advantage of locality to achieve nearly the same performance of the fast memory when most data is in slow memory I.e., solve EAT equation for p (1-p)*(cache_time) + p*(non_cache_time) < (1+x)*(cache_time) EAT x is “acceptable” performance penalty CS-3013, A-Term 2011 Caching Principles and Paging Performance

Goal of Caching (continued) Select size of cache and size of cache items so that p is low enough to meet acceptable performance goal Usually requires simulation of suite of benchmarks CS-3013, A-Term 2011 Caching Principles and Paging Performance

Application to Demand Paging Page Fault Rate (p) 0 < p < 1.0 (measured in average number of faults / reference) Page Fault Overhead = fault service time + read page time + restart process time Fault service time ~ 0.1–10 sec Restart process time ~ 0.1–10–100 sec Read page time ~ 8-20 milliseconds! Dominated by time to read page in from disk! CS-3013, A-Term 2011 Caching Principles and Paging Performance

Demand Paging Performance (continued) Effective Access Time (EAT) = (1-p) * (memory access time) + p * (page fault overhead) Want EAT to degrade no more than, say, 10% from true memory access time i.e., EAT < (1 + 10%) * memory access time CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Performance Example Memory access time = 100 nanosec = 10-7 Page fault overhead = 25 millisec = 0.025 Page fault rate = 1/1000 = 10-3 EAT = (1-p) * 10-7 + p * (0.025) = (0.999) * 10-7 + 10-3 * 0.025  25 microseconds per reference! I.e., 250 * memory access time! CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Performance Goal To achieve less than 10% degradation (1-p) * 10-7 + p * (0.025) < 1.1 * 10-7 i.e., p < (0.1 * 10-7) / (0.025 - 10-7)  0.0000004 I.e., 1 fault in 2,500,000 accesses! CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Working Set Size Assume average swap time of 25 millisec. For memory access time = 100 nanoseconds Require < 1 page fault per 2,500,000 accesses For memory access time = 1 microsecond Require < 1 page fault per 250,000 accesses For memory access time = 10 microseconds Require < 1 page fault per 25,000 accesses CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Object Lesson Working sets must get larger in proportion to memory speed! Disk speed ~ constant (nearly) I.e., faster computers need larger physical memories to exploit the speed! CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Class Discussion What is first thing to do when the PC you bought last year becomes too slow? What else might help? Can we do the same analysis on TLB performance? CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and TLB fault performance Assumptions m = memory access time = 100 nsec t = TLB load time from memory = 300 nsec = 3 * m Goal is < 5% penalty for TLB misses I.e., EAT < 1.05 * m How low does TLB fault rate need to be? CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and TLB fault performance Assumptions m = memory access time = 100 nsec t = TLB load time from memory = 300 nsec = 3 * m Goal is < 5% penalty for TLB misses I.e., EAT < 1.05 * m EAT = (1-p) * m + p * t < 1.05 *m  p < (0.05 * m) / (t – m) = 0.05 * m / 2 * m = 0.025 I.e., TLB fault rate should be < 1 per 40 accesses! CS-3013, A-Term 2011 Caching Principles and Paging Performance

TLB fault performance (continued) Q: How large should TLB be so that TLB faults are not onerous, in these circumstances? A: Somewhat less than 40 entries Assuming a reasonable degree of locality! Note: pathological cases may prevent goal altogether! CS-3013, A-Term 2011 Caching Principles and Paging Performance

What if Software Loaded TLB? E.g., with hashed or inverted page tables? Assume TLB load time is 100 * m Work out on white board CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Summary of this Topic A quantitative way of estimating how large the cache needs to be to avoid excessive thrashing, where Cache = Working set in physical memory Cache = TLB size in hardware Applicable to all forms of caching CS-3013, A-Term 2011 Caching Principles and Paging Performance

General Observation on Caching We create caches because There is not enough fast memory to hold everything we need Memory that is large enough is too slow Performance metric for all caches is EAT Effective Access Time Goal is to make overall performance close to cache memory performance By taking advantage of locality — temporal and spatial By burying a small number of accesses to slow memory under many, many accesses to fast memory CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Cache Applications Physical memory: cache of virtual memory I.e., RAM over disk TLB: cache of page table entries I.e., Registers over RAM Processor L2 cache: over RAM I.e., nanosecond memory over 10’s of nanoseconds Processor L1 cache: over L2 cache I.e., picosecond registers over nanosecond memory … CS-3013, A-Term 2011 Caching Principles and Paging Performance

Cache Applications (continued) Recently accessed blocks of a file I.e., RAM over disk blocks Today’s airline flights I.e., local disk over remote disk CS-3013, A-Term 2011 Caching Principles and Paging Performance

Caching Principles and Questions? CS-3013, A-Term 2011 Caching Principles and Paging Performance