Chapter 10: Virtual Memory

Slides:



Advertisements
Similar presentations
Background Virtual memory – separation of user logical memory from physical memory. Only part of the program needs to be in memory for execution. Logical.
Advertisements

Chapter 9: Virtual Memory
國立台灣大學 資訊工程學系 Chapter 9: Virtual Memory. 資工系網媒所 NEWS 實驗室 Objectives To describe the benefits of a virtual memory system To explain the concepts of demand.
Page 15/4/2015 CSE 30341: Operating Systems Principles Allocation of Frames  How should the OS distribute the frames among the various processes?  Each.
Advanced Operating Systems - Spring 2009 Lecture 17 – March 23, 2009 Dan C. Marinescu Office: HEC 439 B. Office hours:
Chapter 9: Virtual-Memory Management. 9.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts Chapter 9: Virtual Memory Background Demand Paging.
Segmentation and Paging Considerations
Virtual-Memory Management
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Lecture 17: Wrapping up Virtual Memory.
Gordon College Stephen Brinton
Virtual Memory CS 3100 Virutal Memory.
Chapter 9: Virtual Memory
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 10: Virtual Memory Background Demand Paging Process Creation Page Replacement.
Silberschatz, Galvin and Gagne ©2007 Operating System Concepts with Java – 7 th Edition, Nov 15, 2006 Chapter 9: Virtual Memory.
A. Frank - P. Weisberg Operating Systems Virtual Memory Policies.
Chapter 9: Virtual Memory. 9.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 22, 2005 Chapter 9: Virtual Memory Background.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Chapter 9: Virtual Memory.
Chapter 9: Virtual Memory. Background Demand Paging Copy-on-Write Page Replacement Allocation of Frames Thrashing Memory-Mapped Files Allocating Kernel.
Page 19/17/2015 CSE 30341: Operating Systems Principles Optimal Algorithm  Replace page that will not be used for longest period of time  Used for measuring.
Chapter 9: Virtual Memory. 9.2CSCI 380 – Operating Systems Chapter 9: Virtual Memory Background Demand Paging Copy-on-Write Page Replacement Allocation.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Chapter 9: Virtual Memory.
Chapter 9: Virtual-Memory Management. 9.2Operating System Principles Chapter 9: Virtual-Memory Management Background Demand Paging Copy-on-Write Page.
Chapter 9: Virtual Memory. 9.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 22, 2005 Background Virtual memory –
Operating System Concepts 7th Edition Abraham SilBerschatz Peter Baer Galvin Greg Gagne Prerequisite: CSE212.
Chapter 9: Virtual Memory. 9.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 22, 2005 Chapter 9: Virtual Memory Background.
Chapter 9: Virtual Memory. 9.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 22, 2005 Chapter 9: Virtual Memory Background.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition Virtual Memory.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 10: Virtual Memory Background Demand Paging Process Creation Page Replacement.
Chapter 10: Virtual Memory Background Demand Paging Process Creation Page Replacement Allocation of Frames Thrashing Operating System Examples Operating.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 10: Virtual Memory Background Demand Paging Page Replacement Allocation of.
Chapter 9: Virtual Memory. 9.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 22, 2005 Outline n Background n Demand.
Chapter 9: Virtual Memory. 9.2 Chapter 9: Virtual Memory Background Demand Paging Copy-on-Write Page Replacement Allocation of Frames Thrashing Memory-Mapped.
9.1 Silberschatz, Galvin and Gagne ©2005 Operating System Principles 9.5 Allocation of Frames Each process needs minimum number of pages Example: machine.
9.1 Silberschatz, Galvin and Gagne ©2009 Operating System Concepts with Java – 8 th Edition Chapter 9: Virtual Memory.
Chapter 9: Virtual Memory
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Chapter 9: Virtual Memory.
Silberschatz, Galvin and Gagne ©2007 Operating System Concepts with Java – 7 th Edition, Nov 15, 2006 Chapter 9: Virtual Memory.
9.1 Silberschatz, Galvin and Gagne ©2009 Operating System Concepts with Java – 8 th Edition Chapter 9: Virtual Memory.
Silberschatz, Galvin and Gagne ©2011 Operating System Concepts Essentials – 8 th Edition Chapter 9: Virtual Memory.
Chapter 9: Virtual Memory. Background Demand Paging Copy-on-Write Page Replacement Allocation of Frames Thrashing Memory-Mapped Files Allocating Kernel.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition Chapter 9: Virtual Memory.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Chapter 9: Virtual-Memory Management.
Chapter 9: Virtual Memory
OPERATING SYSTEM CONCEPTS AND PRACTISE
Chapter 9: Virtual Memory – Part II
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Virtual Memory.
Module 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
What Happens if There is no Free Frame?
5: Virtual Memory Background Demand Paging
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 9: Virtual-Memory Management
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 9: Virtual Memory
Chapter 10: Virtual Memory
Virtual Memory.
Module 9: Virtual Memory
Presentation transcript:

Chapter 10: Virtual Memory

Chapter 10: Virtual Memory Background Demand Paging Copy-on-Write Page Replacement Allocation of Frames Thrashing Memory-Mapped Files Allocating Kernel Memory Other Considerations Operating-System Examples

Objectives To describe the benefits of a virtual memory system To explain the concepts of demand paging, page-replacement algorithms, and allocation of page frames To discuss the principle of the working-set model To examine the relationship between shared memory and memory-mapped files To explore how kernel memory is managed

Working-Set Model  = 10 instructions   working-set window  a fixed number of page references Example: 10,000 instructions WSSi (working set of Process Pi) = total number of pages referenced in the most recent  (varies in time) if  too small will not encompass entire locality if  too large will encompass several localities if  =   will encompass entire program D =  WSSi  total demand for frames Approximation of locality if D > m  Thrashing (m = total number of available frames) Policy if D > m, then suspend or swap out one of the processes  = 10 instructions

Keeping Track of the Working Set Approximate with fixed interval timer + a reference bit Example:  = 10,000 references Timer interrupts after every 5000 references Whenever a timer interrupts, copy and set the values of reference bit to 0 for each page Keep in-memory 2 bits for each page If one of these bits = 1  page in working set Why is this not completely accurate? we cannot tell where, within an interval of 5,000, a reference occurred Improvement = 10 bits and interrupt every 1000 time units However, the cost to service these more frequent interrupts will be correspondingly higher.

Page-Fault Frequency More direct approach than WSS Establish “acceptable” page-fault frequency (PFF) rate and use local replacement policy If actual rate too low, process loses frame If actual rate too high, process gains frame

Working Sets and Page Fault Rates Direct relationship between working set of a process and its page-fault rate Working set changes over time Peaks and valleys over time transition, stable

Memory-Mapped Files Memory-mapped file I/O allows file I/O to be treated as routine memory access by mapping a disk block to a page in memory A file is initially read using demand paging A page-sized portion of the file is read from the file system into a physical page Subsequent reads/writes to/from the file are treated as ordinary memory accesses Simplifies and speeds file access by driving file I/O through memory rather than read() and write() system calls Also allows several processes to map the same file allowing the pages in memory to be shared But when does written data make it to disk? Periodically and / or at file close() time For example, when the pager scans for dirty pages

Memory-Mapped File Technique for all I/O Some OSes uses memory mapped files for standard I/O Process can explicitly request memory mapping a file via mmap() system call Now file mapped into process address space For standard I/O (open(), read(), write(), close()), mmap anyway But map file into kernel address space Process still does read() and write() Copies data to and from kernel space and user space Uses efficient memory management subsystem Avoids needing separate subsystem COW can be used for read/write non-shared pages Memory mapped files can be used for shared memory (although again via separate system calls)

Memory Mapped Files

Shared Memory via Memory-Mapped I/O

Shared Memory in Windows API First create a file mapping for file to be mapped Then establish a view of the mapped file in process’s virtual address space Consider producer / consumer Producer create shared-memory object using memory mapping features Open file via CreateFile(), returning a HANDLE Create mapping via CreateFileMapping() creating a named shared-memory object Create view via MapViewOfFile() Sample code in Textbook

Allocating Kernel Memory When a process running in user mode requests additional memory, pages are allocated from the list of free page frames maintained by the kernel Treated differently from user memory Often allocated from a free-memory pool different from the list used to satisfy ordinary user-mode processes Kernel requests memory for structures of varying sizes, some of which are less than a page in size; minimize waste due to fragmentation many operating systems do not subject kernel code or data to the paging system. Pages allocated to user-mode processes do not necessarily have to be in contiguous physical memory certain hardware devices interact directly with physical memory—without the benefit of a virtual memory interface, and consequently may require memory residing in physically contiguous pages Two strategies for managing free memory that is assigned to kernel processes: the “buddy system” and slab allocation

Buddy System Allocates memory from fixed-size segment consisting of physically-contiguous pages Memory allocated using power-of-2 allocator Satisfies requests in units sized as power of 2 (4 KB, 8 KB, 16 KB, …) Request rounded up to next highest power of 2 (11 KB satisfied with a 16-KB segment When smaller allocation needed than is available, current chunk split into two buddies of next-lower power of 2 Continue until appropriate sized chunk available For example, assume 256KB chunk available, kernel requests 21KB Split into two buddies AL and AR of 128KB each One further divided into BL and BR of 64KB One further into CL and CR of 32KB each – one used to satisfy request Advantage – quickly coalesce unused chunks into larger chunk Disadvantage - fragmentation

Buddy System Allocator

Slab Allocator Alternate strategy Slab is one or more physically contiguous pages Cache consists of one or more slabs Single cache for each unique kernel data structure Each cache filled with objects – instantiations of the data structure For example, the cache representing semaphores stores instances of semaphore objects When cache created, objects marked as free are allocated to the cache When structures stored, objects assigned from cache are marked as used If slab is full of used objects, next object allocated from empty slab If no empty slabs, new slab allocated Benefits include no fragmentation, fast memory request satisfaction

Slab Allocation

Slab Allocator in Linux For example process descriptor is of type struct task_struct Approx 1.7KB of memory New task -> allocate new struct from cache Will use existing free struct task_struct Slab can be in three possible states Full – all used Empty – all free Partial – mix of free and used Upon request, slab allocator Uses free struct in partial slab If none, takes one from empty slab If no empty slab, create new empty

Slab Allocator in Linux (Cont.) Slab started in Solaris, now wide-spread for both kernel mode and user memory in various OSes Linux 2.2 had SLAB, now has both SLOB and SLUB allocators SLOB for systems with limited memory, e.g. embedded systems Simple List of Blocks – maintains 3 list objects for small (< 256 bytes), medium (<1024 bytes), large objects (< size of page) Memory requests are allocated from an object on the appropriate list using a first-fit policy. SLUB is performance-optimized SLAB, removes per-CPU queues, metadata stored in page structure

Other Considerations -- Prepaging To reduce the large number of page faults that occurs at process startup Prepage all or some of the pages a process will need, before they are referenced But if prepaged pages are unused, I/O and memory was wasted Assume s pages are prepaged and a fraction α (0 ≤ α ≤ 1) of the pages is used Is cost of s * α saved pages faults > or < than the cost of prepaging s * (1- α) unnecessary pages? α near zero  prepaging loses; if α is close to 1, prepaging wins Prepaging an executable program may be difficult, as it may be unclear exactly what pages should be brought in. Prepaging a file may be more predictable, since files are often accessed sequentially. The Linux readahead() system call prefetches the contents of a file into memory, subsequent accesses to the file will take place in main memory.

Other Issues – Page Size Sometimes OS designers have a choice Especially if running on custom-built CPU Page size selection must take into consideration: Fragmentation Page table size Resolution I/O overhead Number of page faults Locality TLB size and effectiveness Always power of 2, usually in the range 212 (4,096 bytes) to 222 (4,194,304 bytes) On average, growing over time

Other Issues – TLB Reach TLB Reach - The amount of memory accessible from the TLB TLB Reach = (TLB Size) X (Page Size) Ideally, the working set of each process is stored in the TLB Otherwise there is a high degree of page faults Increase the Page Size This may lead to an increase in fragmentation as not all applications require a large page size Provide Multiple Page Sizes This allows applications that require larger page sizes the opportunity to use them without an increase in fragmentation The default page size on Linux systems is 4 KB; however, Linux also provides huge pages, a feature that designates a region of physical memory where larger pages (for example, 2 MB) may be used.

Other Issues – Program Structure int[128,128] data; Each row is stored in one page Program 1 for (j = 0; j <128; j++) for (i = 0; i < 128; i++) data[i,j] = 0; 128 x 128 = 16,384 page faults Program 2 for (i = 0; i < 128; i++) for (j = 0; j < 128; j++) data[i,j] = 0; 128 page faults

Other Issues – I/O interlock I/O Interlock – Pages must sometimes be locked into memory Consider I/O - Pages that are used for copying a file from a device must be locked from being selected for eviction by a page replacement algorithm Pinning of pages to lock into memory

Operating System Examples Linux Windows Solaris

Linux Linux uses demand paging, allocating pages from a list of free frames and uses a global page-replacement policy Two types of page lists: an active_list and an inactive_list active_list – pages in use; inactive_list – not recently referenced Each page has an accessed bit that is set whenever the page is referenced. When a page is first allocated, accessed bit is set, added to rear of active_list When a page in active_list is referenced, accessed bit is set, moved to rear of active_list Over time, the least recently used page will be at the front of the active_list, from there, it may migrate to the rear of the inactive_list. If a page in the inactive_list is referenced, it moves back to the rear of the active_list.

Windows Uses demand paging with clustering. Clustering brings in pages surrounding the faulting page For a data page, a cluster contains three pages (the page before and the page after the faulting page); all other page faults have a cluster size of seven. Processes are assigned working set minimum (50 pages) and working set maximum (345 pages) Working set minimum is the minimum number of pages the process is guaranteed to have in memory A process may be assigned as many pages up to its working set maximum Windows uses the LRU-approximation clock algorithm, with a combination of local and global page-replacement policies. When the amount of free memory in the system falls below a threshold, automatic working set trimming is performed to restore the amount of free memory Working set trimming removes pages from processes that have pages in excess of their working set minimum

Solaris Maintains a list of free pages to assign faulting processes Lotsfree – threshold parameter (amount of free memory) to begin paging Desfree – threshold parameter to increasing paging Minfree – threshold parameter to being swapping Paging is performed by pageout process Pageout scans pages using modified clock algorithm Scanrate is the rate at which pages are scanned. This ranges from slowscan to fastscan Pageout is called more frequently depending upon the amount of free memory available Priority paging gives priority to process code pages

Solaris 2 Page Scanner

End of Chapter 10