Chapter 9: Memory Management

Similar documents
Memory Management Outline. Background Swapping Contiguous Memory Allocation Paging Segmentation Segmented Paging

The Deadlock Problem. Deadlocks. Deadlocks. Bridge Crossing Example

Memory management basics (1) Requirements (1) Objectives. Operating Systems Part of E1.9 - Principles of Computers and Software Engineering

OPERATING SYSTEM - MEMORY MANAGEMENT

Chapter 7 Memory Management

Chapter 7 Memory Management

OPERATING SYSTEMS MEMORY MANAGEMENT

Memory unit sees only the addresses, and not how they are generated (instruction counter, indexing, direct)

Operating Systems, 6 th ed. Test Bank Chapter 7

Memory Management 1. Memory Management. Multitasking without memory management is like having a party in a closet.

Virtual vs Physical Addresses

W4118: segmentation and paging. Instructor: Junfeng Yang

Technical Properties. Mobile Operating Systems. Overview Concepts of Mobile. Functions Processes. Lecture 11. Memory Management.

CS5460: Operating Systems

CS 61C: Great Ideas in Computer Architecture Virtual Memory Cont.

& Data Processing 2. Exercise 3: Memory Management. Dipl.-Ing. Bogdan Marin. Universität Duisburg-Essen

Operating System Tutorial

COS 318: Operating Systems. Virtual Memory and Address Translation

Two Parts. Filesystem Interface. Filesystem design. Interface the user sees. Implementing the interface

Chapter 12. Paging an Virtual Memory Systems

Advanced Computer Architecture-CS501. Computer Systems Design and Architecture 2.1, 2.2, 3.2

Page 1 of 5. IS 335: Information Technology in Business Lecture Outline Operating Systems

a storage location directly on the CPU, used for temporary storage of small amounts of data during processing.

Chapter 11: File System Implementation. Chapter 11: File System Implementation. Objectives. File-System Structure

Memory management. Chapter 4: Memory Management. Memory hierarchy. In an ideal world. Basic memory management. Fixed partitions: multiple programs

Chapter 3: Operating-System Structures. System Components Operating System Services System Calls System Programs System Structure Virtual Machines

Board Notes on Virtual Memory

Segmentation Segmentation: Generalized Base/Bounds

Traditional IBM Mainframe Operating Principles

x64 Servers: Do you want 64 or 32 bit apps with that server?

Lecture 17: Virtual Memory II. Goals of virtual memory

Memory Allocation. Static Allocation. Dynamic Allocation. Memory Management. Dynamic Allocation. Dynamic Storage Allocation

How To Write A Page Table

Chapter 13: Query Processing. Basic Steps in Query Processing

Stack Allocation. Run-Time Data Structures. Static Structures

OPERATING SYSTEM - VIRTUAL MEMORY

Virtual Memory. Virtual Memory. Paging. CSE 380 Computer Operating Systems. Paging (1)

This tutorial will take you through step by step approach while learning Operating System concepts.

Peter J. Denning, Naval Postgraduate School, Monterey, California

Chapter 3 Operating-System Structures

CHAPTER 17: File Management

Chapter 11: File System Implementation. Operating System Concepts with Java 8 th Edition

Chapter 6, The Operating System Machine Level

1 File Management. 1.1 Naming. COMP 242 Class Notes Section 6: File Management

CPU Scheduling. CPU Scheduling

Convenience: An OS makes a computer more convenient to use. Efficiency: An OS allows the computer system resources to be used in an efficient manner.

Chapter 2 Basic Structure of Computers. Jin-Fu Li Department of Electrical Engineering National Central University Jungli, Taiwan

Chapter 1 Computer System Overview

Chapter 11 I/O Management and Disk Scheduling

Objectives. Chapter 5: Process Scheduling. Chapter 5: Process Scheduling. 5.1 Basic Concepts. To introduce CPU scheduling

A3 Computer Architecture

Secondary Storage. Any modern computer system will incorporate (at least) two levels of storage: magnetic disk/optical devices/tape systems

Computer Architecture

Lecture 25 Symbian OS

Outline: Operating Systems

Operating Systems CSE 410, Spring File Management. Stephen Wagner Michigan State University

CS 377: Operating Systems. Outline. A review of what you ve learned, and how it applies to a real operating system. Lecture 25 - Linux Case Study

Introduction. What is an Operating System?

Introduction to Virtual Machines

361 Computer Architecture Lecture 14: Cache Memory

Chapter 14 MS-DOS Operating System

Objectives. Chapter 5: CPU Scheduling. CPU Scheduler. Non-preemptive and preemptive. Dispatcher. Alternating Sequence of CPU And I/O Bursts

An Introduction to Computer Science and Computer Organization Comp 150 Fall 2008

Objectives. Chapter 2: Operating-System Structures. Operating System Services (Cont.) Operating System Services. Operating System Services (Cont.

Operating System Overview. Otto J. Anshus

Example of Standard API

Record Storage and Primary File Organization

Chapter 11: File System Implementation. Operating System Concepts 8 th Edition

Data Storage - I: Memory Hierarchies & Disks

ELEC 377. Operating Systems. Week 1 Class 3

Physical Data Organization

Operating Systems. Steven Hand. Michaelmas / Lent Term 2008/ lectures for CST IA. Handout 3. Operating Systems N/H/MWF@12

1 Organization of Operating Systems

Disk Space Management Methods

Uses for Virtual Machines. Virtual Machines. There are several uses for virtual machines:

Operating System Components

Computer-System Architecture

Memory Management CS 217. Two programs can t control all of memory simultaneously

Rethinking SIMD Vectorization for In-Memory Databases

FAT32 vs. NTFS Jason Capriotti CS384, Section 1 Winter Dr. Barnicki January 28, 2000

Part III Storage Management. Chapter 11: File System Implementation

Software: Systems and Application Software

Logical Operations. Control Unit. Contents. Arithmetic Operations. Objectives. The Central Processing Unit: Arithmetic / Logic Unit.

Chapter 2 Data Storage

MACHINE ARCHITECTURE & LANGUAGE

ICS Principles of Operating Systems

Chapter 2 System Structures

Performance Monitoring User s Manual

Chapter 10: Virtual Memory. Lesson 03: Page tables and address translation process using page tables

We r e going to play Final (exam) Jeopardy! "Answers:" "Questions:" - 1 -

Computer Organization and Architecture. Characteristics of Memory Systems. Chapter 4 Cache Memory. Location CPU Registers and control unit memory

Chapter 13 File and Database Systems

Chapter 13 File and Database Systems

218 Chapter 5 CPU Scheduling

Unit Storage Structures 1. Storage Structures. Unit 4.3

Full and Para Virtualization

Operating Systems. Virtual Memory

File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System

Bindel, Spring 2010 Applications of Parallel Computers (CS 5220) Week 1: Wednesday, Jan 27

Transcription:

Chapter 9: Memory Management Background Logical versus Physical Address Space Overlays versus Swapping Contiguous Allocation Paging Segmentation Segmentation with Paging Operating System Concepts 9.1 Background Program must be brought into memory and placed within a process for it to be executed. Input queue collection of processes on the disk that are waiting to be brought into memory for execution. Select one process from input queue to load it into memory for execution. After execution (terminates) memory space becomes available. The address space of the computer starts at 00000. The first address of the user process does not need to be 00000. Operating System Concepts 9.2 1

Background User programs go through several steps before being executed: Source Program Compiler or Assembler Object Module Compile time Load time Execution time Linkage Editor Load Module Loader In-Memory Binary Memory Image Other object modules System Library Dynamically loaded system library Operating System Concepts 9.3 Binding of Instructions and Data to Memory Address binding of instructions and data to memory addresses can happen at three different stages: Compile time: If memory location known a priori, absolute code can be generated; must recompile code if starting location changes. Load time: Must generate relocatable code if memory location is not known at compile time. Execution time: Binding delayed until run time if the process can be moved during its execution from one memory segment to another. Need hardware support for address maps (e.g., base and limit registers). Operating System Concepts 9.4 2

Logical vs. Physical Address Space The concept of a logical address space that is bound to a separate physical address space is central to proper memory management. Logical address generated by the CPU; also referred to as virtual address. Physical address address seen by the memory unit. Logical and physical addresses are the same in compiletime and load-time address-binding schemes; logical (virtual) and physical addresses differ in execution-time address-binding scheme. The set of all logical address generated by a program. Operating System Concepts 9.5 Memory-Management Unit (MMU) The run-time mapping from virtual to physical addresses is done by the memory-management unit (MMU), which is a hardware device. The user program deals with logical addresses; it never sees the real physical addresses. Two different types of addresses: Logical: range 0 to max. Physical: range R+0 to R+max; where R is a base value. The user generates only logical addresses and thinks that the process runs in locations 0 to max. Operating System Concepts 9.6 3

Memory-Management Unit (MMU) (Cont.) In MMU scheme, the value in the relocation register is added to every address generated by a user process at the time it is sent to memory. Example: Dynamic relocation using a relocation register. Logical address 345 Relocation register 1400 Physical address 1745 Memory CPU + MMU Operating System Concepts 9.7 Overlays The entire program and data of a process must be in the physical memory for the process to execute. The size of a process is limited to the size of physical memory. If a process is larger than the amount of memory, a technique called overlays can be used. Overlays is to keep in memory only those instructions and data that are needed at any given time. When other instructions are needed, they are loaded into space that was occupied previously by instructions that are no longer needed. Overlays are implemented by user, no special support needed from operating system, programming design of overlay structure is complex. Operating System Concepts 9.8 4

Overlays Example Example: Consider a two-pass assembler. Pass1 constructs a symbol table. Pass2 generates machine-language code. Assume the following: Size (k = 1024 bytes) Pass1 Pass2 Symbol table Common routines Total size 70k 80k 20k 30k 200k To load everything at once, we need 200k of memory. Operating System Concepts 9.9 Overlays Example (Cont.) If only 150K is available, we cannot run our process. Notice that Pass1 and Pass2 do not need to be in memory at same time. So, we define two overlays: Overlay A: symbol table, common routines, and Pass1. Overlay B: symbol table, common routines, and Pass2. We add overlay driver 10k and start with overlay A in memory. When finish Pass1, we jump to overlay driver, which reads overlay B into memory overwriting overlay A and transfer control to Pass2. Operating System Concepts 9.10 5

Overlays Example (Cont.) Overlay A needs 130k and Overlay B needs 140k. Symbol table Common routines Overlay driver Pass1 (70k) Pass2 (80k) Operating System Concepts 9.11 Swapping A process can be swapped temporarily out of memory to a backing store, and then brought back into memory for continued execution. Backing store fast disk large enough to accommodate copies of all memory images for all users; must provide direct access to these memory images. Roll out, roll in swapping variant used for priority-based scheduling algorithms; lower-priority process is swapped out so higher-priority process can be loaded and executed. Normally, a process that is swapped out will be swapped back into the same memory space that it occupied previously. Example: In a multiprogramming environment with Round Robin CPU scheduling, when time slice expires, the memory manager swap out the process that just finished, and swap in another process to the memory space that has been freed. Operating System Concepts 9.12 6

Major part of swap time is transfer time; total transfer time is directly proportional to the amount of memory swapped. The context-switch time in swapping system is high. Modified versions of swapping are found on many systems, i.e., UNIX and Microsoft Windows. Schematic View of Swapping Operating System Concepts 9.13 Example 1: Swapping Let Process P1 size is 100KB and the transfer rate of a hard disk is 1MB/second. To transfer P1 (100KB) to or from memory takes 100K/1000K per second, which is 1/10 second = 100 milliseconds. Assume the average latency is 8 milliseconds, then to swap in or out takes 108 milliseconds. The total swap (in and out) is 216 milliseconds for 100KB process. For efficient CPU utilization, we want our execution time for each process to be long relative to the swap time. A Round Robin scheduling algorithm, the time slice should be larger than 216 milliseconds (from the above example). Operating System Concepts 9.14 7

Contiguous Allocation Main memory usually is divided into two partitions: For the resident operating system For the user processes. 0 100K O.S. User 1MB Operating System Concepts 9.15 Example 2: Swapping A computer system has 1MB of main memory. The O.S. takes 100KB. Maximum size of user process is 900KB. User process may be smaller than 900KB. From previous example if process P1 is 100K the swap time is 108 milliseconds. But, if P1 is 900K then the swap time is 908 milliseconds. As the size of a process increases the swap time increases too. 0 100K 1MB O.S. 100K User 900K Main Memory Operating System Concepts 9.16 8

Single Partition Allocation Single-partition allocation Needs Protection. Protect O.S. code and data from changes by the user processes. Protect user processes from one another. We can provide protection by using a relocationregister with a limit register. Relocation register contains value of smallest physical address. Limit register contains range of logical addresses. Each logical address must be less than the limit register. Operating System Concepts 9.17 Hardware Support for Relocation & Limit Registers Limit Register Relocation Register CPU < Logical Address No Yes + Physical Address Memory Trap, Address Error Operating System Concepts 9.18 9

Multiple - Partition Allocation Several user processes residing in memory at the same time. Memory can be divided into a number of fixed-sized partitions, where each partition may contain exactly one process. Therefore, the degree of multiprogramming is bound by the number of partitions. Or all memory is available for user processes as one large block (hole). When a process arrives and needs memory, we search for a hole large enough for this process. If we find a space, we allocate only as much memory as is needed, keeping the rest available to satisfy future requests. Operating System Concepts 9.19 Multiple - Partition Allocation (Cont.) Hole block of available memory; holes of various size are scattered throughout memory. Operating system maintains information about: a) allocated partitions b) free partitions (hole) OS OS OS OS process 5 process 5 process 5 process 5 process 9 process 9 process 8 process 10 process 2 process 2 process 2 process 2 Operating System Concepts 9.20 10

Multiple - Partition Allocation (Cont.) When no available block of memory (hole) is large enough to hold process, the O.S. waits until a large block is available. In general, there is at any time a set of holes, of various sizes, scattered throughout memory. When a process arrives and needs memory, we search this set for a hole that is large enough for this process. If the hole is too large, it is split into two: One part is allocated to the arriving process. The other is returned to the set of holes. When a process terminates, it releases its block of memory, which is then placed back in the set of holes. If the new hole is adjacent to other holes, we merge these adjacent holes to form one larger hole. Operating System Concepts 9.21 Dynamic Storage-Allocation Problem First-fit, best-fit, and worst fit are the most common strategies used to select a free hole from the set of available holes. First-fit: Allocate the first hole that is big enough. Searching starts at the beginning of the set of holes. We can stop searching as soon as we find a free hole that is large enough. Best-fit: Allocate the smallest hole that is big enough; must search entire list, unless ordered by size. Produces the smallest leftover hole. Worst-fit: Allocate the largest hole; must also search entire list, unless ordered by size. Produces the largest leftover hole, which may be more useful than the smaller leftover hole from best-fit approach. First-fit and best-fit better than worst-fit in terms of speed and storage utilization respectively. Operating System Concepts 9.22 11

Fragmentation External fragmentation total memory space exists to satisfy a request, but it is not contiguous; storage is fragmented into a large number of small holes. Example: We have a total external fragmentation of (300+260)=560KB. If P5 is 500KB, then this space would be large enough to run P5. But the space is not contiguous. The selection of first-fit versus best-fit can affect the amount of fragmentation. 0 400K 1000K 1700K 2000K 2300K 2560K O.S. P1 P4 Free P3 Free Operating System Concepts 9.23 Fragmentation (Cont.) Internal fragmentation Memory that is internal to a partition, but is not being used, because it is too small. Example: Assume next request is for 18462 bytes. If we allocate exactly the requested block, we are left with a hole of 2 bytes. The overhead to keep track of this hole will be larger than the hole itself. So, we ignore this small hole (internal fragmentation). Hole of 18464 bytes O.S. P7 Free P4 Operating System Concepts 9.24 12

Fragmentation (Cont.) One solution to the problem of external fragmentation is compaction. The goal is to shuffle memory contents to place all free memory together in one large block. Example: (Compaction) 3 holes compacted into one large hole 660KB. 0 400K 900K 1000K 1700K 2000K 2300K 2560K O.S. P5 Free 100K P4 Free 300K P3 Free 260K Compact 0 400K 900K 1600K 1900K 2560K O.S. P5 P4 P3 Free 660K Operating System Concepts 9.25 Fragmentation (Cont.) The simplest compaction algorithm is to move all processes toward one end of memory; all holes move in the other direction, producing one large hole of available memory. Example: Comparison of different ways to compact memory. 0 300 500 600 1000 1200 1500 1900 2100 O.S. P1 P2 Free 400 P3 Free 300 P4 Free 200 A) Original Allocation 0 300 500 600 800 1200 2100 Operating System Concepts 9.26 O.S. P1 P2 P3 P4 Free 900 B) Moved 600 If we use the simple algorithm, we must move P3 and P4 for a total of 600K moved. 13

Fragmentation (Cont.) Example (Cont.): Or, we can move P4 above P3 moving only 400K. 0 O.S. 0 O.S. 300 500 600 1000 1200 1500 1900 2100 P1 P2 Free 400 P3 Free 300 P4 Free 200 A) Original Allocation 300 500 600 1000 1200 2100 P1 P2 P4 P3 Free 900 C) Moved 400 Operating System Concepts 9.27 Fragmentation (Cont.) Example (Cont.): Or, we can move P3 below P4 moving only 200K. 0 300 500 600 1000 1200 1500 1900 2100 O.S. P1 P2 Free 400 P3 Free 300 P4 Free 200 A) Original Allocation 0 300 500 600 1500 1900 2100 O.S. P1 P2 Free 900 P4 P3 D) Moved 200 Swapping can also be combined with compaction; a process can be swapped out of main memory to a backing store and swapped in again later. Operating System Concepts 9.28 14

Paging Logical address space of a process can be noncontiguous; process is allocated physical memory whenever the latter is available. Divide physical memory into fixed-sized blocks called frames (size is power of 2, between 512 bytes and 8192 bytes). Divide logical memory into blocks of same size called pages. To run a program of size n pages, need to find n free frames and load program. When a process is to be executed, its pages are loaded into any available memory frame from the backing store. The backing store is divided into fixed-sized blocks that are of the same size as the memory frames. Operating System Concepts 9.29 Address Translation Scheme Set up a page table to translate logical to physical addresses. Address generated by CPU is divided into: Page number (p) used as an index into a page table which contains base address of each page in physical memory. Page offset (d) combined with base address to define the physical memory address that is sent to the memory unit. Operating System Concepts 9.30 15

Address Translation Architecture Operating System Concepts 9.31 Paging Example of Logical & Physical Memory Operating System Concepts 9.32 16

Paging Example for a 32-byte Memory with 4-byte each Page We have 4 pages and each page is 4 bytes. Logical address 0 is page 0 and offset 0. Page 0 is in frame 5. Logical address maps to physical address (5x4)+0=20. Logical address 3 (page 0, offset 3) maps to physical address (5x4)+3=23. Logical address 4 (page 1, offset 0) maps to physical address (6x4)+0=24. Logical address 13 maps to physical address (2x4)+1=9. 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Logical Memory a b c d e f g h i j k l m n o p 0 5 1 6 2 1 3 2 Page Table 0 4 8 i j k l m no p 12 16 20 a b c d 24 e f g h 28 Physical Memory Operating System Concepts 9.33 Example: Internal Fragmentation in Paging When we use a paging scheme, we have no external fragmentation, but we may have some internal Fragmentation. If pages are 2048 bytes, a process of 72,766 bytes would need 35 pages plus 1086 bytes. It would be allocated 36 frames resulting in an internal fragmentation of 2048-1086=962 bytes. In worst case, a process would need n pages plus one byte. It would be allocated n+1 frames, resulting in an internal fragmentation of almost an entire frame. Operating System Concepts 9.34 17

Example: Free Frames Free-Frame List 14 13 18 20 15 Page 0 Page 1 Page 2 Page 3 New Process 13 Free 14 Free 15 Free 16 17 18 Free 19 20 Free 21 Before Allocation Free-Frame List 15 Page 0 Page 1 Page 2 Page 3 New Process 0 1 2 3 14 13 18 20 New Process Page Table 13 Page 1 14 Page 0 15 Free 16 17 18 Page 2 19 20 Page 3 21 After Allocation Operating System Concepts 9.35 Page Table In general, each page of the process needs one frame. If the process requires n pages, there must be at least n frames available in memory. If there are n frames available, they are allocated to this arriving process. The first page of the process is loaded into one of the allocated frames, and the frame number is put in the page table for this process. The next page is loaded into another frame, and its frame number is put into the page table and so on. Operating System Concepts 9.36 18

Frame Table The operating system is managing physical memory, it must be aware of the allocation details of physical memory: Which frames are allocated. Which frames are available. How many total frames there are and so on. This information is kept in a data structure called a frame table. Frame table has one entry for each physical page frame, indicating whether the latter is free or allocated and if it is allocated, to which page of which process or processes. Clear separation between the user s view of memory and the actual physical memory. Paging increases the context-switch time. Operating System Concepts 9.37 Structure of the Page Table Most operating systems allocate a page table for each process. A pointer to the page table is sorted with other register values (like the instruction counter) in the process control block (PCB). Page table is kept in main memory. Page-table base register (PTBR) points to the page table. Page-table length register (PTLR) indicates size of the page table. Simplest way to implement a page table is as a set of dedicated very high-speed registers for small page table (256 entries). Most computers allow the page table to be very large. For this machines, the use of fast registers to implement the page table is not feasible. Operating System Concepts 9.38 19

Translation Look-aside Buffers (TLBs) In this scheme every data/instruction access requires two memory accesses. One for the page table and one for the data/instruction. The two memory access problem can be solved by the use of a special fast-lookup hardware cache called associative registers or translation look-aside buffers (TLBs). In TLBs, each register consists of two parts: a key and a value. When the TLBs are presented with an item, it is compared with all keys simultaneously. If the item is found, the corresponding value field is output. The search is fast, but hardware is expensive. The number of entries in a TLB varies between 8 and 2048. Operating System Concepts 9.39 Paging Hardware with TLB Logical address CPU p d Page number TLB TLB miss Frame number p TLB hit ---- ---- ---- f ---- Page Table f d Physical address Physical Memory Operating System Concepts 9.40 20

Translation Look-aside Buffers (TLBs) (Cont.) When a logical address is generated by the CPU, its page number is presented to a set of associate registers (TLB) that contain page numbers and their corresponding frame numbers. If the page number is found in the TLB, its frame number is immediately available and is used to access memory. If the page number is not in the TLB, a memory reference to the page table must be made. We add the page number and frame number to TLB, so that will be found quickly on the next reference. If the TLB is already full of entries the operating system must select one for replacement. Every time a new page table is selected the TLB must be flushed (erased) to ensure that next executing process does not use wrong translation information. Operating System Concepts 9.41 TLB s Hit Ratio The percentage of times that a page number is found in the TLB is called the hit ratio. An 80-percent hit ratio means that we find the desired page number in the TLB 80 percent of the time. Example: For 80-percent hit ratio. If it takes 20 nanoseconds to search TLB and 100 nanoseconds to access memory, then a mapped memory access takes 120 nanoseconds when the page number is in TLB. If we fail to find the page number in TLB (20 nanoseconds), then we must first access memory for the page table and frame number (100 nanoseconds), then access the desired byte in memory (100 nanoseconds), for a total of 220 nanoseconds. Effective access time = 0.80x120+0.20x220 = 140 nanoseconds. In this example, we suffer a 40-percent slowdown in memory access time (from 100 to 140 nanoseconds). Operating System Concepts 9.42 21

TLB s Hit Ratio (Cont.) Example: For a 98-percent hit ratio: Effective access time=0.98x120+0.02x220=122 nanoseconds. This increased hit ratio produces only a 22-percent slowdown in memory access time. The hit ratio is clearly related to the number of associative registers (TLBs). Motorola 68030 processor (used in Apple Macintosh systems) has a 22-entry TLB. Intel 80486 CPU has 32 registers and claims a 98-percent hit ratio. In general, number of entries in TLB ranging from 16-512, a hit ratio of 80 to 98 percent can be obtained. Operating System Concepts 9.43 Memory Protection Memory protection implemented by associating protection bits with each frame. These bits are kept in the page table. One bit can be define a page to be read and write or read-only. Every reference to memory goes through the page table to find the correct frame number and at the same time the protection bits can be checked to verify that no writes are being made to a read-only page. An attempt to write to a read-only page causes a hardware trap to the operating system (memory protection violation). Valid-invalid bit attached to each entry in the page table: valid indicates that the associated page is in the process s logical address space, and is thus a legal page. invalid indicates that the page is not in the process s logical address space. Illegal addresses are trapped by using valid-invalid bit. The O.S. sets this bit for each page to allow or disallow accesses to that page. Operating System Concepts 9.44 22

Example: Memory Protection A system with 14 bit address space (0 to 2 14 = 16,383). A program use only addresses 0 to 12,287. A page size is 2KB (2048). 00000 10,468 12,287 Page 0 Page 1 Page 2 Page 3 Page 4 Page 5 Page Number 0 1 2 3 4 5 6 7 Frame number Validinvalid bit 2 v 3 v 4 v 7 v 8 v 9 v 0 i 0 i Page Table 0 1 2 3 4 5 6 7 8 9 Page 0 Page 1 Page 2 Page 3 Page 4 Page 5 Page n Operating System Concepts 9.45 Example: Memory Protection (Cont.) Addresses in pages 0, 1, 2, 3, 4, and 5 are mapped normally through page table. Any attempt to generate an address in pages 6 and 7 finds that the valid-invalid bit is set to invalid, and the computer will trap to the operating system (invalid page reference). Notice that references to page 5 are classified as valid, so accesses to addresses up to 12,287 are valid. Only addresses from 12,288 to 16,383 are invalid. Operating System Concepts 9.46 23

Multilevel Paging Most modern computer systems support a very large logical address space. The page table is very large. Example: A system with a 32-bit logical address space. If page size is 4K bytes (2 12 ), then a page table may consist of up to 1 million entries (2 32 / 2 12 ) = (2 20 ). Because each entry consists of 4 bytes (32 bits) and we have one million entries, each process may need up to 4 megabytes of physical address space for the page table alone. A simple solution is to divide the page table into small pieces. One way is to use a two-level paging scheme (the page table itself is also paged). Operating System Concepts 9.47 Two-Level Page-Table Scheme Operating System Concepts 9.48 24

Two-Level Paging Example A logical address (on 32-bit machine with 4K (2 12 ), page size) is divided into: a page number consisting of (32-12)=20 bits. a page offset consisting of 12 bits. Since the page table is paged, the page number is further divided into: a 10-bit page number. a 10-bit page offset. Thus, a logical address is as follows: page number page offset p i p 2 d 10 10 12 where p i is an index into the outer page table, and p 2 is the displacement within the page of the outer page table. Operating System Concepts 9.49 Address-Translation Scheme Address-translation scheme for a two-level 32-bit paging architecture. Operating System Concepts 9.50 25

Multilevel Paging and Performance The SPARC machine with 32-bit addressing supports 3-level paging scheme, and the 32-bit Motorola 68030 supports 4-level paging scheme. How does multilevel paging affect system performance? For example, in 4-level paging scheme, each level is stored as a separate table in memory, it will take 4 memory accesses. If we use cache memory, let hit ratio to be 98 and it takes 20 nanoseconds to search TLB and 100 nanoseconds to access memory, then the effective access time will be: 0.98 x 120 + 0.02 x 520 = 128 nanoseconds. which is only a (128-100) 28 percent slowdown in memory access time. Operating System Concepts 9.51 Inverted Page Table Each process has a page table associated with it. A disadvantage is that each page table may consist of millions of entries.therefore, these tables may consume large amounts of physical memory, which is required just to keep track of how the other physical memory is being used. To solve this problem, we can use an inverted page table. An inverted page table has one entry for each real page (frame) of memory. Each entry consists of the virtual (logical) address of the page stored in that real memory location, with information about the process that owns that page. So, there is only one page table in the system, and it has only one entry for each page of physical memory. The inverted-page table decreases memory needed to store each page table, but increases time needed to search the table when a page reference occurs. To improve searching the table we can use hash table. Operating System Concepts 9.52 26

Inverted Page Table Architecture Examples of systems using this scheme: - IBM RISC Systems 6000 -IBM RT - HP Spectrum workstations. Operating System Concepts 9.53 Shared Pages Another advantage of paging is sharing common code. Shared code One copy of read-only (reentrant) code shared among processes (i.e., text editors, compilers, window systems). Shared code must appear in same location in the logical address space of all processes. Private code and data Each process keeps a separate copy of the code and data. The pages for the private code and data can appear anywhere in the logical address space. Operating System Concepts 9.54 27

Shared Pages (Cont.) Example: A system supports 40 users, each executes a text editor. If the text editor consists of 150K of code and 50K of data space, we would need 8000K to supports the 40 users. That is (150K + 50K) x 40 = 8000K If the code is reentrant (it can be shared), then we need only one copy of the editor (150K) plus 40 copies of the 50K of data space per user. The total space required is now (150K + 40 x 50K) 2150K instead of 8000K (a significant saving of memory). Other used programs can be shared: Compilers; Window Systems; Database Systems To be shared, the code must be reentrant (nonmodifiable). Operating System Concepts 9.55 Shared Pages Example Operating System Concepts 9.56 28

Shared Pages Example (Cont.) Note that frames 3, 4, and 6 were shared by P1, P2, and P3. Reentrant code (sharable code) is non-modifiable code. If the code is reentrant, then it never changes during execution. Two or more processes can execute the same code at the same time. As you can see from the example, only one copy of editor needs to be kept in physical memory. Operating System Concepts 9.57 Segmentation The user s view of memory is not the same as the actual physical memory. A program is a collection of segments. A segment is a logical unit such as: main program, procedure, function, local variables, global variables, common block, stack, symbol table, arrays Operating System Concepts 9.58 29

Logical View of Segmentation User s view of a program as segments. Segmentation is a memory-management scheme that supports this user view of memory. A logical address space is a collection of segments. Subroutine1 Symbol Table3 Stack2 Main Program4 1 4 2 3 user space physical memory space Operating System Concepts 9.59 Segmentation Architecture Logical address consists of a two tuple: <segment-number, offset> Mapping 2-dimensional user-defined addresses into 1- dimensional physical addresses, by using segment table. Each entry of the segment table has a segment base and a segment limit. The segment base contains the starting physical address where the segment resides in memory. The segment limit specifies the length of the segment. The segment table is an array of base-limit register pairs. Operating System Concepts 9.60 30

Segmentation Hardware s Segment Table CPU limit base s d Logical Address < No Yes + Physical Address Physical Memory Trap, Address Error Operating System Concepts 9.61 Subroutine0 Symbol Table4 Segmentation Example Stack3 Main Program2 Sqrt1 Logical Address Space Operating System Concepts 9.62 0 1 2 3 4 Limit 1000 400 400 1100 1000 base 1400 6300 4300 3200 4700 Segment Table Segment2 is 400 bytes long and begins at location 4300. So, a reference to byte 53 of segment2 is mapped onto location 4300+53=4353. However, a reference to byte 1222 of segment0 would result in a trap to the O.S. because segment0 is 1000 bytes long. 1400 2400 3200 4300 4700 5700 6300 Segment 0 Segment 3 Segment 2 Segment 4 6700 segment 1 Physical memory 31

Implementation of Segment Tables Like the page table, the segment table can be put either in fast registers or in memory. Segment-table base register (STBR) points to the segment table s location in memory. Segment-table length register (STLR) indicates number of segments used by a program. For logical address (s, d): Check the segment number (s) is legal if (s < STLR). Then add the segment number to STBR resulting in the address (STBR+s) in memory of the segment table entry. As before, now check the offset against the segment length and compute the physical address of the desired byte as the sum of the segment base and offset. This mapping requires two memory references per logical address, slowing the computer system by a factor of 2. Solution: use a set of associative registers to hold most recently used segment-table entries. Operating System Concepts 9.63 Segmentation: Protection & Sharing Some segments are instructions and other segments are data. Instruction segment can be defined as read-only or execute only. Protection: With each entry in segment table associate: validation bit = 0 illegal segment read/write/execute privileges. Memory-mapping hardware will check the protection bits associated with each segment-table entry to prevent illegal accesses to memory, such as attempts to write into a read-only segment, or to use an execute-only segment as data. Another advantage of segmentation is sharing of code or data. Each process has a segment table associated with it. Segments are shared when entries in the segment tables of two different processes point to the same physical memory. Operating System Concepts 9.64 32

Example: Sharing of Segments Operating System Concepts 9.65 Segmentation: Fragmentation The long-term scheduler (selecting a process from disk into memory) find and allocate memory for all the segments of a user program. Segments are of variable length; whereas pages are of same length. With the variable-sized partition scheme, memory allocation is a dynamic storage-allocation problem, usually solved with a best-fit or first-fit algorithm. Segmentation may cause external fragmentation; whereas paging may cause internal fragmentation. External fragmentation, when all blocks of free memory are too small to accommodate a segment. In this case, the process may have to wait until more memory becomes available. Compaction may be used to create a larger hole. Operating System Concepts 9.66 33

Average Segment Size One extreme: each process to be a segment reduces variable-sized partition scheme. Other extreme: every byte could be put in its own segment, this eliminates external fragmentation. Generally, if average segment size is small external fragmentation will be also small. Operating System Concepts 9.67 Segmentation with Paging MULTICS Paging scheme causes internal fragmentation; whereas segmentation scheme causes external fragmentation. A better solution is to combine both; where each segment vary in size and has a different number of pages. All pages in all segments have same size. Segmentation with paging scheme eliminates external fragmentation. As with paging, the last page of each segment generally will not be completely full (internal fragmentation). The MULTICS system solved problems of external fragmentation and lengthy search times by paging the segments. Operating System Concepts 9.68 34

Segmentation with Paging MULTICS (Cont.) Logical address in MULTICS: Segment # offset s d 18 bits 16 bits Each page in MULTICS consists of 1K words; that is: Segment # s 18 bits Offset (d) p d 6 bits 10 bits Solution differs from pure segmentation in that the segmenttable entry contains not the base address of the segment, but rather the base address of a page table for this segment. Operating System Concepts 9.69 MULTICS Address Translation Scheme Operating System Concepts 9.70 35

Segmentation with Paging MULTICS (Cont.) In MULTICS the segment number is 18-bit, we could have up to 2 18 = 262,144 segments, requiring a large segment table. To solve this problem, MULTICS pages the segment table. The segment number (18 bits) is broken into an 8-bit page number and 10-bit page offset. The segment table is represented by a page table consisting of up to 2 8 entries. So, a logical address in MULTICS is as follows: Segment number s1 s2 8-bit 10-bit Offset d1 d2 6-bit 10-bit Operating System Concepts 9.71 Address Translation in MULTICS Segment number s1 s2 8-bit 10-bit s1 Offset d1 d2 6-bit 10-bit s2 Logical address Page table for segment table Page of segment table d1 Page table for segment d2 Page of segment Operating System Concepts 9.72 36

Performance in MULTICS To ensure reasonable performance: 16 associative register are available that contain the address of the 16 most recently referred pages. Each register consists of two parts: a key and a value. The key is 24-bit field that is the concatenation of a segment number and a page number. The value is the frame number. Segment number Offset s1 s2 d1 d2 8-bit 10-bit 6-bit 10-bit Page # Key 24-bit Operating System Concepts 9.73 Segmentation with Paging Intel 386 As shown in the following diagram, the Intel 386 uses segmentation with paging for memory management with a two-level paging scheme. Operating System Concepts 9.74 37

Intel 30386 address translation Operating System Concepts 9.75 Comparing Memory-Management Strategies Hardware support Performance Fragmentation Relocation Swapping Sharing Protection Operating System Concepts 9.76 38