SlideShare a Scribd company logo
1 of 65
Operating Systems:
Internals and Design Principles, 6/E
          William Stallings



         Chapter 8
      Virtual Memory



                   Patricia Roy
        Manatee Community College, Venice, FL
                ©2008, Prentice Hall
Hardware and Control
            Structures
• Memory references are dynamically
  translated into physical addresses at run
  time
  – A process may be swapped in and out of main
     memory such that it occupies different
    regions
Hardware and Control
           Structures
• A process may be broken up into pieces,
  which do not need to be located
  contiguously in main memory
• It is not necessary for all pieces of a
  process to be loaded in main memory
  during execution of the process
Execution of a Program
• Operating system brings into main
  memory a few pieces of the program
• Resident set - portion of process that is in
  main memory
• An interrupt is generated when an address
  is needed that is not in main memory
• Operating system places the process in a
  blocking state
Execution of a Program
• Piece of process that contains the logical
  address is brought into main memory
  – Operating system issues a disk I/O Read
    request
  – Another process is dispatched to run while the
    disk I/O takes place
  – An interrupt is issued when disk I/O complete
    which causes the operating system to place
    the affected process in the Ready state
Improved System Utilization
• More processes may be maintained in
  main memory
  – Only load in some of the pieces of each
    process
  – With so many processes in main memory, it is
    very likely a process will be in the Ready state
    at any particular time
• A process may be larger than all of main
  memory
Types of Memory
• Real memory
  – Main memory
• Virtual memory
  – Memory on disk
  – Allows for effective multiprogramming and
    relieves the user of tight constraints of main
    memory
Thrashing
• Swapping out a piece of a process just
  before that piece is needed
• The processor spends most of its time
  swapping pieces rather than executing
  user instructions
Principle of Locality
• Program and data references within a
  process tend to cluster
• Only a few pieces of a process will be
  needed over a short period of time
• Possible to make intelligent guesses
  about which pieces will be needed in the
  future
• This suggests that virtual memory may
  work efficiently
Support Needed for Virtual
           Memory
• Hardware must support paging and
  segmentation
• Operating system must be able to do the
  management the movement of pages
  and/or segments between secondary
  memory and main memory
Paging
• Each process has its own page table
• Each page table entry contains the frame
  number of the corresponding page in main
  memory
• A bit is needed to indicate whether the
  page is in main memory or not
Paging
Modify Bit in Page Table
• Modify bit is needed to indicate if the page
  has been altered since it was last loaded
  into main memory
• If no change has been made, the page
  does not have to be written to the disk
  when it needs to be replaced
Address Translation
Two-Level Hierarchical Page
          Table
Page Tables
• Page tables are also stored in virtual
  memory
• When a process is running, part of its
  page table is in main memory
Address Translation
Inverted Page Table
• Used on PowerPC, UltraSPARC, and IA-
  64 architecture
• Page number portion of a virtual address
  is mapped into a hash value
• Hash value points to inverted page table
• Fixed proportion of real memory is
  required for the tables regardless of the
  number of processes
Inverted Page Table
•   Page number
•   Process identifier
•   Control bits
•   Chain pointer
Inverted Page Table
Translation Lookaside Buffer
• Each virtual memory reference can cause
  two physical memory accesses
  – One to fetch the page table
  – One to fetch the data
• To overcome this problem a high-speed
  cache is set up for page table entries
  – Called a Translation Lookaside Buffer (TLB)
Translation Lookaside Buffer
• Contains page table entries that have
  been most recently used
Translation Lookaside Buffer
• Given a virtual address, processor
  examines the TLB
• If page table entry is present (TLB hit), the
  frame number is retrieved and the real
  address is formed
• If page table entry is not found in the TLB
  (TLB miss), the page number is used to
  index the process page table
Translation Lookaside Buffer
• First checks if page is already in main
  memory
  – If not in main memory a page fault is issued
• The TLB is updated to include the new
  page entry
Translation Lookaside Buffer
Translation Lookaside Buffer
Translation Lookaside Buffer
Translation Lookaside Buffer
Page Size
• Smaller page size, less amount of internal
  fragmentation
• Smaller page size, more pages required
  per process
• More pages per process means larger
  page tables
• Larger page tables means large portion of
  page tables in virtual memory
Page Size
• Secondary memory is designed to
  efficiently transfer large blocks of data so
  a large page size is better
Page Size
• Small page size, large number of pages
  will be found in main memory
• As time goes on during execution, the
  pages in memory will all contain portions
  of the process near recent references.
  Page faults low.
• Increased page size causes pages to
  contain locations further from any recent
  reference. Page faults rise.
Page Size
Example Page Size
Segmentation
• May be unequal, dynamic size
• Simplifies handling of growing data
  structures
• Allows programs to be altered and
  recompiled independently
• Lends itself to sharing data among
  processes
• Lends itself to protection
Segment Tables
• Starting address corresponding segment
  in main memory
• Each entry contains the length of the
  segment
• A bit is needed to determine if segment is
  already in main memory
• Another bit is needed to determine if the
  segment has been modified since it was
  loaded in main memory
Segment Table Entries
Segmentation
Combined Paging and
         Segmentation
• Paging is transparent to the programmer
• Segmentation is visible to the programmer
• Each segment is broken into fixed-size
  pages
Combined Paging and
   Segmentation
Address Translation
Protection Relationships
Fetch Policy
• Determines when a page should be
  brought into memory
• Demand paging only brings pages into
  main memory when a reference is made
  to a location on the page
  – Many page faults when process first started
• Prepaging brings in more pages than
  needed
  – More efficient to bring in pages that reside
    contiguously on the disk
Placement Policy
• Determines where in real memory a
  process piece is to reside
• Important in a segmentation system
• Paging or combined paging with
  segmentation hardware performs address
  translation
Replacement Policy
• Which page is replaced?
• Page removed should be the page least
  likely to be referenced in the near future
• Most policies predict the future behavior
  on the basis of past behavior
Replacement Policy
• Frame Locking
  – If frame is locked, it may not be replaced
  – Kernel of the operating system
  – Key control structures
  – I/O buffers
  – Associate a lock bit with each frame
Basic Replacement Algorithms
• Optimal policy
  – Selects for replacement that page for which
    the time to the next reference is the longest
  – Impossible to have perfect knowledge of
    future events
Basic Replacement Algorithms
• Least Recently Used (LRU)
  – Replaces the page that has not been
    referenced for the longest time
  – By the principle of locality, this should be the
    page least likely to be referenced in the near
    future
  – Each page could be tagged with the time of
    last reference. This would require a great
    deal of overhead.
Basic Replacement Algorithms
• First-in, first-out (FIFO)
  – Treats page frames allocated to a process as
    a circular buffer
  – Pages are removed in round-robin style
  – Simplest replacement policy to implement
  – Page that has been in memory the longest is
    replaced
  – These pages may be needed again very soon
Basic Replacement Algorithms
• Clock Policy
  – Additional bit called a use bit
  – When a page is first loaded in memory, the
    use bit is set to 1
  – When the page is referenced, the use bit is
    set to 1
  – When it is time to replace a page, the first
    frame encountered with the use bit set to 0 is
    replaced.
  – During the search for replacement, each use
    bit set to 1 is changed to 0
Clock Policy
Clock Policy
Clock Policy
Comparison
Behavior of Page Replacement
          Algorithms
Basic Replacement Algorithms
• Page Buffering
  – Replaced page is added to one of two lists
    • Free page list if page has not been modified
    • Modified page list
Resident Set Size
• Fixed-allocation
  – Gives a process a fixed number of pages
    within which to execute
  – When a page fault occurs, one of the pages of
    that process must be replaced
• Variable-allocation
  – Number of pages allocated to a process
    varies over the lifetime of the process
Fixed Allocation, Local Scope
• Decide ahead of time the amount of
  allocation to give a process
• If allocation is too small, there will be a
  high page fault rate
• If allocation is too large there will be too
  few programs in main memory
  – Processor idle time
  – Swapping
Variable Allocation, Global
              Scope
• Easiest to implement
• Adopted by many operating systems
• Operating system keeps list of free frames
• Free frame is added to resident set of
  process when a page fault occurs
• If no free frame, replaces one from
  another process
Variable Allocation, Local Scope
• When new process added, allocate
  number of page frames based on
  application type, program request, or other
  criteria
• When page fault occurs, select page from
  among the resident set of the process that
  suffers the fault
• Reevaluate allocation from time to time
Cleaning Policy
• Demand cleaning
  – A page is written out only when it has been
    selected for replacement
• Precleaning
  – Pages are written out in batches
Cleaning Policy
• Best approach uses page buffering
  – Replaced pages are placed in two lists
    • Modified and unmodified
  – Pages in the modified list are periodically
    written out in batches
  – Pages in the unmodified list are either
    reclaimed if referenced again or lost when its
    frame is assigned to another page
Load Control
• Determines the number of processes that
  will be resident in main memory
• Too few processes, many occasions when
  all processes will be blocked and much
  time will be spent in swapping
• Too many processes will lead to thrashing
Multiprogramming
Process Suspension
• Lowest priority process
• Faulting process
  – This process does not have its working set in
    main memory so it will be blocked anyway
• Last process activated
  – This process is least likely to have its working
    set resident
Process Suspension
• Process with smallest resident set
  – This process requires the least future effort to
    reload
• Largest process
  – Obtains the most free frames
• Process with the largest remaining
  execution window

More Related Content

What's hot

40 demand paging
40 demand paging40 demand paging
40 demand paging
myrajendra
 
Virtual memory
Virtual memoryVirtual memory
Virtual memory
Anuj Modi
 
Connection Machine
Connection MachineConnection Machine
Connection Machine
butest
 
chap 18 multicore computers
chap 18 multicore computers chap 18 multicore computers
chap 18 multicore computers
Sher Shah Merkhel
 

What's hot (20)

virtual memory
virtual memoryvirtual memory
virtual memory
 
40 demand paging
40 demand paging40 demand paging
40 demand paging
 
Computer organization memory
Computer organization memoryComputer organization memory
Computer organization memory
 
Virtual memory
Virtual memoryVirtual memory
Virtual memory
 
Virtual memory
Virtual memoryVirtual memory
Virtual memory
 
Chapter 9 - Virtual Memory
Chapter 9 - Virtual MemoryChapter 9 - Virtual Memory
Chapter 9 - Virtual Memory
 
04 cache memory.ppt 1
04 cache memory.ppt 104 cache memory.ppt 1
04 cache memory.ppt 1
 
Virtual memory
Virtual memory Virtual memory
Virtual memory
 
Operating Systems: Virtual Memory
Operating Systems: Virtual MemoryOperating Systems: Virtual Memory
Operating Systems: Virtual Memory
 
Unit 5 Advanced Computer Architecture
Unit 5 Advanced Computer ArchitectureUnit 5 Advanced Computer Architecture
Unit 5 Advanced Computer Architecture
 
Virtual memory
Virtual memoryVirtual memory
Virtual memory
 
Processor Organization
Processor OrganizationProcessor Organization
Processor Organization
 
Cache memory
Cache memoryCache memory
Cache memory
 
Connection Machine
Connection MachineConnection Machine
Connection Machine
 
chap 18 multicore computers
chap 18 multicore computers chap 18 multicore computers
chap 18 multicore computers
 
Cs8493 unit 3
Cs8493 unit 3Cs8493 unit 3
Cs8493 unit 3
 
Virtual Memory
Virtual MemoryVirtual Memory
Virtual Memory
 
Virtual memory ppt
Virtual memory pptVirtual memory ppt
Virtual memory ppt
 
Unix Memory Management - Operating Systems
Unix Memory Management - Operating SystemsUnix Memory Management - Operating Systems
Unix Memory Management - Operating Systems
 
Operating system memory management
Operating system memory managementOperating system memory management
Operating system memory management
 

Viewers also liked

Virtual memory
Virtual memoryVirtual memory
Virtual memory
Mohd Arif
 
08 virtual memory
08 virtual memory08 virtual memory
08 virtual memory
Kamal Singh
 

Viewers also liked (19)

Virtual memory
Virtual memoryVirtual memory
Virtual memory
 
Virtual Memory and Paging
Virtual Memory and PagingVirtual Memory and Paging
Virtual Memory and Paging
 
VIRTUAL MEMORY
VIRTUAL MEMORYVIRTUAL MEMORY
VIRTUAL MEMORY
 
Operating Systems - Virtual Memory
Operating Systems - Virtual MemoryOperating Systems - Virtual Memory
Operating Systems - Virtual Memory
 
Virtual memory
Virtual memoryVirtual memory
Virtual memory
 
Virtual memory
Virtual memoryVirtual memory
Virtual memory
 
Virtual memory
Virtual memoryVirtual memory
Virtual memory
 
Operating System-Memory Management
Operating System-Memory ManagementOperating System-Memory Management
Operating System-Memory Management
 
Virtual Memory - Part1
Virtual Memory - Part1Virtual Memory - Part1
Virtual Memory - Part1
 
Virtual Memory - Part2
Virtual Memory - Part2Virtual Memory - Part2
Virtual Memory - Part2
 
Paging and Segmentation in Operating System
Paging and Segmentation in Operating SystemPaging and Segmentation in Operating System
Paging and Segmentation in Operating System
 
08 virtual memory
08 virtual memory08 virtual memory
08 virtual memory
 
Mô hình mạng pert
Mô hình mạng pertMô hình mạng pert
Mô hình mạng pert
 
Ch09
Ch09Ch09
Ch09
 
Ch10: Virtual Memory
Ch10: Virtual MemoryCh10: Virtual Memory
Ch10: Virtual Memory
 
Os8
Os8Os8
Os8
 
Virtual Memory (Making a Process)
Virtual Memory (Making a Process)Virtual Memory (Making a Process)
Virtual Memory (Making a Process)
 
Virtual Memory
Virtual MemoryVirtual Memory
Virtual Memory
 
Virtual memory
Virtual memoryVirtual memory
Virtual memory
 

Similar to virtual memory

Lecture 8- Virtual Memory Final.pptx
Lecture 8- Virtual Memory Final.pptxLecture 8- Virtual Memory Final.pptx
Lecture 8- Virtual Memory Final.pptx
Amanuelmergia
 
Virtual Memory sjkdhikejv vsdkjnksnv vkjhfvk
Virtual Memory sjkdhikejv vsdkjnksnv vkjhfvkVirtual Memory sjkdhikejv vsdkjnksnv vkjhfvk
Virtual Memory sjkdhikejv vsdkjnksnv vkjhfvk
HKShab
 

Similar to virtual memory (20)

Chapter08
Chapter08Chapter08
Chapter08
 
Virtual Memory.pdf
Virtual Memory.pdfVirtual Memory.pdf
Virtual Memory.pdf
 
08 operating system support
08 operating system support08 operating system support
08 operating system support
 
16. PagingImplementIssused.pptx
16. PagingImplementIssused.pptx16. PagingImplementIssused.pptx
16. PagingImplementIssused.pptx
 
08 operating system support
08 operating system support08 operating system support
08 operating system support
 
08 operating system support
08 operating system support08 operating system support
08 operating system support
 
Demand paging
Demand pagingDemand paging
Demand paging
 
Operating system 37 demand paging
Operating system 37 demand pagingOperating system 37 demand paging
Operating system 37 demand paging
 
07-MemoryManagement.ppt
07-MemoryManagement.ppt07-MemoryManagement.ppt
07-MemoryManagement.ppt
 
08 operating system support
08 operating system support08 operating system support
08 operating system support
 
Virtual Memory Management Part - I.pdf
Virtual Memory Management Part - I.pdfVirtual Memory Management Part - I.pdf
Virtual Memory Management Part - I.pdf
 
Lecture 8- Virtual Memory Final.pptx
Lecture 8- Virtual Memory Final.pptxLecture 8- Virtual Memory Final.pptx
Lecture 8- Virtual Memory Final.pptx
 
virtual_memory (3).pptx
virtual_memory (3).pptxvirtual_memory (3).pptx
virtual_memory (3).pptx
 
PAGIN AND SEGMENTATION.docx
PAGIN AND SEGMENTATION.docxPAGIN AND SEGMENTATION.docx
PAGIN AND SEGMENTATION.docx
 
OS UNIT4.pptx
OS UNIT4.pptxOS UNIT4.pptx
OS UNIT4.pptx
 
Chap8 Virtual Memory. 1997-2003.ppt
Chap8 Virtual Memory. 1997-2003.pptChap8 Virtual Memory. 1997-2003.ppt
Chap8 Virtual Memory. 1997-2003.ppt
 
Memory Management in Operating Systems for all
Memory Management in Operating Systems for allMemory Management in Operating Systems for all
Memory Management in Operating Systems for all
 
Lecture-7 Main Memroy.pptx
Lecture-7 Main Memroy.pptxLecture-7 Main Memroy.pptx
Lecture-7 Main Memroy.pptx
 
Memory Management
Memory ManagementMemory Management
Memory Management
 
Virtual Memory sjkdhikejv vsdkjnksnv vkjhfvk
Virtual Memory sjkdhikejv vsdkjnksnv vkjhfvkVirtual Memory sjkdhikejv vsdkjnksnv vkjhfvk
Virtual Memory sjkdhikejv vsdkjnksnv vkjhfvk
 

Recently uploaded

Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
WSO2
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Victor Rentea
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
panagenda
 

Recently uploaded (20)

Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with Milvus
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
CNIC Information System with Pakdata Cf In Pakistan
CNIC Information System with Pakdata Cf In PakistanCNIC Information System with Pakdata Cf In Pakistan
CNIC Information System with Pakdata Cf In Pakistan
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
 
Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)
 
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
 
FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
 
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
 
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
 
Vector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptxVector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptx
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
 
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWEREMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
 
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
 

virtual memory

  • 1. Operating Systems: Internals and Design Principles, 6/E William Stallings Chapter 8 Virtual Memory Patricia Roy Manatee Community College, Venice, FL ©2008, Prentice Hall
  • 2. Hardware and Control Structures • Memory references are dynamically translated into physical addresses at run time – A process may be swapped in and out of main memory such that it occupies different regions
  • 3. Hardware and Control Structures • A process may be broken up into pieces, which do not need to be located contiguously in main memory • It is not necessary for all pieces of a process to be loaded in main memory during execution of the process
  • 4. Execution of a Program • Operating system brings into main memory a few pieces of the program • Resident set - portion of process that is in main memory • An interrupt is generated when an address is needed that is not in main memory • Operating system places the process in a blocking state
  • 5. Execution of a Program • Piece of process that contains the logical address is brought into main memory – Operating system issues a disk I/O Read request – Another process is dispatched to run while the disk I/O takes place – An interrupt is issued when disk I/O complete which causes the operating system to place the affected process in the Ready state
  • 6. Improved System Utilization • More processes may be maintained in main memory – Only load in some of the pieces of each process – With so many processes in main memory, it is very likely a process will be in the Ready state at any particular time • A process may be larger than all of main memory
  • 7. Types of Memory • Real memory – Main memory • Virtual memory – Memory on disk – Allows for effective multiprogramming and relieves the user of tight constraints of main memory
  • 8. Thrashing • Swapping out a piece of a process just before that piece is needed • The processor spends most of its time swapping pieces rather than executing user instructions
  • 9. Principle of Locality • Program and data references within a process tend to cluster • Only a few pieces of a process will be needed over a short period of time • Possible to make intelligent guesses about which pieces will be needed in the future • This suggests that virtual memory may work efficiently
  • 10. Support Needed for Virtual Memory • Hardware must support paging and segmentation • Operating system must be able to do the management the movement of pages and/or segments between secondary memory and main memory
  • 11. Paging • Each process has its own page table • Each page table entry contains the frame number of the corresponding page in main memory • A bit is needed to indicate whether the page is in main memory or not
  • 13. Modify Bit in Page Table • Modify bit is needed to indicate if the page has been altered since it was last loaded into main memory • If no change has been made, the page does not have to be written to the disk when it needs to be replaced
  • 16. Page Tables • Page tables are also stored in virtual memory • When a process is running, part of its page table is in main memory
  • 18. Inverted Page Table • Used on PowerPC, UltraSPARC, and IA- 64 architecture • Page number portion of a virtual address is mapped into a hash value • Hash value points to inverted page table • Fixed proportion of real memory is required for the tables regardless of the number of processes
  • 19. Inverted Page Table • Page number • Process identifier • Control bits • Chain pointer
  • 21. Translation Lookaside Buffer • Each virtual memory reference can cause two physical memory accesses – One to fetch the page table – One to fetch the data • To overcome this problem a high-speed cache is set up for page table entries – Called a Translation Lookaside Buffer (TLB)
  • 22. Translation Lookaside Buffer • Contains page table entries that have been most recently used
  • 23. Translation Lookaside Buffer • Given a virtual address, processor examines the TLB • If page table entry is present (TLB hit), the frame number is retrieved and the real address is formed • If page table entry is not found in the TLB (TLB miss), the page number is used to index the process page table
  • 24. Translation Lookaside Buffer • First checks if page is already in main memory – If not in main memory a page fault is issued • The TLB is updated to include the new page entry
  • 29. Page Size • Smaller page size, less amount of internal fragmentation • Smaller page size, more pages required per process • More pages per process means larger page tables • Larger page tables means large portion of page tables in virtual memory
  • 30. Page Size • Secondary memory is designed to efficiently transfer large blocks of data so a large page size is better
  • 31. Page Size • Small page size, large number of pages will be found in main memory • As time goes on during execution, the pages in memory will all contain portions of the process near recent references. Page faults low. • Increased page size causes pages to contain locations further from any recent reference. Page faults rise.
  • 34. Segmentation • May be unequal, dynamic size • Simplifies handling of growing data structures • Allows programs to be altered and recompiled independently • Lends itself to sharing data among processes • Lends itself to protection
  • 35. Segment Tables • Starting address corresponding segment in main memory • Each entry contains the length of the segment • A bit is needed to determine if segment is already in main memory • Another bit is needed to determine if the segment has been modified since it was loaded in main memory
  • 38. Combined Paging and Segmentation • Paging is transparent to the programmer • Segmentation is visible to the programmer • Each segment is broken into fixed-size pages
  • 39. Combined Paging and Segmentation
  • 42. Fetch Policy • Determines when a page should be brought into memory • Demand paging only brings pages into main memory when a reference is made to a location on the page – Many page faults when process first started • Prepaging brings in more pages than needed – More efficient to bring in pages that reside contiguously on the disk
  • 43. Placement Policy • Determines where in real memory a process piece is to reside • Important in a segmentation system • Paging or combined paging with segmentation hardware performs address translation
  • 44. Replacement Policy • Which page is replaced? • Page removed should be the page least likely to be referenced in the near future • Most policies predict the future behavior on the basis of past behavior
  • 45. Replacement Policy • Frame Locking – If frame is locked, it may not be replaced – Kernel of the operating system – Key control structures – I/O buffers – Associate a lock bit with each frame
  • 46. Basic Replacement Algorithms • Optimal policy – Selects for replacement that page for which the time to the next reference is the longest – Impossible to have perfect knowledge of future events
  • 47. Basic Replacement Algorithms • Least Recently Used (LRU) – Replaces the page that has not been referenced for the longest time – By the principle of locality, this should be the page least likely to be referenced in the near future – Each page could be tagged with the time of last reference. This would require a great deal of overhead.
  • 48. Basic Replacement Algorithms • First-in, first-out (FIFO) – Treats page frames allocated to a process as a circular buffer – Pages are removed in round-robin style – Simplest replacement policy to implement – Page that has been in memory the longest is replaced – These pages may be needed again very soon
  • 49. Basic Replacement Algorithms • Clock Policy – Additional bit called a use bit – When a page is first loaded in memory, the use bit is set to 1 – When the page is referenced, the use bit is set to 1 – When it is time to replace a page, the first frame encountered with the use bit set to 0 is replaced. – During the search for replacement, each use bit set to 1 is changed to 0
  • 54. Behavior of Page Replacement Algorithms
  • 55. Basic Replacement Algorithms • Page Buffering – Replaced page is added to one of two lists • Free page list if page has not been modified • Modified page list
  • 56. Resident Set Size • Fixed-allocation – Gives a process a fixed number of pages within which to execute – When a page fault occurs, one of the pages of that process must be replaced • Variable-allocation – Number of pages allocated to a process varies over the lifetime of the process
  • 57. Fixed Allocation, Local Scope • Decide ahead of time the amount of allocation to give a process • If allocation is too small, there will be a high page fault rate • If allocation is too large there will be too few programs in main memory – Processor idle time – Swapping
  • 58. Variable Allocation, Global Scope • Easiest to implement • Adopted by many operating systems • Operating system keeps list of free frames • Free frame is added to resident set of process when a page fault occurs • If no free frame, replaces one from another process
  • 59. Variable Allocation, Local Scope • When new process added, allocate number of page frames based on application type, program request, or other criteria • When page fault occurs, select page from among the resident set of the process that suffers the fault • Reevaluate allocation from time to time
  • 60. Cleaning Policy • Demand cleaning – A page is written out only when it has been selected for replacement • Precleaning – Pages are written out in batches
  • 61. Cleaning Policy • Best approach uses page buffering – Replaced pages are placed in two lists • Modified and unmodified – Pages in the modified list are periodically written out in batches – Pages in the unmodified list are either reclaimed if referenced again or lost when its frame is assigned to another page
  • 62. Load Control • Determines the number of processes that will be resident in main memory • Too few processes, many occasions when all processes will be blocked and much time will be spent in swapping • Too many processes will lead to thrashing
  • 64. Process Suspension • Lowest priority process • Faulting process – This process does not have its working set in main memory so it will be blocked anyway • Last process activated – This process is least likely to have its working set resident
  • 65. Process Suspension • Process with smallest resident set – This process requires the least future effort to reload • Largest process – Obtains the most free frames • Process with the largest remaining execution window