CSC/ECE 506 Spring 2014/7b ss

From Expertiza_Wiki
Revision as of 01:28, 25 March 2014 by Sjantli (talk | contribs) (→‎Validation Based Approach)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

TLB (Translation Lookaside Buffer) Coherence in Multiprocessing


Overview

The cache coherence problem in multiprocessing architectures and their hardware based protocol solutions have received great attention. Another coherence problem in multiprocesing is that of TLBs (Transaction Lookaside Buffers). A TLB is a fully associate hardware cache that maintains virtual to physical mapping of most recently used pages. TLB is used by CPU for fast look up of the physical page frame number for a virtual memory location. The same lookup from the page-table for the process (software based construct) is slower. A page may become invalid due to a swap out to memory or may change protection level (read v/s read-write). Maintaining coherence between TLB entries on different processors for the same page (as its state changes) is called TLB coherence problem. This coherence is mantained most often by software based solutions by the operating systems. In this note, we explain a) the background for used of TLB hardware, b) current solutions for maintaining TLB coherence and c) finally a proposal for a unified cache-TLB coherence protocol (that is hardware based).

Background - Virtual Memory, Paging and TLB

In this section we introduce the basic terminology and the set-up where TLBs are used.

A process running on a CPU has its own view of memory - "Virtual Memory" (one-single space) that is mapped to actual physical memory. Virtual memory management scheme allows programs to exceed the size of physical memory space. Virtual memory operates by executing programs only partially resident in memory while relying on hardware and the operating system to bring the missing items into main memory when needed.

Paging is a memory-management scheme that permits the physical address space of a process to be non-contiguous. The basic method involves breaking physical memory into fixed-size blocks called frames and breaking the virtual memory into blocks of the same size called pages. When a process is to be executed its pages are loaded into any available memory frames from the backing storage (for example-a hard drive).

Each process has a page table that maintains a mapping of virtual pages to physical pages. A page table is a software construct managed by operating system that is kept in main memory. For a process, a pointer to the page table (Page-Table Base Register-PTBR) is stored in a register. Changing page table requires changing only this one register, substantially reducing the context-switch time. (A page table, with virtual page number that is 4 bytes long (32 bits) can point to 2^32 physical page frames. If frame size is 4Kb then a system with 4-byte entries can address 2^44 bytes (or 16 TB) of physical memory).

Every address is divided into two parts – a page number (p) and a page offset (d). The page number is used as an index into the page table. The page table contains the base address of each page in physical memory. The base address is combined with the page offset to define the physical memory address that is sent to the memory unit. The problem with this approach is the time required to access a user memory location. If we want to access location, we must first index into the page table (that requires a memory access of PTBR) and then find the frame number from the page table and thereafter access the required frame number. The memory access is slowed down by a factor of 2.

Virtual to physical page mapping using TLB and Page table (Source-Operating System Concepts - Silberschatz, Galvin, Gagne)

The standard solution to this problem is to use a special, small fast lookup hardware cache called a Translation Look-Aside Buffer (TLB). The TLB is fully associate, high-speed memory. Each entry in TLB consists of two parts –a key (tag) and a value. When the associative memory is presented with an item, the item is compared with all keys simultaneously. If the item is found, the corresponding value field is returned. The search is fast; the hardware, however is expensive. Typically, the number of entries in a TLB is small, often numbering between 64 to 1024.

The TLB is used with page tables in the following way. The TLB contains only a few of the page-table entries. When a logical address is generated by the CPU, its page number is presented to the TLB. If the page number is found its frame number is immediately available and is used to access memory.

If the page number is not in the TLB (known as a TLB miss), a memory reference to the page table must be made. When the frame is obtained, we can use it to access memory. In addition, we add the page number and frame number to the TLB, so that they will be found quickly on the next reference. If the TLB is already full of entries, the operating system must select one for replacement. Replacement policies range from least recently used (LRU) to random.

A page table typically has an additional bit per entry to indicate whether the frame number is read only or both read-write (protection level). Another bit valid-invalid is also added to indicate whether the frame is part of that particular process.

Mutilevel pages tables are often used rather than single page table per process. The first level page table points to the entry in the next level page table and so on until the mapping of virtual to physical page is obtained from the last level page table.

It may be noted that caches typically use "Virtual Indexed Physically Tagged" solution that enables simultaneous look-up of both L-1 cache and TLB. If the page block is not found in L-1 cache, a TLB look-up of address is used to refresh the page block. Simultanous look-up of both L-1 cache and TLB makes the process efficient.

TLB coherence problem in multiprocessing

In multiprocessor systems with shared virtual memory and multiple Memory Management Units (MMUs), the mapping of a virtual address to physical address can be simultaneously cached in multiple TLBs. (This happens because of sharing of data across processors and process migration from one processor to another). Since the contents of these mapping can change in response to changes of state (and the related page in memory), multiple copies of a given mapping may pose a consistency problem. Keeping all copies of a mapping descriptor consistent with each other is sometimes referred to as TLB coherence problem.

A TLB entry changes because of the following events - a) Mapping modifications i.e. changing the mapping of a virtual page to a physical frame. b) Increasing the page protection privileges (e.g., from read-write to read-only). c) Invalidation of memory mapping of virtual to physical memory. d) Decreasing the page privileges (e.g., from read-only to read-write).

There are two types of changes namely safe and unsafe changes. Unsafe changes will lead to TLB coherence problem while the safe changes can be ignored. Unsafe changes include a) Mapping modifications i.e. changing the mapping of a virtual page to a physical frame, b) Increasing the page protection privileges (e.g., from read-write to read-only) and c) Invalidation of memory mapping of virtual to physical memory. Safe changes include d) Decreasing the page privileges (e.g., from read-only to read-write). Some architectures do not follow TLB coherence for each datum (i.e., TLBs may contain different values). These architectures require TLB coherence only for unsafe changes made to address translations.

A variety of solutions have been used for TLB coherence. Software solutions through operating systems are popular since TLB coherence operations are less frequent than cache coherence operations. The exact solution depends on whether PTEs (Page Table Entries) are loaded into TLBs directly by hardware or under software control. Hardware solutions are used where TLBs are not visible to software.

In the next sections, we cover following approaches to TLB coherence: virtually addressed caches, software TLB shootdown, address space Identifiers, hardware TLB coherence.

We also review UNified Instruction/Translation/Data (UNITD) Coherence, a hardware coherence framework that integrates TLB and cache coherence protocol.

TLB Coherence Approaches

TLB coherence refers to the consistency of the information stored in page-map tables (PMTs) with the cached copies of this data in each processor’s TLB. A PMT entry (and its TLB counterpart) store various data elements, including:

  • The physical memory frame number for each virtual page resident in main memory.
  • A protection field indicating how the page may be addressed (e.g., read-only or read-write).
  • A presence bit that indicates whether the page is in main memory.
  • A dirty bit that indicates whether the page was modified while in main memory.
  • Page reference history used to indicate whether a page has been referenced during some time interval.<ref>

"Teller, Patricia J. Translation-Lookaside Buffer Consistency"</ref>

Consistency among TLBs and PMTs need not be maintained for all types of changes to the above data elements. For example, it is possible to safely modify page reference history in main memory without also updating the TLBs. Updates are therefore categorized as either safe or unsafe.

It is important to note that these inconsistencies are not ignored but rather are handled downstream by other mechanisms. For example, if the operating system kernel detects a process is attempting to write to a page marked as read-only, it will verify whether the page’s protection settings have been reduced before throwing a protection fault. If the protection settings in main memory have been reduced by another processor, the kernel would invalidate the acting processor’s TLB and retry the operation.

The above example also illustrates why a change to the same data element in the opposite direction (an increase in protection) would be unsafe without a TLB coherence scheme as there would be no mechanism for the operating system to trap the condition and update the processor’s TLB.

Any modifications to the mapping between a virtual page and main memory frame are also unsafe without a TLB coherence mechanism. This is in large part due to the fact that page numbers are formed by splitting a finite address space into discrete units. These pages in turn are mapped to physical memory frames as needed by the operating system. The mapping of a page to a frame is a transient condition and the same page may be mapped to different frames across time. Without a TLB coherence mechanism, it is possible that a processor may attempt to use a stale page-frame mapping to reference a frame that has been reassigned.

TLB Coherence Strategies

There are several approaches to maintaining coherence or consistency among multiple TLBs in a multi-processor machine and the page tables in main memory.

Virtually Indexed Caches

Virtually indexed caches may be accessed using the virtual memory reference. This approach eliminates the TLB altogether and its associated coherence issue. This approach requires processors to index data cache values using virtual addresses and thereby avoid the need for virtual-physical address translations between the processor and the L1 cache. Ultimately, however, in order to address physical main memory on an L1 miss (or L2 miss depending on the indexing strategy of the L2 cache), translation is still required from the virtual reference to a physical address in memory. In a non-TLB architecture, this translation uses the page mapping tables directly.

Under this approach, PMT entries may be brought into the processor’s data cache. Though the cache coherence protocol in effect will manage consistency across all processor caches, it does not handle consistency between cached PMT entries and the page tables in main memory. Hence, even in the absence of a TLB there remains a coherence problem between data caches and main memory page tables. As PMT entries in main memory are modified by the operating system kernel, the processor data caches must be reconciled by invalidating stale PMT entries.

Since this coherence issues relates more to PMT-data cache coherence (versus TLB coherence), it will not be discussed further in this section.

TLB Shootdown

The TLB Shootdown method is the most widely used TLB coherence approach. The approach has variants that turn on several factors, how many processors participate in the TLB update (the victim list) and the extent of hardware support for the process, which in turn may determine the visibility of the TLB by operating system kernel.

In the discussion below, the initiating processor is defined as the processor servicing the operating system kernel during the course of the PMT update.

Broadly speaking, the shootdown approach includes the following two components:

  • A synchronization strategy across all processors during the PMT update. This orchestration begins by the initiating processor locking the page table (or some portion thereof).
  • A message broadcasting mechanism that allows the initiating processor to notify other processors that their TLB cache may be out-of-date and to suspend their use of the TLB during the course of the update.

This process is described in detail next followed by a discussion of implementations and their shootdown variants.

Details of the TLB Shootdown

  1. The initiating processor must first disable local interrupts, preventing another thread from preempting and modifying its TLB. It also clears its active flag for TLB usage. This flag is used to indicate whether a processor is currently using its TLB cache.
  2. The initiating processor locks the page table it is modifying or potentially some smaller portion of it. This prevents modification to the same page data by other processors.
  3. The shootdown method is primarily based on software that invalidates TLB entries. The initiating processor must place an inter-processor interrupt (IPI) request to its interrupt controller. The interrupt includes a data structure (IDT) that references the location of the shootdown program and the target TLB entries that should be invalidated. The initiator waits until all processors have responded.
  4. The recipients of the interrupt respond by disabling their local inter-processor interrupts and invalidating the affected TLB entries or the entire TLB (depending on the implementation) using the shootdown program. The receiving processor then clears its TLB active flag and enters into a wait state after servicing the interrupt. The wait state may be implemented by attempting to gain the same lock held by the initiating processor on the page table or polling shared memory.
  5. Once the interrupts have completed, the initiating processor makes changes to the PMTs and updates its own TLB cache. Before releasing the page lock, it sets its TLB active flag to true.
  6. The receiving processors exit their wait state and update their TLB cache and set their TLB active flag.

Implementations

Variations in implementations may occur at various steps above:

  • A modified TLB shootdown approach is implemented in IBM’s RP3 architecture that eliminates the need for victim processors to busy-wait while the PMT is being updated by the initiator. <ref>

"Teller, Patricia J. Translation-Lookaside Buffer Consistency" </ref>

  • Some editions of Windows kernels (including XP, PAE, and 2003 Enterprise Edition) attempt to batch updates to the PMT entries in order to minimize shootdowns and their associated performance dip. <ref>

"Operating Systems and PAE Support" </ref>

  • The Intel Nehalem platform uses a hierarchical TLB. <ref>

"First the Tick, Now the Tock: Next Generation Intel Microarchitecture (Nehalem)" </ref>

  • Carnegie Mellon University's Mach operating system which has been ported to BBN Butterfly, Encore's Multimax, IBM's RP3 and Sequent's Balance and Symmetry systems. <ref>

"Teller, Patricia J. Translation-Lookaside Buffer Consistency" </ref>

  • Linux and Windows Operating Systems running on Intel and AMD chips. <ref>

"Baumann, Andrew et al. The Multikernel: A new OS architecture for scalable multicore systems" </ref>

Hardware TLB Coherence

Some processors have modified their instruction set to include a TLB invalidate instruction. The instruction is invoked by the operating system kernel to broadcast the modified page address from the initiating processor. The instruction is placed on the shared bus and a snooper on each processor detects the instruction and invalidates the appropriate TLB entries.

Implementations

Examples of this include the Intel Itanium architecture (via the ptc.g and ptc.ga instructions) and the tlbivax instruction on IBM’s Power series (for Power ISA 2.06).

An example of an operating system that uses the Itanium support for instruction based TLB invalidation is the Linux kernel.<ref> "Bryant, Raj and Hawkes, John. Linux Scalability for Large NUMA Systems" </ref>

Address Space Identifier (ASID) based Approach

The MIPS family of processors assigns each TLB entry a unique identifier called an ASID. The ASID is similar to a process identifier (PID) except that its uniqueness is not guaranteed for the life of a process. Each process is assigned a unique ASID on each processor. Hence, a PID may have many ASIDs. A separate array is used to keep track of this mapping.

When a process modifies a PTE in main memory, the kernel sets all the process’ ASIDs to 0 on the other processors. Note: This does not modify the ASIDs in the TLB entries, only the array that maps a process to a processor.<ref>Culler, David E. et al. Parallel Computer Architecture: A Hardware/Software Approach, 1999, p. 440,441.</ref>

When the kernel find that a process is assigned a zero ASID value on a processor, it assigns it a new ASID. This new ASID will not match the stale TLB entries (and their ASIDs) on that processor, forcing a subsequent TLB invalidation.

Implementations

The IRIX 5.2 operating system uses this mechanism on MIPS as well as the AMD64 operating system. <ref> "Biemueller, Sebastian. ASID Management in Xen AMD-V: Partioning the physical TLB with SVM ASIDs" </ref>

Validation Based Approach

Under this approach, TLB invalidation is postponed until memory is accessed. This approach eliminates the need for interrupts and synchronization. Validation table has mapping of frame and the latest count. When a processor sends a memory requests, the requests also contains the generation count which is stored in the processor’s TLB. If the count in the request and validation table is matched, then the access is granted else the TLB entry of that processor is invalidated.<ref>"Teller, Patricia J. Translation-Lookaside Buffer Consistency" </ref>

Modified TLB Shootdown

Any processor that uses the page table need not do a busy-wait for synchronizing with the other processors. Although the page table modifications are serialized, shared page tables can be used because page table modifications are atomic. The main difference from the original TLB shoot down algorithm is that at any point in time, one processor can use the page's old translation while another processor can use the page's new translation.<ref>"Teller, Patricia J. Translation-Lookaside Buffer Consistency" </ref>

Read Locked TLBs

The premise of Read locked TLBs is that: preventing a TLB entry from becoming invalid prevents inconsistency in the TLBs. This stratergy does not allow any un-safe changes to be done if the valid copy of the translation resides in more than one TLB. The processor will maintain a read lock on every PTE till the translation for that page resides in the TLB. This will prevent the PTE from becoming inconsistent. The unsafe change will have to be postponed, also this limits the sharing of memory among processes.<ref>"Teller, Patricia J. Translation-Lookaside Buffer Consistency" </ref>

Unified Cache and TLB coherence solution

In this section the salient features of UNified Instruction/Translation/Data (UNITD) Coherence protocol proposed by Romanescu/Lebeck/Sorin/Bracy in their paper <ref>UNified Instruction Translation Data UNITD Coherence One Protocol to Rule Them All, Romanescu/Lebeck/Sorin/Bracy</ref> are discussed. This provides an example of hardware based unified protocol for cache-TLB coherence.

UNITD is a unified hardware coherence framework that integrates TLB coherence into existing cache coherence protocol. In this protocol, the TLBs participate in cache coherence updates without needing any change to existing cache coherence protocol. This protocol is an improvement over the software based shootdown approach for TLB coherence and reduces the performance penalty due to TLB coherence significantly.

Background

Figure 3: Per-shootdown Latency Source-UNITD protocol paper by Romanescu/Lebeck/Sorin/Bracy
Figure 4: Shootdown performance overhead on Phoenix benchmarks. Source-UNITD protocol paper by Romanescu/Lebeck/Sorin/Bracy

Cache coherence protocols are all-hardware based because this provides a) higher performance and b) decoupling with architectural issues. On the Other hand, the shootdown protocol used for TLB coherence is essentially software based. A hardware based TLB coherence can a) improve TLB coherence performance significantly for high number of processors and b)provide a cleaner interface to the Operating system.

TLB shootdowns are invisible to user applications, although they directly impact user performance. The performance impact depends on - a) position of TLB in memory hierarchy, b) shoot down algorithm used, c) number of processors. The position of TLB in memory heirarchy refers to whether the TLB is placed close to processor or is close to memory. Shoot down algorithms trade performance v/s complexity. The performance penalty for shootdown increases with the number of processors as can be seen from Figure-3 that shows latency of the processor issuing shootdown and the ones receiving shootdowns. The effective latency is more than shown in the graphs because of TLB invalidations resulting in extra cycles spent in repopulating the TLB that can happen either from either cache or main memory depending on the case. The latency from TLB shootdowns can be higher for application that use the routine more often. (Refer to Figure-4).

UNITD Coherence Implementation

At a high level, UNITD integrates the TLBs into the existing cache coherence protocol (such as typical MOSI -Modified Owned Share Invalid coherence states). TLBs are simply additional caches that participate in the coherence protocol however like read-only instruction caches. UNITD has no impact on the cache coherence protocol and thus does not increase its complexity.

As mentioned before, TLB entries are read-only. (Virtual to physical address mapping are never modified in the TLBs themselves but rather in the page table entries). There are only two coherence states: Shared (read-only) and Invalid. UNITD uses a Valid bit in TLB to maintain an entry’s coherence state. When a translation is inserted into a TLB, it is marked as Shared. The cached translation can be accessed by the local core as long as it is in the Shared state. The translation remains in Shared state until a TLB coherence message invalidates it. Thereafter subsequent memory accesses depending on it will miss in the TLB and reacquire the translation from the memory system.

alt text
Figure 1: (Source-UNITD protocol paper by Romanescu/Lebeck/Sorin/Bracy) Shows how the PCAM is integrated into the system, with interfaces to the TLB insertion/eviction mechanism (for inserting/evicting the corresponding PCAM entries), the coherence controller (for receiving coherence invalidations) and the processor core . The PCAM is off the critical path of a memory access; it is not accessed during regular TLB lookups for obtaining translations.

.

In order to integrate TLB coherence with the existing cache coherence protocol with minimal architectural changes, we relax the correspondence of the translations to the memory block containing the PTE, rather than the PTE itself. Maintaining translation granularity at a coarser grain (i.e., cache block, rather than PTE) trades a small performance penalty for ease of integration. Because multiple PTEs can be placed in the same cache block, the PCAM can hold multiple copies of the same datum. For simplicity, PCAM entries are referred as PTE addresses.

alt text
Figure 2: (Source-UNITD protocol paper by Romanescu/Lebeck/Sorin/Bracy) Shows the two operations associated with the PCAM: (a) inserting an entry into the PCAM and (b) performing a coherence invalidation at the PCAM. PTE addresses are added in the PCAM simultaneously with the insertion of their corresponding translations in the TLB.

Because the PCAM has the same structure as the TLB, a PTE address is inserted in the PCAM at the same index as its corresponding translation in the TLB (physical address 12 in Figure 2(a) above. Note that there can be multiple PCAM entries with the same physical address, as in Figure 2a; this situation occurs when multiple cached translations correspond to PTEs residing in the same cache block.

PCAM entries are removed as a result of the replacement of the corresponding translation in the TLB or due to an incoming coherence request for read-write access. If a coherence request hits in the PCAM, the Valid bit for the corresponding TLB entry is cleared. If multiple TLB translations have the same PTE block address, a PCAM lookup on this block address results in the identification of all associated TLB entries. Figure 2(b) above illustrates a coherence invalidation of physical address 12 that hits in two PCAM entries.

Definitions

PMT(Page Mapping Table)
This is an in-memory data structure that primarily helps the operating system map two address spaces two one another: virtual memory and physical memory. In addition to this mapping information, the PMT also stores additional data fields to keep track of whether the virtual memory has been loaded in physical memory, permissions and utilization metrics. Coupled with the TLB, the PMT is a critical component of virtual memory systems. Learn more about page tables here.
PTE (Page Table Entry)
A row in a page mapping table that relates a single page in virtual memory to a physical memory address or a page frame.
TLB(Translation Look-Aside Buffer)
Associative, high-speed memory used to cache page table information and speed up virtual memory references from the processor.
Presence Bit
One of the data elements in each PMT entry. Indicates whether the page is loaded in main memory (versus stored to disk).
Safe Change
A type of update that can be made to a PMT safely without also updating the cached TLB counterpart.
Unsafe Change
The change that will cause TLB coherence problem.
TLB Coherence Strategy
A process or system for maintaining the consistency of information between the processor TLBs and the PMT entries.
IDT (Interrupt Descriptor Table)
A vector-based data structure used to communicate information between processors during an inter-processor interrupt.
Interrupt
A hardware or software based method for signalling to a processor that there is work for it.
ASID (Address Space Identifier)
A number assigned to process that is unique within the scope of a processor for a specified time period. A process may be assigned multiple ASIDs across time. Used with the TLB entries, the ASID helps the discard stale TLB entries.
Modified TLB Shootdown
Busy waiting for synchronization with other processors while using page table is avoided.
Read Locked TLBs
Does not allow unsafe changes to be done if valid copy exists in more than one TLB.
Page Table Base Register
Pointer to page table location is maintained in this register
UNITD protocol
A protocol that combine cache and TLB coherence proposed by Romanescu/Lebeck/Sorin/Bracy
Validation based approach
In this approach TLB entries are not updated until used
Virtual Index Cache
Cache that stores data based on virtual page number
TLB Shootdown
Software based TLB coherence method that requires one processor to shoot-down/disable TLB entries in other processors before providing them with a new TLB entry

Links

Address Resolution and the TLB

References

1. Operating System Concepts - Silberschatz, Galvin, Gagne. Seventh Edition, Wiley publication, 2005

2. Fundamentals of Parallel Computer Architecture, Yan Solihin, 2008-2009

3. Culler, David E. et al. Parallel Computer Architecture: A Hardware/Software Approach, 1999, Gulf Professional Publishing

4. Microprocessor Memory Management Unit, Milan Milenkovic, IEEE, Vol10, Issue-2, 1990, Pages 70-85

Notes

<references>

</references>

Quiz

1. Page Table is maintained as a software construct (choose one answer)

a) True
b) False

2. TLB keeps a mapping of ___ to ____

a) Virtual Page to Physical Page
b) Virtual Address to Physical Address
c) Physical Page to Virtual Page
d) Physical Address to Virtual Address

3. Currently TLB coherence is most often achieved through

a) Software solutions
b) Hardware solutions

4. UNITD is (choose one answer)

a) A protocol for TLB coherence only
b) A protocol for cache coherence only
c) A protocol for cache and TLB coherence

5. Which of the following approaches provides a solution for TLB coherence (choose at least one)

a) Virtual Cache address
b) Shootdown
c) Invalidation
d) Hardware solutions

6. What messaging implementation is typically used with TLB Shootdown: (choose one answer)

a) TLB active flag
b) Hardware instructions
c) Inter-Processor Interrupts
d) PMT entries in main memory

7. Which data elements of a PMT entry are unsafe without a TLB coherence scheme: (choose all that apply)

a) Increase in protection level;
b) Changes to virtual-physical address mappings.
c) Page table reference counts
d) Presence bit

8. An ASID refers to: (choose one answer)

a) The unique ID of a process for its lifetime
b) The unique ID of a process-processor pair at a point in time
c) The unique ID of a PMT entry
d) The unique ID of a page

9. If Virtually Indexed data caches avoid the use of TLBs, where is the coherence issue identified? (choose one answer)

a) Between PMT entries in main memory and the disk storage
b) Between PMT entries in main memory and processor registers
c) Between PMT entries in main memory and the instruction cache
d) Between PMT entries in main memory and the data cache

10. How does the initiating processor prevent updates to the PMT entries by other processors during a PMT update? (Choose one answer)

a) Via a shared memory semaphore
b) Via the interrupt IDC
c) Via a hardware instruction
d) Via a lock to the page table in main memory

11. What is the main difference between Modified TLB shootdown and TLB shootdown methods? (Choose one answer)

a) One processor can use old translation while other processor can use new translation simultaneously.
b) Modified TLB shootdown causes busy waiting for synchronization.
c) TLB shootdown doesn't cause busy waiting for synchronization.
d) Old and new translations can not be accessed simultaneously by two different processors.

12. Best situation to use Read Locked TLBs is? (Choose one answer)

a) In all situation it is the best approach.
b) When there are fewer unsafe changes because they will be postponed.
c) when there are large number of unsafe changes.
d) Does not depend on the number of unsafe changes.

Answers 1-a, 2-a, 3-a, 4-c, 5-a,b,c,d, 6-c, 7-a,b, 8-b, 9-d, 10-d 11-a 12-b