US5479627A - Virtual address to physical address translation cache that supports multiple page sizes - Google Patents

Virtual address to physical address translation cache that supports multiple page sizes Download PDF

Info

Publication number
US5479627A
US5479627A US08/118,398 US11839893A US5479627A US 5479627 A US5479627 A US 5479627A US 11839893 A US11839893 A US 11839893A US 5479627 A US5479627 A US 5479627A
Authority
US
United States
Prior art keywords
virtual address
translation
page
candidate
virtual
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US08/118,398
Inventor
Yousef A. Khalidi
Glen R. Anderson
Stephen A. Chessin
Shing I. Kong
Charles E. Narad
Madhusudhan Talluri
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sun Microsystems Inc
Original Assignee
Sun Microsystems Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sun Microsystems Inc filed Critical Sun Microsystems Inc
Priority to US08/118,398 priority Critical patent/US5479627A/en
Assigned to SUN MICROSYSTEMS, INC. reassignment SUN MICROSYSTEMS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ANDERSON, GLEN R., CHESSIN, STEPHEN A., NARAD, CHARLES E., KHALIDI, YOUSEF A., KONG, SHING IP, TALLURI, MADHUSUDHAN
Assigned to SUN MICROSYSTEMS, INC. reassignment SUN MICROSYSTEMS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ANDERSON, GLEN R., CHESSIN, STEPHEN A., NARAD, CHARLES E., KHALIDI, YOUSEF A., KONG, SHING IP, TALLURI, MADHUSUDHAN
Priority to EP94305846A priority patent/EP0642086B1/en
Priority to DE69408922T priority patent/DE69408922T2/en
Priority to JP6236020A priority patent/JPH07200409A/en
Application granted granted Critical
Publication of US5479627A publication Critical patent/US5479627A/en
Priority to US08/874,201 priority patent/US5956756A/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/10Address translation
    • G06F12/1027Address translation using associative or pseudo-associative address translation means, e.g. translation look-aside buffer [TLB]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/65Details of virtual memory and virtual address translation
    • G06F2212/652Page size control

Definitions

  • the present invention relates to the field of computer systems. More specifically, the present invention relates to translation from a virtual address to a physical address in a virtual memory computer system.
  • Virtual memory is a technique that allows an application to see the system as providing a large uniform primary memory, which in reality may be smaller, more fragmented, and/or partially simulated by secondary storage such as a hard disk.
  • Applications access memory through virtual addresses, which are translated (mapped) by special hardware onto physical addresses. Paging and segmentation are two common implementations of virtual memory.
  • the virtual address space is divided into a number of fixed-size blocks called pages, each of which can be mapped onto any of the similarly sized physical pages available on the system.
  • special memory management hardware such as a memory management unit (MMU) or paged memory management unit (PMMU) performs the address translation from virtual addresses to physical addresses.
  • MMU memory management unit
  • PMMU paged memory management unit
  • any attempt to access data that is not present in physical (system) memory causes the PMMU to send an interrupt signal to the central processing unit (CPU); the operating system then transfers the data in from other storage (such as hard disk), without the application "knowing" about the transfer.
  • a first process may refer to a given physical address using a first virtual address and a second process running; on the same CPU, or on a different CPU in a multi-processor environment, will refer to the given physical address by a second virtual address.
  • the two virtual addresses corresponding to the same physical address are referred to as aliases.
  • Some CPUs are able to operate using more than one size of virtual page. This permits the virtual page size used in a mapping to be selected to best meet the needs of the virtual address space being mapped. If one uses big pages, a large amount of virtual memory can be translated using a single entry in the MMU. Therefore, fewer resources are needed for the MMU thereby increasing performance. The use of big pages can, however, lead to memory fragmentation problems not encountered when small pages are used.
  • a method and apparatus for translating a virtual address to a physical address is disclosed.
  • a virtual address to be translated has a virtual page offset and a virtual page number.
  • the virtual address to be translated addresses a page of memory.
  • the size of this page is unknown.
  • the unknown page size is one of L different possible page sizes where L is a positive integer greater than one.
  • Each of the L different page sizes is selected to be a test page size and a test is performed.
  • a pointer into a translation storage buffer is calculated.
  • the pointer is calculated from the virtual address to be translated by assuming that the virtual address to be translated corresponds to a mapping of the test page size.
  • the pointer points to a candidate translation table entry of the translation storage buffer.
  • the candidate translation table entry has a candidate tag and candidate data.
  • the candidate tag identifies a particular virtual address and the candidate data identifies a particular physical address corresponding to the particular virtual address.
  • a virtual address target tag is extracted from the virtual address to be translated.
  • the virtual address target tag is calculated by assuming that the virtual address to be translated corresponds to a mapping of the test page size.
  • the target tag and the candidate tag are then compared. If the target tag matches the candidate tag, the candidate data is provided as the physical address translation corresponding to the virtual address to be translated.
  • FIG. 1 illustrates a computer system that uses an embodiment of the present invention to perform address translation from a virtual address to a physical address
  • FIG. 2 illustrates a multi-level approach for translating virtual addresses to physical addresses
  • FIG. 3 illustrates the relationship between physical addresses and virtual addresses and how a translation storage buffer index and target are derived from a virtual address
  • FIG. 4 illustrates an embodiment of a translation storage buffer that supports multiple page sizes
  • FIG. 5 illustrates a translation table entry within a translation storage buffer
  • FIG. 6 illustrates the use of hardware logic by the present invention to quickly derive a tag target and translation storage buffer pointers
  • FIG. 7 illustrates an alternate embodiment of a translation storage buffer that supports multiple page sizes
  • FIG. 8 is a flow chart illustrating a translation storage buffer look-up as performed by an embodiment of the present invention.
  • FIG. 9 is a flow chart of a procedure employed by an embodiment of the present invention to check for a translation storage buffer hit.
  • a method and apparatus for implementation of a virtual address to physical address translation cache that supports multiple page sizes is disclosed.
  • hardware provides support for a software search of the translation cache by calculating the tag value to be compared and the pointers into the cache where the comparison is to be made.
  • the hardware support described is simple to implement and efficient. If a translation exists within the translation cache, the translation can be performed by retrieving one translation entry. Because the embodiment permits the translation to be retrieved from the cache by software, frequently used portions of the translation cache can themselves, be cached in a virtually-indexed data cache.
  • the translation cache can be configured to be one cache that intermixes translations for different page sizes or configured as one cache divided into sub caches wherein each sub cache contains translations for a single page size.
  • Multiple translation caches can also be employed. Therefore, in one embodiment, separate translation caches can be provided, one for instructions and the other for data. Alternate embodiments can also be implemented wherein there is one global translation cache for all CPUs and contexts; one translation cache per CPU, with each translation cache containing all contexts; one translation cache per context, but shared by all CPUs, or one translation cache per context per CPU.
  • system translations are treated as a separate context thereby permitting separate translation caches for user translations and for system translations.
  • central processing unit 110 is coupled to primary memory 130 and secondary memory 140 by a memory interconnect such as memory bus 120.
  • central processing unit 110 can access primary memory 130 faster than it can access secondary memory 140.
  • Secondary memory 140 is much larger, and therefore can hold more data, than primary memory 130.
  • secondary memory 140 is divided into pages and the pages are swapped in and out of primary memory 130 as they are needed by central processing unit 110. In this way, central processing unit 110 can access more memory than is available in primary memory 130 at a speed that is roughly the same as if all of the memory in secondary memory 140 could be accessed with the speed of primary memory 130.
  • Central processing unit 110 is supported by a data cache 170.
  • Data cache 170 is smaller than primary memory 130 but can be accessed faster than can primary memory 130.
  • data cache 170 is first checked to determine whether the data is available there. If the data is not available in data cache 170, then primary memory 130 is checked to determine whether primary memory 130 includes a page that contains the required data. If primary memory 130 does not have the required data, then a page containing the required data is copied into primary memory 130 from secondary memory 140 so that it can be accessed by central processing unit 110. At least part of the required data will also be copied into data cache 170 so that it can be accessed quickly by central processing unit 110.
  • Translation lookaside buffer (TLB) 150 is a hardware device that contains virtual address to physical address translation information for a small set of recently, or frequently, used virtual addresses. Because TLB 150 is implemented in hardware, it can quickly provide central processing unit 110 with a physical address translation of a requested virtual address. TLB 150, however, contains translations for only a small set of virtual addresses. If TLB 150 cannot provide the address translation requested by central processing unit 110, then translation storage buffer 160 is checked to see if it can provide the address translation requested.
  • Translation storage buffer 160 contains more address translations than does TLB 150. Because translation buffer 160 is stored in primary memory 130, it cannot be accessed by central processing unit 110 as quickly as can TLB 150. On the other hand, because translation storage buffer 160 is larger than TLB 150, TSB 160 provides more opportunities to find the requested address translation than does TLB 150.
  • TSB 160 resides in primary memory 130, it can be treated by system 100 as any other memory object and therefore all, or part of TSB 160 can migrate into data cache 170.
  • TLB 150 cannot provide a physical address translation of a virtual address for CPU 110, there is a good chance that CPU 110 will find the requested translation in TSB 160.
  • TSB 160 will often enable the desired virtual address to physical address translation to be provided quickly to central processing unit 110 in cases where the requested translation cannot be provided by TLB 150.
  • Level I is depicted as being a memory management unit (MMU) such as translation lookaside buffer (TLB) 220.
  • MMU memory management unit
  • TLB translation lookaside buffer
  • Level I is typically implemented in hardware, and is characterized by providing a fast translation for a small number of addresses.
  • the set of addresses that can be translated at Level I changes dynamically with elements of the set being updated in order to insure that the most likely needed translations will be available at Level I.
  • Level II is exemplified in FIG. 2 by translation storage buffer (TSB) 250.
  • TSB 250 is implemented in primary memory. Therefore, TSB 250, while not as fast as TLB 220, can maintain a larger set of translation elements than can TLB 220. Thus, when TLB 220 is unable to provide a translation (a TLB "miss"), TSB 250 is tested to determine whether the translation is stored in TSB 250. If the translation can be provided by TSB 250 (a TSB "hit") then the translation is provided to the CPU and to the translation lookaside buffer 220. If, on the other hand, TSB 250 cannot provide the translation (a TSB "miss”) then a Level III translation is performed.
  • a Level III translation is characterized by being slow but also as being comprehensive.
  • the virtual memory operating system will provide a big software translation table (BST) 280 that is stored in data storage.
  • the data storage can be primary memory, secondary memory or a combination of primary and secondary memory.
  • BST 280 contains every translation from virtual address space to physical address space that exists. Therefore, while a Level III translation will take a relatively long time, it will result in a translation being found if one exists.
  • file results are provided to the CPU, to the translation storage buffer 250 and to the translation lookaside buffer 220.
  • a Level I translation is performed when the CPU generates a virtual address request (210) to a Level I translator such as TLB 220. If the Level I translation is successful (a hit), the Level I translator will provide the requested physical address (230)that corresponds to the requested virtual address 210.
  • Level I translation fails (a miss)
  • a virtual address request 240 will be made to the Level II translator such as TSB 250. If the Level II translation is successful, then the physical address 260 corresponding to the virtual address request 240 will be returned to the Level I translator (TLB 220) and then to the CPU (physical address 230).
  • Level II translation fails, then virtual address request 270 will be sent to a Level III translator such as BST 280. Because BST 280 is comprehensive, if a physical address translation exists, it will be returned to the Level II translator (TSB 250) as physical address 290. The translation will also be provided to the Level I translator (TLB 220) as physical address 260 and will ultimately be provided to the CPU as physical address 230.
  • BST 280 is comprehensive, if a physical address translation exists, it will be returned to the Level II translator (TSB 250) as physical address 290.
  • the translation will also be provided to the Level I translator (TLB 220) as physical address 260 and will ultimately be provided to the CPU as physical address 230.
  • FIG. 3 further illustrates the derivation of a translation storage buffer index and target.
  • physical address 310 is depicted as being comprised of a physical page offset 320 and a physical page number 330.
  • a physical page is a contiguous portion of physical primary memory.
  • the physical page can be thought of as being divided into a set of "rows" wherein a row represents the smallest set of data that can be addressed individually.
  • Physical page offset 320 specifies the row being addressed within a given physical page and physical page number 330 specifies which page from the set of physical pages is being selected.
  • Physical pages can be of varying sizes. Thus, a big page will have more rows than a small page. It follows that a given amount of physical memory can be divided into fewer big pages than can be divided into small pages. Therefore, more bits of physical address 310 will be devoted to physical page offset 320 for a big page than for a small page. Correspondingly, fewer bits of physical address 310 will be devoted to physical page number 330 for a big page than for a small page.
  • the physical address 310 can be represented as an M+N bit binary number wherein M bits of physical address 310 are used to designate physical page number 330 and N bits of physical address 310 are used to designate physical page offset 320.
  • a 40 bit physical address space is used and is mapped using small pages that are 8KB in size and big pages that are 64KB in size.
  • physical address 310 is a 40 bit binary number.
  • the 13 least significant bits (LSbs) of physical address 310 will represent physical page offset 320.
  • the 27 most significant bits (MSbs) of physical address 310 will specify the physical page number 330 for a given small page.
  • the 16 least significant bits of physical address 310 will be used to provide the physical page offset 320 for the big page.
  • the 24 MSbs of physical address 310 will be used to specify a given physical page number 330 for a given big page.
  • virtual address 340 is divided into a virtual page offset 350 and a virtual page number 360 in a manner similar to the division of physical address 310.
  • the virtual page offset 350 will equal the physical page offset 320.
  • Virtual memory techniques are typically used to map a larger virtual address space into a smaller physical address space, so that for a given page size, there will usually be more virtual pages than physical pages.
  • virtual address 340 is a 64 bit binary number representing a virtual address space of size 2 64 .
  • virtual page offset 350 will be the 13 LSbs of virtual address 340 and virtual page number 360 will be the 51 MSbs of virtual address 340.
  • virtual page offset 350 will be the 16 LSbs of virtual address 340 and virtual page number 360 will be the 48 MSbs of virtual address 340.
  • virtual page offset 350 will be the 16 LSbs of virtual address 340 and virtual page number 360 will be the 48 MSbs of virtual address 340.
  • FIG. 1 An example embodiment having a 40 bit physical address space, a 64 bit virtual address space and big and small page sizes of 64KB and 8KB, respectively will be described.
  • different sizes can be used for the virtual and physical address spaces, as well as for the big and small pages.
  • an alternate embodiment exists wherein more than two page sizes can be used.
  • a direct-mapped translation storage buffer is used to provide virtual address to physical address translations.
  • the translation storage buffer contains translation table entries (TTEs). Each TTE provides a particular translation from a virtual address to a physical address. If the translation storage buffer is constrained to contain a positive integral power of two number of TTEs, then an N bit TSB index 370 can be used to directly map into the TSB buffer where N is the base two logarithm of the number of entries of the TSB. Therefore, in an example embodiment wherein a translation storage buffer having 512 entries is employed, the index into the translation storage buffer can be represented by a 9 bit binary number.
  • the virtual page number 360 can be divided into two fields.
  • the first field is the N least significant bits of the virtual page number 360 and is the TSB index 370.
  • the second field is the remaining most significant bits of virtual page number 360.
  • This field is a translation storage buffer target 380.
  • Target 380 can be used to determine whether the TTE entry stored at the location specified by TSB index 370 is the TTE for the given virtual address 340.
  • TSB index 370 the 9 LSbs of a virtual page number 360 for a given virtual address 340 will provide TSB index 370.
  • the remaining MSbs of the virtual page number 360 will provide the translation storage buffer target 380.
  • TSB index 370 is defined to be a known number of least significant bits of virtual page number 360, but without knowing the page size, the least significant bits of virtual page number 360 cannot be determined.
  • the solution to this dilemma is to calculate the TSB index 370 and TSB target 380 for each possible page size and then, for each page size, check the TTE corresponding to the TSB index for that page size against the TSB target until a TSB hit is found. If no hit is found after checking the TTE corresponding to the TSB index for each page size, then the TSB does not contain the requested translation and a higher level translator (e.g. BST) is employed.
  • a higher level translator e.g. BST
  • translation storage buffer 400 is a contiguous set of memory containing a 2 N element array of translation information.
  • tag O 410 and data O 420 together represent the first translation table entry of TSB 400 and tag 2 N -1 430 and tag data 2 N -1 440 together represent the last translation table entry of TSB 400.
  • the tag field of the candidate TEE is compared to the tag field being sought. If they do not match, then the TTE does not correspond to the virtual address to be translated. If, on the other hand, the tag of the candidate TEE matches the tag calculated from the virtual address to be translated, then the physical address for the virtual address to be translated can be found in the corresponding data field of the candidate TEE and a TSB hit has occurred.
  • Base 450 points to the start of the translation storage buffer 400. Because the size of each TEE is known, a pointer into TSB 400 can be calculated for a given TSB index by adding the base 450 address to the product of the TSB index and the size of an individual TTE. Because base 450 can be specified as a virtual address, the pointers into the translation storage buffer can also be virtual addresses. The entire translation storage buffer 400 can be stored in virtual memory just as can any other memory object. Therefore, as mentioned above, in an embodiment of a computer system that has a data storage cache, all, or part, of TSB 400 will migrate in and out of the data storage cache. When the candidate TTEs of TSB 400 are available in the data cache, the look-up of the translation storage buffer will be very fast.
  • recursive virtual address translations will be required since the pointers used to access TSB 400 will themselves need to be translated from virtual to physical addresses.
  • Recursive virtual address translations can be provided in many ways.
  • the translation information for the virtual addresses of the pointers to the TSB are stored in an entry of the translation lookaside buffer that is locked into the TLB so that the pointer translations will always be available within the translation lookaside buffer.
  • TSBs can be implemented thereby effectively providing a set-associative TSB.
  • big and small pointers are calculated for each TSB and one TTE per TSB is checked for each page size. If an additional TSB is appended to TSB 400, the big pointer to the additional TSB is calculated by adding the size of TSB 400 to the big pointer of TSB 400. Similarly, the small pointer to the additional TSB is calculated by adding the size of TSB 400 to the small pointer of TSB 400.
  • FIG. 5 a sample translation table entry of a translation storage buffer is illustrated.
  • translation table entry 500 is divided into two parts, TSB tag 510 and TSB data 550.
  • TSB tag field 510 contains all of the information necessary to determine whether the translation table entry 500 Corresponds to the virtual address to be translated.
  • TSB data field 550 provides all of the information necessary to translate the virtual address to be translated into a physical address.
  • TSB tag 510 is comprised of three fields.
  • the first field contains the most significant bits of the virtual address 520 for the virtual address that is translated by TTE 500.
  • the minimum number of most significant bits of the virtual address to include in field 520 can be calculated by subtracting the sum of the number of bits in the page offset for the smallest page mapping and the number of bits required to specify the TSB index from the total number of bits in the virtual address.
  • the 42 most significant bits (64-(13+9)) of the virtual address corresponding to TTE 500 will be stored in field 520.
  • the big page size is 64KB, then only the 39 most significant bits (64-(16+9)) of the virtual address are required to determine a match for a big page. So long as the number of bits in the page offset for the biggest page does not exceed the sum of the number of bits for the page offset of the smallest page and the number of bits required to represent the virtual index, there is no harm in including "extra bits" of the most significant bits for the big pages.
  • each context can have a separate virtual address space or, alternately, some contexts can use the same virtual address space thereby providing a "global" virtual address space.
  • Global bit 540 is used to specify whether translation table entry 500 is context specific or not. If global bit 540 is set, then a hit will occur whenever field 520 matches the most significant bits of the virtual address to be translated regardless of the context requesting the translation. If, on the other hand, global bit 540 is not set, then the virtual address corresponding to translation table entry 500 is context specific and a hit will only occur if field 520 matches the most significant bits of the virtual address to be translated and the value stored in context number field 530 matches the context number for the context requesting the virtual address to be translated.
  • TSB data 550 includes page attributes field 560, physical page number field 570, page size field 580, and valid bit 590.
  • Physical page number field 570 contains the physical page number corresponding to the virtual address translated by TTE 500.
  • Physical page number 570 is a field that has as many bits allocated to it as are required to specify the physical page number for the smallest page size being used in the mapping. Thus, in the case where the physical address space is specified by 40 bits, and the smallest physical page mapping is an 8KB page, physical page number field 570 will be 27 bits in size (40-13).
  • translation table entry 500 corresponds to a mapping for a page size bigger than the smallest page size mapping
  • only the appropriate number of most significant bits of physical page number field 570 are used to contain the physical number for the translation table entry and the remaining least significant bits are ignored.
  • the 24 most significant bits of physical page number field 570 would be needed to specify the physical page number for the translation and the remaining three least significant bits of physical page number field 570 would be ignored.
  • the least significant bits of TSB data 550 are used to store page attributes 560 for the page being translated.
  • page attributes are bits that indicate whether the page of the translation can be cached or whether the page has read/write or read only privileges.
  • TSB data field 550 also includes a page size field 580.
  • page size field 580 is two bits in size and can be used to specify the page size of the four possible page sizes corresponding to the page of translation table entry 500.
  • TSB data field 550 has a valid bit 590 that is the most significant bit of TSB data field 550. If valid bit field 590 is set, then the translation stored in TTE 500 is valid and can be used. If, on the other hand the valid bit 590 is not set, then the translation table entry 500 is not valid and cannot be used to provide a translation. As will be discussed below in connection with FIG. 9, by placing the valid bit in the most significant bit of TSB data field 550, this embodiment exploits the fact that the sign of an integer number is specified in the most significant bit of a representation of the number. As will be seen, a branch on negative instruction can be used to efficiently branch when a valid translation has been found in the translation storage buffer.
  • tag access register 610 is a register that contains the context number of the context requesting a Virtual address to be translated and the most significant bits of the virtual address to be translated that are required by the translation storage buffer lookup.
  • the tag access register 610 is a 64 bit register that contains the 51 most significant bits of the virtual address to be translated in the 51 most significant bits of the tag access register 610.
  • the 13 least significant bits of tag access register 610 are used to store the context number that requested the virtual address translation.
  • tag access register 610 In the case where the translation storage buffer lookup is being performed after failure of a translation lookaside buffer lookup, tag access register 610 would be filled by the translation lookaside buffer when it requested a translation storage buffer lookup be performed. If, on the other hand, the translation storage buffer were being used as a Level I translator, then tag access register 610 would be filled by the CPU requesting the address translation at the time that it requested the address translation.
  • TSB tag target logic 620 is bit shifting logic that fills TSB tag target register 630 with a TSB tag target to be compared with the TSB tag of each translation table entry of the translation storage buffer to be checked.
  • TSB pointer logic 640 takes the value stored in the tag access register 610 and calculates one TSB pointer per page size. The TSB pointers are then stored in TSB pointers register 650. In order to perform the calculation of the TSB pointers, TSB pointer logic 640 is supplied with TSB parameters 660. The values stored in TSB parameters 660 are set at the time that the operating system is initialized. In an alternate embodiment, the parameters are dynamic and are changed as changes are needed. For example, the values of the TSB parameters can be changed during a context switch. By storing the information for the TSB parameters 660 in registers, a high degree of flexibility can be maintained in structuring the TSB. This flexibility permits the TSB to be tuned so that it will operate most effectively for a particular computer system implementation.
  • the TSB parameters 660 that can be set at initialization include TSB split 662, TSB bases 664, number of page sizes 666, page sizes 668 and TSB size 670.
  • the parameter TSB split 662 is a one bit parameter.
  • TSB split is set to zero, a translation storage buffer structure such as that of translation storage buffer 400 of FIG. 4 is established. In that structure, big and small page translation table entries are stored intermixed in the same translation storage buffer and the big and small pointers point to entries of the same translation table entry array.
  • FIG. 7 an alternate translation storage buffer 700 is depicted. This structure is implemented when the parameter TSB split 662 of FIG. 6 is set to equal one.
  • the translation storage buffer 700 of FIG. 7 has been split or bifurcated into two separate sub translation storage buffers. There is a small page sub translation storage buffer 710 and a big page sub translation storage buffer 750.
  • Small page sub translation storage buffer 710 has a base value small base 740 and a pointer small pointer 745 into the array of translation table entries for small pages.
  • big page sub translation storage buffer 750 has a base value of big base 780 and a pointer of big pointer 785 into the big page translation table entries.
  • Translation storage buffer 700 has the page size included implicitly in the structure, therefore there is no need to check the page size after a candidate tag has been found to match the tag target. This is because small page translation table 710 is comprised only of translation table entries for small pages and big page translation storage buffer 750 is comprised only of translation tattle entries for big pages.
  • the structure of translation storage buffer 400 can be more efficient that the structure of translation storage buffer 700.
  • the need to translate virtual address may be dominated by either the small or big page size translations.
  • either small page sub TSB 710 or big page sub TSB 750 will be under-utilized while the other sub TSB is over-utilized.
  • one of the sub TSBs will be thrashing while the other will be accessed only infrequently.
  • the number of page sizes parameter 666 is set to two, when TSB split parameter 662 is set to one, the split will occur into two sub TSBs. If, on the other hand, the number of page sizes parameter 666 is a number other than two, setting the TSB split parameter 662 will cause the TSB structure to be split into a corresponding number of sub TSBs.
  • TSB bases 664 if the TSB split has not been specified, only one TSB base will be stored in TSB bases 664. If, on the other hand, the TSB split has been specified, the bases for each of the sub TSBs will be stored as parameters in TSB bases 664.
  • the field page sizes 668 will store the actual sizes of the pages being translated. Thus, in the case where there are big and small pages, there will be two entries in page sizes 668, one entry will include the size of a big page and the other the size of a small page. In an alternate embodiment, the sizes could be implicit in hardware and therefore need not be stored in registers.
  • parameter TSB size 670 includes the number of entries in a particular translation storage buffer.
  • TSB size will specify the number of entries in a sub TSB. In an alternate embodiment, there will be a separate TSB size for each sub TSB.
  • TSB pointer logic 640 can quickly calculate the TSB pointers to be used in a TSB lookup and store this information in TSB pointers register 650.
  • the TSB lookup process 800 begins at start bubble 810 and proceeds to process block 820 where a translation lookaside buffer miss is detected thereby causing the loading of the tag access register.
  • the translation storage buffer pointers are calculated using the contents of the tag access register and the parameters that describe the TSB.
  • the TSB tag target is also calculated in process block 830.
  • the TSB pointers and tag target are calculated by hardware as illustrated in FIG. 6. Alternately, the TSB pointers and tag target can be calculated by software.
  • process block 840 a process is performed in process block 840 to check to determine whether one of the translation table entries of the translation storage buffer pointed to by the TSB pointers contains the translation for the virtual address to be translated.
  • An embodiment of a procedure for checking for TSB hits will be discussed below in greater detail in connection with FIG. 9.
  • decision block 850 determines whether or not a TSB hit occurred. If a hit occurred then decision path 854 is taken to process block 860 where the translation lookaside buffer is loaded with the translation data required to perform the virtual address to physical address translation requested.
  • the translation lookaside buffer has been structured so that it has a translation lookaside buffer tag and translation lookaside buffer data field for each translation entry of the TLB.
  • the translation lookaside buffer data field has the same structure as the translation storage buffer data field 550 of FIG. 5. Because the translation lookaside buffer tag field has information that is known at the time that the request was made for a translation, at the time that a translation lookaside buffer miss is detected, a tag register is filled with the contents of a translation lookaside buffer tag field formatted so that it will comply with the structure of a translation lookaside buffer tag.
  • the translation lookaside buffer tag for the translation entry can be provided directly from the tag register of the translation lookaside buffer. Furthermore, because the translation storage buffer data field is formatted to be the same as that of the translation lookaside buffer data field, the translation information can be loaded into the translation lookaside buffer quickly since there is no need to reformat the field contents.
  • process block 870 a miss has occurred because the TSB does not contain the translation for the virtual address to be translated. Therefore, a TSB miss handling procedure is executed.
  • the miss handling procedure upon a TSB miss, the miss handling procedure will consult the big software translation table 280 managed by the virtual memory operating system to retrieve the address translation if it exists.
  • the miss handler could use a translator of an intermediate level between Level II and Level III before the Level III translator is consulted.
  • An example of an intermediate level translator would be yet another translation storage buffer.
  • FIG. 9 a flow chart is illustrated of a procedure that checks to determine whether a TSB hit has occurred.
  • the procedure check for TSB hit 900 begins at start bubble 910 and proceeds to process block 920 where the translation storage buffer pointers and tag target are retrieved from their respective registers. Then, in process block 930, the translation storage buffer candidate tags stored in the translation storage buffer at the locations pointed to by the translation storage buffer pointers are retrieved. This is followed by the retrieval of the corresponding translation storage buffer candidate data in process block 940.
  • a candidate tag for each page size is tested against the target tag in process block 950.
  • decision block 960 a test is made to determine whether a match was found. If no match was found, then decision path 964 is taken to process block 970 where a miss is declared. Processing then completes at terminal bubble 995 with a return from the procedure.
  • decision path 962 is taken to decision block 980 where a test is performed to determine whether the matching translation table entry is valid.
  • the valid bit data field for the successful candidate is the most significant bit, the valid bit can be tested by a branch on negative instruction as discussed previously.
  • decision path 982 is taken to process block 970 and a miss is declared. If, on the other hand, the matching translation table entry is valid, then decision path 984 is taken to process block 990 where a hit is declared before terminating the procedure at terminal bubble 995.
  • procedure 900 is implemented in hardware. This permits the checking of the TSB entries that are pointed to by the pointers for each page size to be checked in a parallel manner. Alternately, procedure 900 is performed through the execution of software. A software implementation provides greater flexibility than that afforded by hardware. Furthermore, procedure 900 is structured so that operations within the procedure can be interleaved. Interleaving permits the checking of multiple entries of the translation storage buffer without a linear increase in the amount of time necessary to make the checks. Thus, if only one TSB entry were being tested, idle cycles would be "wasted" while the central processing unit was waiting for operations that have a long latency to complete.
  • TSB translation table entry is checked for each page size, some of the idle cycles caused by a test for a first page size can be used to perform the test for the second page size. Moreover, if procedure 900 is executed on a superscalar processor (i.e. one that can execute multiple instructions concurrently), the translation table entries of the TSB can be checked in parallel even in a software implementation.

Abstract

A method and apparatus for translating a virtual address to a physical address. A virtual address to be translated has a virtual page offset and a virtual page number. The virtual address to be translated addresses a page of memory. The size of this page is unknown. There are L different possible page sizes where L is a positive integer greater than one. Each of the L different page sizes is selected to be a test page size and a test is performed. During the test, a pointer into a translation storage buffer is calculated. The pointer is calculated from the virtual address to be translated by assuming that the virtual address to be translated corresponds to a mapping of the test page size. The pointer points to a candidate translation table entry of the translation storage buffer. The candidate translation table entry has a candidate tag and candidate data. The candidate tag identifies a particular virtual address and the candidate data identifies a particular physical address corresponding to the particular virtual address. A virtual address target tag is extracted from the virtual address to be translated. The virtual address target tag is calculated by assuming that the virtual address to be translated corresponds to a mapping of the test page size. The target tag and the candidate tag are then compared. If the target tag matches the candidate tag, the candidate data is provided as the physical address translation corresponding to the virtual address to be translated.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to the field of computer systems. More specifically, the present invention relates to translation from a virtual address to a physical address in a virtual memory computer system.
2. Art Background
Virtual memory is a technique that allows an application to see the system as providing a large uniform primary memory, which in reality may be smaller, more fragmented, and/or partially simulated by secondary storage such as a hard disk. Applications access memory through virtual addresses, which are translated (mapped) by special hardware onto physical addresses. Paging and segmentation are two common implementations of virtual memory.
When implementing virtual memory using the paging technique, the virtual address space is divided into a number of fixed-size blocks called pages, each of which can be mapped onto any of the similarly sized physical pages available on the system. Typically, special memory management hardware such as a memory management unit (MMU) or paged memory management unit (PMMU) performs the address translation from virtual addresses to physical addresses. In this type of memory management, any attempt to access data that is not present in physical (system) memory causes the PMMU to send an interrupt signal to the central processing unit (CPU); the operating system then transfers the data in from other storage (such as hard disk), without the application "knowing" about the transfer.
In one virtual memory scheme, there is only one mapping from virtual memory space to physical memory space. In other implementations however, there can be several address spaces, each with its own mapping to the physical address space. Thus, a first process may refer to a given physical address using a first virtual address and a second process running; on the same CPU, or on a different CPU in a multi-processor environment, will refer to the given physical address by a second virtual address. In such a case, the two virtual addresses corresponding to the same physical address are referred to as aliases.
Some CPUs are able to operate using more than one size of virtual page. This permits the virtual page size used in a mapping to be selected to best meet the needs of the virtual address space being mapped. If one uses big pages, a large amount of virtual memory can be translated using a single entry in the MMU. Therefore, fewer resources are needed for the MMU thereby increasing performance. The use of big pages can, however, lead to memory fragmentation problems not encountered when small pages are used.
SUMMARY OF THE INVENTION
A method and apparatus for translating a virtual address to a physical address is disclosed. A virtual address to be translated has a virtual page offset and a virtual page number. The virtual address to be translated addresses a page of memory. The size of this page is unknown. The unknown page size is one of L different possible page sizes where L is a positive integer greater than one.
Each of the L different page sizes is selected to be a test page size and a test is performed. During the test, a pointer into a translation storage buffer is calculated. The pointer is calculated from the virtual address to be translated by assuming that the virtual address to be translated corresponds to a mapping of the test page size. The pointer points to a candidate translation table entry of the translation storage buffer. The candidate translation table entry has a candidate tag and candidate data. The candidate tag identifies a particular virtual address and the candidate data identifies a particular physical address corresponding to the particular virtual address.
A virtual address target tag is extracted from the virtual address to be translated. The virtual address target tag is calculated by assuming that the virtual address to be translated corresponds to a mapping of the test page size. The target tag and the candidate tag are then compared. If the target tag matches the candidate tag, the candidate data is provided as the physical address translation corresponding to the virtual address to be translated.
BRIEF DESCRIPTION OF THE DRAWINGS
The objects, features, and advantages of the method and apparatus of the present invention will be apparent from the following detailed description of the invention in which:
FIG. 1 illustrates a computer system that uses an embodiment of the present invention to perform address translation from a virtual address to a physical address;
FIG. 2 illustrates a multi-level approach for translating virtual addresses to physical addresses;
FIG. 3 illustrates the relationship between physical addresses and virtual addresses and how a translation storage buffer index and target are derived from a virtual address;
FIG. 4 illustrates an embodiment of a translation storage buffer that supports multiple page sizes;
FIG. 5 illustrates a translation table entry within a translation storage buffer;
FIG. 6 illustrates the use of hardware logic by the present invention to quickly derive a tag target and translation storage buffer pointers;
FIG. 7 illustrates an alternate embodiment of a translation storage buffer that supports multiple page sizes;
FIG. 8 is a flow chart illustrating a translation storage buffer look-up as performed by an embodiment of the present invention; and
FIG. 9 is a flow chart of a procedure employed by an embodiment of the present invention to check for a translation storage buffer hit.
DETAILED DESCRIPTION OF THE INVENTION
A method and apparatus for implementation of a virtual address to physical address translation cache that supports multiple page sizes is disclosed. In an embodiment described below, hardware provides support for a software search of the translation cache by calculating the tag value to be compared and the pointers into the cache where the comparison is to be made. The hardware support described is simple to implement and efficient. If a translation exists within the translation cache, the translation can be performed by retrieving one translation entry. Because the embodiment permits the translation to be retrieved from the cache by software, frequently used portions of the translation cache can themselves, be cached in a virtually-indexed data cache.
Moreover, in an embodiment that uses a software implemented search structure, great flexibility in the structure of the translation cache can be achieved. Thus, the translation cache can be configured to be one cache that intermixes translations for different page sizes or configured as one cache divided into sub caches wherein each sub cache contains translations for a single page size. Multiple translation caches can also be employed. Therefore, in one embodiment, separate translation caches can be provided, one for instructions and the other for data. Alternate embodiments can also be implemented wherein there is one global translation cache for all CPUs and contexts; one translation cache per CPU, with each translation cache containing all contexts; one translation cache per context, but shared by all CPUs, or one translation cache per context per CPU. Finally, an embodiment can be implemented where system translations are treated as a separate context thereby permitting separate translation caches for user translations and for system translations.
In the following description, for purposes of explanation, specific numbers, materials and configurations are set forth in order to provide a thorough understanding of the present invention. It will be apparent to one skilled in the art, however, that the present invention may be practiced without the specific details. In other instances, well known systems are shown in diagrammatic or block diagram form in order not to obscure the present invention unnecessarily.
Referring now to FIG. 1, an embodiment of a virtual memory computer system is illustrated that uses a virtual address (VA) to physical address (PA) translation cache that supports multiple page sizes. In computer system 100 of FIG. 1, central processing unit (CPU) 110 is coupled to primary memory 130 and secondary memory 140 by a memory interconnect such as memory bus 120. In an embodiment of computer system 100, central processing unit 110 can access primary memory 130 faster than it can access secondary memory 140. Secondary memory 140, however, is much larger, and therefore can hold more data, than primary memory 130. In this embodiment, secondary memory 140 is divided into pages and the pages are swapped in and out of primary memory 130 as they are needed by central processing unit 110. In this way, central processing unit 110 can access more memory than is available in primary memory 130 at a speed that is roughly the same as if all of the memory in secondary memory 140 could be accessed with the speed of primary memory 130.
Central processing unit 110 is supported by a data cache 170. Data cache 170 is smaller than primary memory 130 but can be accessed faster than can primary memory 130. Thus, when central processing unit 110 needs data, data cache 170 is first checked to determine whether the data is available there. If the data is not available in data cache 170, then primary memory 130 is checked to determine whether primary memory 130 includes a page that contains the required data. If primary memory 130 does not have the required data, then a page containing the required data is copied into primary memory 130 from secondary memory 140 so that it can be accessed by central processing unit 110. At least part of the required data will also be copied into data cache 170 so that it can be accessed quickly by central processing unit 110.
When a process executing within central processing unit 110 requires data, the process will specify the required data using a virtual address. Translation lookaside buffer (TLB) 150 is a hardware device that contains virtual address to physical address translation information for a small set of recently, or frequently, used virtual addresses. Because TLB 150 is implemented in hardware, it can quickly provide central processing unit 110 with a physical address translation of a requested virtual address. TLB 150, however, contains translations for only a small set of virtual addresses. If TLB 150 cannot provide the address translation requested by central processing unit 110, then translation storage buffer 160 is checked to see if it can provide the address translation requested.
Translation storage buffer 160 contains more address translations than does TLB 150. Because translation buffer 160 is stored in primary memory 130, it cannot be accessed by central processing unit 110 as quickly as can TLB 150. On the other hand, because translation storage buffer 160 is larger than TLB 150, TSB 160 provides more opportunities to find the requested address translation than does TLB 150.
Furthermore, because TSB 160 resides in primary memory 130, it can be treated by system 100 as any other memory object and therefore all, or part of TSB 160 can migrate into data cache 170. Thus, if TLB 150 cannot provide a physical address translation of a virtual address for CPU 110, there is a good chance that CPU 110 will find the requested translation in TSB 160. Moreover, there is also a good chance that the portion of TSB 160 that contains the requested translation will reside in data cache 170. Thus, TSB 160 will often enable the desired virtual address to physical address translation to be provided quickly to central processing unit 110 in cases where the requested translation cannot be provided by TLB 150.
Referring now to FIG. 2, a hierarchical translation structure for one embodiment of the present invention is depicted. In FIG. 2, a three-level translation hierarchy is illustrated. Whenever a CPU requires a virtual address to be translated to a physical address, a translation attempt will be made at Level I. In FIG. 2, Level I is depicted as being a memory management unit (MMU) such as translation lookaside buffer (TLB) 220. Level I is typically implemented in hardware, and is characterized by providing a fast translation for a small number of addresses. The set of addresses that can be translated at Level I changes dynamically with elements of the set being updated in order to insure that the most likely needed translations will be available at Level I.
Level II is exemplified in FIG. 2 by translation storage buffer (TSB) 250. In an embodiment of the present invention, TSB 250 is implemented in primary memory. Therefore, TSB 250, while not as fast as TLB 220, can maintain a larger set of translation elements than can TLB 220. Thus, when TLB 220 is unable to provide a translation (a TLB "miss"), TSB 250 is tested to determine whether the translation is stored in TSB 250. If the translation can be provided by TSB 250 (a TSB "hit") then the translation is provided to the CPU and to the translation lookaside buffer 220. If, on the other hand, TSB 250 cannot provide the translation (a TSB "miss") then a Level III translation is performed.
A Level III translation is characterized by being slow but also as being comprehensive. Typically, the virtual memory operating system will provide a big software translation table (BST) 280 that is stored in data storage. The data storage can be primary memory, secondary memory or a combination of primary and secondary memory. BST 280 contains every translation from virtual address space to physical address space that exists. Therefore, while a Level III translation will take a relatively long time, it will result in a translation being found if one exists. When a Level III translation is made, file results are provided to the CPU, to the translation storage buffer 250 and to the translation lookaside buffer 220.
To summarize FIG. 2, a Level I translation is performed when the CPU generates a virtual address request (210) to a Level I translator such as TLB 220. If the Level I translation is successful (a hit), the Level I translator will provide the requested physical address (230)that corresponds to the requested virtual address 210.
If the Level I translation fails (a miss), then a virtual address request 240 will be made to the Level II translator such as TSB 250. If the Level II translation is successful, then the physical address 260 corresponding to the virtual address request 240 will be returned to the Level I translator (TLB 220) and then to the CPU (physical address 230).
If, on the other hand, the Level II translation fails, then virtual address request 270 will be sent to a Level III translator such as BST 280. Because BST 280 is comprehensive, if a physical address translation exists, it will be returned to the Level II translator (TSB 250) as physical address 290. The translation will also be provided to the Level I translator (TLB 220) as physical address 260 and will ultimately be provided to the CPU as physical address 230.
While the present invention is being described herein in an example embodiment containing three levels of translation, alternate embodiments exist that contain more, or less, levels of translation. Furthermore, while the present invention is described in an embodiment of a Level II translation storage buffer such as TSB 250, an alternate embodiment of the present invention would be one wherein a translation storage buffer was used as a Level I translator.
Referring now to FIG. 3, the relationship between physical and virtual addresses is depicted. FIG. 3 further illustrates the derivation of a translation storage buffer index and target. In FIG. 3, physical address 310 is depicted as being comprised of a physical page offset 320 and a physical page number 330. A physical page is a contiguous portion of physical primary memory. Conceptually, the physical page can be thought of as being divided into a set of "rows" wherein a row represents the smallest set of data that can be addressed individually. Physical page offset 320 then specifies the row being addressed within a given physical page and physical page number 330 specifies which page from the set of physical pages is being selected.
Physical pages can be of varying sizes. Thus, a big page will have more rows than a small page. It follows that a given amount of physical memory can be divided into fewer big pages than can be divided into small pages. Therefore, more bits of physical address 310 will be devoted to physical page offset 320 for a big page than for a small page. Correspondingly, fewer bits of physical address 310 will be devoted to physical page number 330 for a big page than for a small page. If the number of rows in a given page is constrained to be an integral power of two, and if the maximum number of physical pages for a given page size is also constrained to be an integral power of two, then it follows that the physical address 310 can be represented as an M+N bit binary number wherein M bits of physical address 310 are used to designate physical page number 330 and N bits of physical address 310 are used to designate physical page offset 320.
In one embodiment, a 40 bit physical address space is used and is mapped using small pages that are 8KB in size and big pages that are 64KB in size. In this embodiment, physical address 310 is a 40 bit binary number. When a small (8KB) page is being addressed, the 13 least significant bits (LSbs) of physical address 310 will represent physical page offset 320. The 27 most significant bits (MSbs) of physical address 310 will specify the physical page number 330 for a given small page. If, on the other hand, a large (64KB) page is being used, the 16 least significant bits of physical address 310 will be used to provide the physical page offset 320 for the big page. In the case of big pages, the 24 MSbs of physical address 310 will be used to specify a given physical page number 330 for a given big page.
Referring now to virtual address 340, it can be seen that virtual address 340 is divided into a virtual page offset 350 and a virtual page number 360 in a manner similar to the division of physical address 310. For a given page size, the virtual page offset 350 will equal the physical page offset 320. Virtual memory techniques are typically used to map a larger virtual address space into a smaller physical address space, so that for a given page size, there will usually be more virtual pages than physical pages. In an example embodiment, virtual address 340 is a 64 bit binary number representing a virtual address space of size 264. In the case where a small page of size 8KB is used, virtual page offset 350 will be the 13 LSbs of virtual address 340 and virtual page number 360 will be the 51 MSbs of virtual address 340. Similarly, in the case where a big page of size 64KB is used, virtual page offset 350 will be the 16 LSbs of virtual address 340 and virtual page number 360 will be the 48 MSbs of virtual address 340. Throughout the rest of the detailed description, unless otherwise noted, an example embodiment having a 40 bit physical address space, a 64 bit virtual address space and big and small page sizes of 64KB and 8KB, respectively will be described. In alternate embodiments, different sizes can be used for the virtual and physical address spaces, as well as for the big and small pages. Furthermore, an alternate embodiment exists wherein more than two page sizes can be used. Moreover, alternate embodiments exist where there is a larger physical address space than virtual address space.
In one embodiment of the present invention, a direct-mapped translation storage buffer is used to provide virtual address to physical address translations. In this embodiment, the translation storage buffer contains translation table entries (TTEs). Each TTE provides a particular translation from a virtual address to a physical address. If the translation storage buffer is constrained to contain a positive integral power of two number of TTEs, then an N bit TSB index 370 can be used to directly map into the TSB buffer where N is the base two logarithm of the number of entries of the TSB. Therefore, in an example embodiment wherein a translation storage buffer having 512 entries is employed, the index into the translation storage buffer can be represented by a 9 bit binary number.
Conceptually, the virtual page number 360 can be divided into two fields. The first field is the N least significant bits of the virtual page number 360 and is the TSB index 370. The second field is the remaining most significant bits of virtual page number 360. This field is a translation storage buffer target 380. Target 380 can be used to determine whether the TTE entry stored at the location specified by TSB index 370 is the TTE for the given virtual address 340. Thus, in the case of a 512 entry TSB, the 9 LSbs of a virtual page number 360 for a given virtual address 340 will provide TSB index 370. The remaining MSbs of the virtual page number 360 will provide the translation storage buffer target 380.
The problem with the above described TSB indexing scheme is that at the time that a virtual address is to be translated, the page size corresponding to the virtual address space of the virtual address to be translated will not be known. Because the page size will not be known, it cannot be determined which part of virtual address 340 is devoted to virtual page offset 350 and which part is devoted to virtual page number 360. This is a problem because TSB index 370 is defined to be a known number of least significant bits of virtual page number 360, but without knowing the page size, the least significant bits of virtual page number 360 cannot be determined. The solution to this dilemma is to calculate the TSB index 370 and TSB target 380 for each possible page size and then, for each page size, check the TTE corresponding to the TSB index for that page size against the TSB target until a TSB hit is found. If no hit is found after checking the TTE corresponding to the TSB index for each page size, then the TSB does not contain the requested translation and a higher level translator (e.g. BST) is employed.
Referring now to FIG. 4, a translation storage buffer that intermixes translation table entries for big and small sized pages is illustrated. In FIG. 4, it can be seen that translation storage buffer 400 is a contiguous set of memory containing a 2N element array of translation information. Thus, tag O 410 and data O 420 together represent the first translation table entry of TSB 400 and tag 2N -1 430 and tag data 2N -1 440 together represent the last translation table entry of TSB 400.
When a TTE is tested, the tag field of the candidate TEE is compared to the tag field being sought. If they do not match, then the TTE does not correspond to the virtual address to be translated. If, on the other hand, the tag of the candidate TEE matches the tag calculated from the virtual address to be translated, then the physical address for the virtual address to be translated can be found in the corresponding data field of the candidate TEE and a TSB hit has occurred.
Thus, if the TEE containing tag O 410 and data O 420 were being tested, a comparison would be made between the translation storage buffer target corresponding to the virtual address to be translated and tag O 410. If tag O 410 matched the translation storage buffer target of the virtual address to be translated, then the information stored in data O 420 would provide the physical address for the virtual address to be translated.
As described above, in the case where there is a translation storage buffer 400 that maps two page sizes (small and big), then two entries of TSB 400 will be checked to see if there is a hit. Small pointer 460 will point to the candidate TEE of TSB 400 to be tested to determine whether a hit has occurred for a small page size virtual address and big pointer 470 will point to the candidate TEE of TSB 400 to be used to determine whether a TSB hit has occurred for the case where the virtual address to be translated is a big page size virtual address.
Base 450 points to the start of the translation storage buffer 400. Because the size of each TEE is known, a pointer into TSB 400 can be calculated for a given TSB index by adding the base 450 address to the product of the TSB index and the size of an individual TTE. Because base 450 can be specified as a virtual address, the pointers into the translation storage buffer can also be virtual addresses. The entire translation storage buffer 400 can be stored in virtual memory just as can any other memory object. Therefore, as mentioned above, in an embodiment of a computer system that has a data storage cache, all, or part, of TSB 400 will migrate in and out of the data storage cache. When the candidate TTEs of TSB 400 are available in the data cache, the look-up of the translation storage buffer will be very fast.
In the case where TSB 400 is virtually indexed, recursive virtual address translations will be required since the pointers used to access TSB 400 will themselves need to be translated from virtual to physical addresses. Recursive virtual address translations can be provided in many ways. In one embodiment, the translation information for the virtual addresses of the pointers to the TSB are stored in an entry of the translation lookaside buffer that is locked into the TLB so that the pointer translations will always be available within the translation lookaside buffer.
Multiple TSBs can be implemented thereby effectively providing a set-associative TSB. In a set-associative TSB embodiment, there is at least one additional translation storage buffer like TSB 400. In this embodiment, big and small pointers are calculated for each TSB and one TTE per TSB is checked for each page size. If an additional TSB is appended to TSB 400, the big pointer to the additional TSB is calculated by adding the size of TSB 400 to the big pointer of TSB 400. Similarly, the small pointer to the additional TSB is calculated by adding the size of TSB 400 to the small pointer of TSB 400.
Referring now to FIG. 5, a sample translation table entry of a translation storage buffer is illustrated. In FIG. 5, it can be seen that translation table entry 500 is divided into two parts, TSB tag 510 and TSB data 550. TSB tag field 510 contains all of the information necessary to determine whether the translation table entry 500 Corresponds to the virtual address to be translated. TSB data field 550, on the other hand, provides all of the information necessary to translate the virtual address to be translated into a physical address.
Referring now to TSB tag 510, it can be seen that TSB tag 510 is comprised of three fields. The first field contains the most significant bits of the virtual address 520 for the virtual address that is translated by TTE 500. The minimum number of most significant bits of the virtual address to include in field 520 can be calculated by subtracting the sum of the number of bits in the page offset for the smallest page mapping and the number of bits required to specify the TSB index from the total number of bits in the virtual address. Thus, in the example where the smallest page size to be mapped is 8KB, there is a 512 translation table entry (TTE) TSB and a 64 bit virtual address space, the 42 most significant bits (64-(13+9)) of the virtual address corresponding to TTE 500 will be stored in field 520. Note that, in this example, if the big page size is 64KB, then only the 39 most significant bits (64-(16+9)) of the virtual address are required to determine a match for a big page. So long as the number of bits in the page offset for the biggest page does not exceed the sum of the number of bits for the page offset of the smallest page and the number of bits required to represent the virtual index, there is no harm in including "extra bits" of the most significant bits for the big pages. The advantage of doing so is that under these circumstances, only one TSB tag need be calculated for the virtual address to be translated and that one TSB tag can be compared against a candidate TSB tag for each page size. In alternate embodiments that violate this constraint, one can calculate one TSB tag per page size to compare against the corresponding candidate TSB tag for that page size.
In the TTE embodiment of FIG. 5, each context can have a separate virtual address space or, alternately, some contexts can use the same virtual address space thereby providing a "global" virtual address space. Global bit 540 is used to specify whether translation table entry 500 is context specific or not. If global bit 540 is set, then a hit will occur whenever field 520 matches the most significant bits of the virtual address to be translated regardless of the context requesting the translation. If, on the other hand, global bit 540 is not set, then the virtual address corresponding to translation table entry 500 is context specific and a hit will only occur if field 520 matches the most significant bits of the virtual address to be translated and the value stored in context number field 530 matches the context number for the context requesting the virtual address to be translated.
Referring now to TSB data field 550, TSB data 550 includes page attributes field 560, physical page number field 570, page size field 580, and valid bit 590. Physical page number field 570 contains the physical page number corresponding to the virtual address translated by TTE 500. Physical page number 570 is a field that has as many bits allocated to it as are required to specify the physical page number for the smallest page size being used in the mapping. Thus, in the case where the physical address space is specified by 40 bits, and the smallest physical page mapping is an 8KB page, physical page number field 570 will be 27 bits in size (40-13). In the case where translation table entry 500 corresponds to a mapping for a page size bigger than the smallest page size mapping, only the appropriate number of most significant bits of physical page number field 570 are used to contain the physical number for the translation table entry and the remaining least significant bits are ignored. Thus, in the case where a 64KB page was being translated and the smallest page size was 8KB, only the 24 most significant bits of physical page number field 570 would be needed to specify the physical page number for the translation and the remaining three least significant bits of physical page number field 570 would be ignored.
In the translation table entry 500 embodiment of FIG. 5, the least significant bits of TSB data 550 are used to store page attributes 560 for the page being translated. Examples of page attributes are bits that indicate whether the page of the translation can be cached or whether the page has read/write or read only privileges.
TSB data field 550 also includes a page size field 580. In this embodiment, the two mappings being used are chosen from a universe of four possible page sizes. Therefore, in this embodiment, page size field 580 is two bits in size and can be used to specify the page size of the four possible page sizes corresponding to the page of translation table entry 500.
Finally, TSB data field 550 has a valid bit 590 that is the most significant bit of TSB data field 550. If valid bit field 590 is set, then the translation stored in TTE 500 is valid and can be used. If, on the other hand the valid bit 590 is not set, then the translation table entry 500 is not valid and cannot be used to provide a translation. As will be discussed below in connection with FIG. 9, by placing the valid bit in the most significant bit of TSB data field 550, this embodiment exploits the fact that the sign of an integer number is specified in the most significant bit of a representation of the number. As will be seen, a branch on negative instruction can be used to efficiently branch when a valid translation has been found in the translation storage buffer.
Referring now to FIG. 6, hardware that is used in one embodiment to provide the TSB tag target and TSB pointers is illustrated. In FIG. 6, tag access register 610 is a register that contains the context number of the context requesting a Virtual address to be translated and the most significant bits of the virtual address to be translated that are required by the translation storage buffer lookup. Thus, in the embodiment where the smallest page size is an 8KB page, and the virtual address space is specified by 64 bits, the tag access register 610 is a 64 bit register that contains the 51 most significant bits of the virtual address to be translated in the 51 most significant bits of the tag access register 610. In this embodiment, because a context number can be represented in 13 bits or less, the 13 least significant bits of tag access register 610 are used to store the context number that requested the virtual address translation.
In the case where the translation storage buffer lookup is being performed after failure of a translation lookaside buffer lookup, tag access register 610 would be filled by the translation lookaside buffer when it requested a translation storage buffer lookup be performed. If, on the other hand, the translation storage buffer were being used as a Level I translator, then tag access register 610 would be filled by the CPU requesting the address translation at the time that it requested the address translation. TSB tag target logic 620 is bit shifting logic that fills TSB tag target register 630 with a TSB tag target to be compared with the TSB tag of each translation table entry of the translation storage buffer to be checked.
In a similar manner, TSB pointer logic 640 takes the value stored in the tag access register 610 and calculates one TSB pointer per page size. The TSB pointers are then stored in TSB pointers register 650. In order to perform the calculation of the TSB pointers, TSB pointer logic 640 is supplied with TSB parameters 660. The values stored in TSB parameters 660 are set at the time that the operating system is initialized. In an alternate embodiment, the parameters are dynamic and are changed as changes are needed. For example, the values of the TSB parameters can be changed during a context switch. By storing the information for the TSB parameters 660 in registers, a high degree of flexibility can be maintained in structuring the TSB. This flexibility permits the TSB to be tuned so that it will operate most effectively for a particular computer system implementation.
In the embodiment described in FIG. 6, the TSB parameters 660 that can be set at initialization include TSB split 662, TSB bases 664, number of page sizes 666, page sizes 668 and TSB size 670. The parameter TSB split 662 is a one bit parameter. When TSB split is set to zero, a translation storage buffer structure such as that of translation storage buffer 400 of FIG. 4 is established. In that structure, big and small page translation table entries are stored intermixed in the same translation storage buffer and the big and small pointers point to entries of the same translation table entry array.
Referring now to FIG. 7, an alternate translation storage buffer 700 is depicted. This structure is implemented when the parameter TSB split 662 of FIG. 6 is set to equal one. The translation storage buffer 700 of FIG. 7 has been split or bifurcated into two separate sub translation storage buffers. There is a small page sub translation storage buffer 710 and a big page sub translation storage buffer 750. Small page sub translation storage buffer 710 has a base value small base 740 and a pointer small pointer 745 into the array of translation table entries for small pages. Similarly, big page sub translation storage buffer 750 has a base value of big base 780 and a pointer of big pointer 785 into the big page translation table entries.
An advantage of the structure of translation storage buffer 700 over the structure of translation storage buffer 400 is that when a hit is found in translation storage buffer 400, it is necessary to check the page size of the candidate before one can be sure that an actual hit was achieved. Translation storage buffer 700, on the other hand, has the page size included implicitly in the structure, therefore there is no need to check the page size after a candidate tag has been found to match the tag target. This is because small page translation table 710 is comprised only of translation table entries for small pages and big page translation storage buffer 750 is comprised only of translation tattle entries for big pages.
On the other hand, in so, he implementations the structure of translation storage buffer 400 can be more efficient that the structure of translation storage buffer 700. This is because, in some implementations, the need to translate virtual address may be dominated by either the small or big page size translations. In such a case, either small page sub TSB 710 or big page sub TSB 750 will be under-utilized while the other sub TSB is over-utilized. In such a case, one of the sub TSBs will be thrashing while the other will be accessed only infrequently.
Referring again to FIG. 6, by including the parameter TSB split 662, one can easily change the structure of the translation storage buffer to suit the needs of a particular system. If the number of page sizes parameter 666 is set to two, when TSB split parameter 662 is set to one, the split will occur into two sub TSBs. If, on the other hand, the number of page sizes parameter 666 is a number other than two, setting the TSB split parameter 662 will cause the TSB structure to be split into a corresponding number of sub TSBs.
Referring now to parameter TSB bases 664, if the TSB split has not been specified, only one TSB base will be stored in TSB bases 664. If, on the other hand, the TSB split has been specified, the bases for each of the sub TSBs will be stored as parameters in TSB bases 664.
The field page sizes 668 will store the actual sizes of the pages being translated. Thus, in the case where there are big and small pages, there will be two entries in page sizes 668, one entry will include the size of a big page and the other the size of a small page. In an alternate embodiment, the sizes could be implicit in hardware and therefore need not be stored in registers.
Finally, parameter TSB size 670 includes the number of entries in a particular translation storage buffer. In the case where translation storage buffer split parameter 662 has been set, TSB size will specify the number of entries in a sub TSB. In an alternate embodiment, there will be a separate TSB size for each sub TSB.
When provided with all of the information stored in TSB parameters register 660 and the tag access register 610, TSB pointer logic 640 can quickly calculate the TSB pointers to be used in a TSB lookup and store this information in TSB pointers register 650.
Referring now to FIG. 8, a flow chart that depicts an embodiment of a TSB lookup is illustrated. The TSB lookup process 800 begins at start bubble 810 and proceeds to process block 820 where a translation lookaside buffer miss is detected thereby causing the loading of the tag access register. In process block 830, the translation storage buffer pointers are calculated using the contents of the tag access register and the parameters that describe the TSB. The TSB tag target is also calculated in process block 830. In one embodiment, the TSB pointers and tag target are calculated by hardware as illustrated in FIG. 6. Alternately, the TSB pointers and tag target can be calculated by software.
Once the TSB pointers and tag target have been calculated, a process is performed in process block 840 to check to determine whether one of the translation table entries of the translation storage buffer pointed to by the TSB pointers contains the translation for the virtual address to be translated. An embodiment of a procedure for checking for TSB hits will be discussed below in greater detail in connection with FIG. 9.
Referring again to FIG. 8, after the check for TSB hit procedure has been performed in process block 840, a test is made in decision block 850 to determine whether or not a TSB hit occurred. If a hit occurred then decision path 854 is taken to process block 860 where the translation lookaside buffer is loaded with the translation data required to perform the virtual address to physical address translation requested.
Referring again briefly to FIG. 5, in one embodiment, the translation lookaside buffer has been structured so that it has a translation lookaside buffer tag and translation lookaside buffer data field for each translation entry of the TLB. In this embodiment, the translation lookaside buffer data field has the same structure as the translation storage buffer data field 550 of FIG. 5. Because the translation lookaside buffer tag field has information that is known at the time that the request was made for a translation, at the time that a translation lookaside buffer miss is detected, a tag register is filled with the contents of a translation lookaside buffer tag field formatted so that it will comply with the structure of a translation lookaside buffer tag. Then, if a TSB hit occurs, there is only a need to transfer the translation lookaside data field contents from the translation storage buffer. The translation lookaside buffer tag for the translation entry can be provided directly from the tag register of the translation lookaside buffer. Furthermore, because the translation storage buffer data field is formatted to be the same as that of the translation lookaside buffer data field, the translation information can be loaded into the translation lookaside buffer quickly since there is no need to reformat the field contents.
Once the translation lookaside buffer has been loaded with translation data in process block 860, then processing of the TSB lookup procedure 800 ends at terminal bubble 880.
Referring again to decision block 850, in the case where a TSB hit did not occur, then processing continues along decision path 852 to process block 870. In process block 870, a miss has occurred because the TSB does not contain the translation for the virtual address to be translated. Therefore, a TSB miss handling procedure is executed.
Referring briefly to FIG. 2, upon a TSB miss, the miss handling procedure will consult the big software translation table 280 managed by the virtual memory operating system to retrieve the address translation if it exists. Alternately, the miss handler could use a translator of an intermediate level between Level II and Level III before the Level III translator is consulted. An example of an intermediate level translator would be yet another translation storage buffer.
Referring now to FIG. 9, a flow chart is illustrated of a procedure that checks to determine whether a TSB hit has occurred. In FIG. 9, the procedure check for TSB hit 900 begins at start bubble 910 and proceeds to process block 920 where the translation storage buffer pointers and tag target are retrieved from their respective registers. Then, in process block 930, the translation storage buffer candidate tags stored in the translation storage buffer at the locations pointed to by the translation storage buffer pointers are retrieved. This is followed by the retrieval of the corresponding translation storage buffer candidate data in process block 940.
A candidate tag for each page size is tested against the target tag in process block 950. In decision block 960, a test is made to determine whether a match was found. If no match was found, then decision path 964 is taken to process block 970 where a miss is declared. Processing then completes at terminal bubble 995 with a return from the procedure.
If, on the other hand, a tag, match was detected in decision block 960, then decision path 962 is taken to decision block 980 where a test is performed to determine whether the matching translation table entry is valid. In the case where the valid bit data field for the successful candidate is the most significant bit, the valid bit can be tested by a branch on negative instruction as discussed previously.
If the matching translation table entry is determined in decision block 980 to be invalid, then decision path 982 is taken to process block 970 and a miss is declared. If, on the other hand, the matching translation table entry is valid, then decision path 984 is taken to process block 990 where a hit is declared before terminating the procedure at terminal bubble 995.
In one embodiment, procedure 900 is implemented in hardware. This permits the checking of the TSB entries that are pointed to by the pointers for each page size to be checked in a parallel manner. Alternately, procedure 900 is performed through the execution of software. A software implementation provides greater flexibility than that afforded by hardware. Furthermore, procedure 900 is structured so that operations within the procedure can be interleaved. Interleaving permits the checking of multiple entries of the translation storage buffer without a linear increase in the amount of time necessary to make the checks. Thus, if only one TSB entry were being tested, idle cycles would be "wasted" while the central processing unit was waiting for operations that have a long latency to complete. Because one TSB translation table entry is checked for each page size, some of the idle cycles caused by a test for a first page size can be used to perform the test for the second page size. Moreover, if procedure 900 is executed on a superscalar processor (i.e. one that can execute multiple instructions concurrently), the translation table entries of the TSB can be checked in parallel even in a software implementation.
While the method and apparatus of the present invention has been described in terms of its presently preferred and alternate embodiments, those skilled in the art will recognize that the present invention may be practiced with modification and alternation within the spirit and scope of the appended claims. The specifications and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims (17)

What is claimed is:
1. A method for translating a first virtual address to a physical address, the method comprising the steps of:
receiving the first virtual address, the first virtual address comprising a virtual page offset and a virtual page number, the first virtual address being mapped by an unknown page size, the unknown page size being one of L different page sizes where L is a positive integer greater than one;
then, for each of the L page sizes, selecting a different one of the L page sizes to be a test page size,
(i) calculating a pointer into a translation storage buffer, the pointer being calculated from the first virtual address by assuming that the first virtual address corresponds to a mapping of the test page size, the pointer pointing to a candidate translation table entry of the translation storage buffer, the candidate translation table entry comprising a candidate tag and candidate data, the candidate tag identifying a particular virtual address and the candidate data identifying a particular physical address corresponding to the particular virtual address,
(ii) extracting a virtual address target tag from the first virtual address, the virtual address target tag being calculated by assuming that the first virtual address corresponds to a mapping of the test page size,
(iii) comparing target tag to the candidate tag, and
(iv) if the target tag matches the candidate tag, providing the candidate data as the physical address translation corresponding to the first virtual address,
wherein the translation storage buffer is comprised of an Nth power of two translation table entries and each translation table entry is M words in size where N and M are positive integers, and the step of calculating the pointer includes the steps of:
forming a translation storage buffer index by extracting N least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the test page size;
adding the product of the translation storage buffer index and M to a base address to form the pointer.
2. The method as set forth in claim 1 wherein the pointer is a second virtual address.
3. The method as set forth in claim 1 wherein:
a smallest page size is selected from among the L page sizes;
the target tag is formed by extracting all but N least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the smallest page size; and
the candidate tag is formed by extracting all but N least significant bits of the virtual page number of the particular virtual address assuming the particular virtual address is mapped using the smallest page size.
4. The method as set forth in claim 1, the method further comprising the step of signaling that a translation failed if the target tag has not matched the candidate tag after testing every page size.
5. A method for translating a first virtual address to a physical address, the method comprising the steps of:
receiving the first virtual address, the first virtual address comprising a virtual page offset and a virtual page number, the first virtual address being mapped by an unknown page size, the unknown page size being one of L different page sizes where L is a positive integer greater than one;
then, for each of the L page sizes, selecting a different one of the L page sizes to be a test page size,
(i) calculating a pointer into a translation storage buffer, the pointer being calculated from the first virtual address by assuming that the first virtual address corresponds to a mapping of the test page size, the pointer pointing to a candidate translation table entry of the translation storage buffer, the candidate translation table entry comprising a candidate tag and candidate data, the candidate tag identifying a particular virtual address and the candidate data identifying a particular physical address corresponding to the particular virtual address,
(ii) extracting a virtual address target tag from the first virtual address, the virtual address target tag being calculated by assuming that the first virtual address corresponds to a mapping of the test page size,
(iii) comparing the target tag to the candidate tag, and
(iv) if the target tag matches the candidate tag, providing the candidate data as the physical address translation corresponding to the first virtual address,
wherein the translation steerage buffer is comprised of sub translation storage buffers, each sub translation storage buffer corresponding to a page size and is comprised of a positive integral power of two translation table entries and each translation table entry is M words in size where M is a positive integer, and the step of calculating the pointer includes the steps of:
forming a translation storage buffer index by extracting N least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the test page size where N is the base two logarithm of the number of translation table entries of the sub translation storage buffer corresponding to the test page size; and
adding the product of the translation storage buffer index and M to a base address to form the pointer, the base address corresponding to the test page size, the base address being one of L base addresses.
6. The method as set forth in claim 5 wherein multiple translation storage buffers are tested.
7. A method for translating a first virtual address to a physical address, the method comprising the steps of:
receiving the first virtual address to be translated, the first virtual address comprising a virtual page offset and a virtual page number, the first virtual address being mapped by an unknown page size, the unknown page size being a first or second page size, the first page size not equal to the second page size;
calculating first and second pointers into a translation storage buffer, the first pointer being calculated from the first virtual address by assuming that the first virtual address corresponds to a mapping of the first page size, the first pointer pointing to a first candidate translation table entry of the translation storage buffer, the first candidate translation table entry comprising a first candidate tag and first candidate data, the first candidate tag identifying a first particular virtual address and the first candidate data identifying a first particular physical address corresponding to the first particular virtual address, the second pointer being calculated from the virtual address to be translated by assuming that the first virtual address corresponds to a mapping of the second page size, the second pointer pointing to a second candidate translation table entry of the translation storage buffer, the second candidate translation table entry comprising a second candidate tag and second candidate data, the second candidate tag identifying a second particular virtual address and the second candidate data identifying a second particular physical address corresponding to the second particular virtual address;
extracting first and second virtual address target tags from the first virtual address, the first virtual address target tag being calculated by assuming that the first virtual address corresponds to a mapping of the first page size, the second virtual address target tag being calculated by assuming that the first virtual address corresponds to a mapping of the second page size;
comparing the first target tag to the first candidate tag;
if the first target tag matches the first candidate tag, providing the first candidate data as the physical address translation corresponding to the first virtual address;
comparing the second target tag to the second candidate tag;
if the second target tag matches the second candidate tag, providing the second candidate data as the physical address translation corresponding to the first virtual address,
wherein the translation storage buffer is comprised of sub translation storage buffers, each sub translation storage buffer corresponding to a page size of the first and second page sizes and is comprised of a positive integral power of two translation table entries and each translation table entry is M words in size where M is a positive integer, the step of calculating the first and second pointers includes the steps of:
forming a first translation storage buffer index by extracting N1 least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the first page size where N1 is the base two logarithm of the number of translation-table entries of the sub translation storage buffer corresponding to the first page size;
adding the product of the first translation storage buffer index and M to a first base address to form the first pointer;
forming a second translation storage buffer index by extracting N2 least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the second page size where N2 is the base two logarithm of the number of translation table entries of the sub translation storage buffer corresponding to the second page size; and
adding the product of the second translation storage buffer index and M to a second base address to form the second pointer.
8. The method as set forth in claim 7 wherein the first base address is equal to the second base address.
9. The method as set forth in claim 7 wherein the first pointer is a second virtual address and the second pointer is a third virtual address.
10. The method as set forth in claim 7 wherein:
the translation storage buffer is comprised of an Nth power of two translation table entries where N is a positive integer;
the first page size is smaller than the second page size;
the first target tag is formed by extracting all but N least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the first page size;
the first candidate tag is formed by extracting all but N least significant bits of the virtual page number of the first particular virtual address assuming the first particular virtual address is mapped using the first page size;
the second target tag is formed by extracting all but N least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the first page size; and
the second candidate tag is formed by extracting all but N least significant bits of the virtual page number of the second particular virtual address assuming the second particular virtual address is mapped using the first page size.
11. The method as set forth in claim 7, the method further comprising the step of signaling that a translation failed if both the first target tag does not match the first candidate tag and the second target tag does not match the second candidate tag.
12. The method as set forth in claim 7 wherein multiple translation storage buffers are tested.
13. An apparatus for translating a first virtual address to a physical address, the apparatus comprising:
a receiver adapted to receiving the first virtual address, the virtual address comprising a virtual page offset and a virtual page number, the first virtual address being mapped by an unknown page size, the unknown page size being a first or second page size, the first page size not equal to the second page size;
a calculator adapted to calculating first and second pointers into a translation storage buffer, the calculator coupled to the receiver, the first pointer being calculated from the first virtual address by assuming that the first virtual address corresponds to a mapping of the first page size, the first pointer pointing to a first candidate translation table entry of the translation storage buffers, the first candidate translation table entry comprising a first candidate tag and first candidate data, the first candidate tag identifying a first particular virtual address and the first candidate data identifying a first particular physical address corresponding to the first particular virtual address, the second pointer being calculated from the first virtual address by assuming that the first virtual address corresponds to a mapping of the second page size, the second pointer pointing to a second candidate translation table entry of the translation storage buffer, the second candidate translation table entry comprising a second candidate tag and second candidate data, the second candidate tag identifying a second particular virtual address and the second candidate data identifying a second particular physical address corresponding to the second particular virtual address;
an extractor adapted to extracting first and second virtual address target tags from the first virtual address, the extractor coupled to the receiver, the first virtual address target tag being calculated by assuming that the first virtual address corresponds to a mapping of the first page size, the second virtual address target tag being calculated by assuming that the first virtual address corresponds to a mapping of the second page size;
a first comparator coupled to the calculator and the extractor, the first comparator adapted to comparing the first target tag to the first candidate tag;
a first provider coupled to the first comparator, the first provider adapted to providing the first candidate data as the physical address translation corresponding to the first virtual address if the first target tag matches the first candidate tag;
a second comparator coupled to the calculator and the extractor, the second comparator adapted to comparing the second target tag to the second candidate tag;
a second provider coupled to the second comparator, the second provider adapted to providing the second candidate data as the physical address translation corresponding to the first virtual address if the second target tag matches the second candidate tag,
wherein the translation storage buffer is comprised of sub translation storage buffers, each sub translation storage buffer corresponds to a page size of the first and second page sizes and is comprised of an integral power of two translation table entries and each translation table entry is M words in size where M is a positive integer, the calculator comprising:
a first indexer adapted to forming a first translation storage buffer index by extracting N1 least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the first page size where N1 is the base two logarithm of the number of translation table entries of the sub translation storage buffer corresponding to the first page size;
a first pointer calculator adapted to adding the product of the first translation storage buffer index and M to a first base address to form the first pointer;
a second indexer adapted to forming a second translation storage buffer index by extracting N2 least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the second page size where N2 is the base two logarithm of the number of translation table entries of the sub translation storage buffer corresponding to the second page size; and
a second pointer calculator adapted to adding the product of the second translation storage buffer index and M to a second base address to form the second pointer.
14. The apparatus as set forth in claim 13 wherein the first pointer is a second virtual address and the second pointer is a third virtual address.
15. The apparatus as set forth in claim 13 wherein the first base address is equal to the second base address.
16. The apparatus as set forth in claim 13 wherein:
the translation storage buffer is comprised of an Nth power of two translation table entries where N is a positive integer;
the first page size is smaller than the second page size;
the first target tag is formed by extracting all but N least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the first page size;
the first candidate tag is formed by extracting all but N least significant bits of the virtual page number of the first particular virtual address assuming the first particular virtual address is mapped using the first page size;
the second target tag is formed by extracting all but N least significant bits of the virtual page number of the first virtual address assuming the first virtual address is mapped using the first page size; and
the second candidate tag is formed by extracting all but N least significant bits of the virtual page number of the second particular virtual address assuming the second particular virtual address is mapped using the first page size.
17. The apparatus as set forth in claim 13, the apparatus further comprising a signaler coupled to the first and second comparator, the signaler adapted to signaling that a translation failed if both the first target tag does not match the first candidate tag and the second target tag does not match the second candidate tag.
US08/118,398 1993-09-08 1993-09-08 Virtual address to physical address translation cache that supports multiple page sizes Expired - Lifetime US5479627A (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US08/118,398 US5479627A (en) 1993-09-08 1993-09-08 Virtual address to physical address translation cache that supports multiple page sizes
EP94305846A EP0642086B1 (en) 1993-09-08 1994-08-05 Virtual address to physical address translation cache that supports multiple page sizes
DE69408922T DE69408922T2 (en) 1993-09-08 1994-08-05 Address translation cache for converting virtual addresses to physical addresses that supports multiple page sizes
JP6236020A JPH07200409A (en) 1993-09-08 1994-09-06 Method and apparatus for conversion of virtual address into physical address
US08/874,201 US5956756A (en) 1993-09-08 1997-06-06 Virtual address to physical address translation of pages with unknown and variable sizes

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US08/118,398 US5479627A (en) 1993-09-08 1993-09-08 Virtual address to physical address translation cache that supports multiple page sizes

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US46585395A Division 1993-09-08 1995-06-06

Publications (1)

Publication Number Publication Date
US5479627A true US5479627A (en) 1995-12-26

Family

ID=22378330

Family Applications (2)

Application Number Title Priority Date Filing Date
US08/118,398 Expired - Lifetime US5479627A (en) 1993-09-08 1993-09-08 Virtual address to physical address translation cache that supports multiple page sizes
US08/874,201 Expired - Lifetime US5956756A (en) 1993-09-08 1997-06-06 Virtual address to physical address translation of pages with unknown and variable sizes

Family Applications After (1)

Application Number Title Priority Date Filing Date
US08/874,201 Expired - Lifetime US5956756A (en) 1993-09-08 1997-06-06 Virtual address to physical address translation of pages with unknown and variable sizes

Country Status (4)

Country Link
US (2) US5479627A (en)
EP (1) EP0642086B1 (en)
JP (1) JPH07200409A (en)
DE (1) DE69408922T2 (en)

Cited By (98)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5623692A (en) * 1995-05-15 1997-04-22 Nvidia Corporation Architecture for providing input/output operations in a computer system
US5640591A (en) * 1995-05-15 1997-06-17 Nvidia Corporation Method and apparatus for naming input/output devices in a computer system
US5652872A (en) * 1994-03-08 1997-07-29 Exponential Technology, Inc. Translator having segment bounds encoding for storage in a TLB
US5680566A (en) * 1995-03-03 1997-10-21 Hal Computer Systems, Inc. Lookaside buffer for inputting multiple address translations in a computer system
US5694567A (en) * 1995-02-09 1997-12-02 Integrated Device Technology, Inc. Direct-mapped cache with cache locking allowing expanded contiguous memory storage by swapping one or more tag bits with one or more index bits
US5732404A (en) * 1996-03-29 1998-03-24 Unisys Corporation Flexible expansion of virtual memory addressing
US5751990A (en) * 1994-04-26 1998-05-12 International Business Machines Corporation Abridged virtual address cache directory
US5784707A (en) * 1994-01-12 1998-07-21 Sun Microsystems, Inc. Method and apparatus for managing virtual computer memory with multiple page sizes
US5809562A (en) * 1996-05-20 1998-09-15 Integrated Device Technology, Inc. Cache array select logic allowing cache array size to differ from physical page size
US5822576A (en) * 1997-03-26 1998-10-13 International Business Machines Corporation Branch history table with branch pattern field
EP0887736A1 (en) * 1997-06-26 1998-12-30 Sun Microsystems, Inc. Flexible translation storage buffers for virtual address translation caching
US5860147A (en) * 1996-09-16 1999-01-12 Intel Corporation Method and apparatus for replacement of entries in a translation look-aside buffer
US5897666A (en) * 1996-12-09 1999-04-27 International Business Machines Corporation Generation of unique address alias for memory disambiguation buffer to avoid false collisions
US5897660A (en) * 1995-04-07 1999-04-27 Intel Corporation Method for managing free physical pages that reduces trashing to improve system performance
US5905509A (en) * 1997-09-30 1999-05-18 Compaq Computer Corp. Accelerated Graphics Port two level Gart cache having distributed first level caches
US5918251A (en) * 1996-12-23 1999-06-29 Intel Corporation Method and apparatus for preloading different default address translation attributes
US5936640A (en) * 1997-09-30 1999-08-10 Compaq Computer Corporation Accelerated graphics port memory mapped status and control registers
US5949436A (en) * 1997-09-30 1999-09-07 Compaq Computer Corporation Accelerated graphics port multiple entry gart cache allocation system and method
US5986677A (en) * 1997-09-30 1999-11-16 Compaq Computer Corporation Accelerated graphics port read transaction merging
US5999743A (en) * 1997-09-09 1999-12-07 Compaq Computer Corporation System and method for dynamically allocating accelerated graphics port memory space
US6012132A (en) * 1997-03-31 2000-01-04 Intel Corporation Method and apparatus for implementing a page table walker that uses a sliding field in the virtual addresses to identify entries in a page table
US6047354A (en) * 1994-09-09 2000-04-04 Hitachi, Ltd. Data processor for implementing virtual pages using a cache and register
US6055617A (en) * 1997-08-29 2000-04-25 Sequent Computer Systems, Inc. Virtual address window for accessing physical memory in a computer system
US6057863A (en) * 1997-10-31 2000-05-02 Compaq Computer Corporation Dual purpose apparatus, method and system for accelerated graphics port and fibre channel arbitrated loop interfaces
US6065010A (en) * 1997-06-10 2000-05-16 Daikin Us Corporation Computer implemented method of generating virtual files for sharing information of physical information file
US6069638A (en) * 1997-06-25 2000-05-30 Micron Electronics, Inc. System for accelerated graphics port address remapping interface to main memory
US6078338A (en) * 1998-03-11 2000-06-20 Compaq Computer Corporation Accelerated graphics port programmable memory access arbiter
US6081881A (en) * 1998-02-20 2000-06-27 Unisys Corporation Method of and apparatus for speeding up the execution of normal extended mode transfer instructions
US6088780A (en) * 1997-03-31 2000-07-11 Institute For The Development Of Emerging Architecture, L.L.C. Page table walker that uses at least one of a default page size and a page size selected for a virtual address space to position a sliding field in a virtual address
US6108761A (en) * 1998-02-20 2000-08-22 Unisys Corporation Method of and apparatus for saving time performing certain transfer instructions
US6128684A (en) * 1997-06-30 2000-10-03 Nec Corporation Bus bridge
US6157398A (en) * 1997-12-30 2000-12-05 Micron Technology, Inc. Method of implementing an accelerated graphics port for a multiple memory controller computer system
US6167476A (en) * 1998-09-24 2000-12-26 Compaq Computer Corporation Apparatus, method and system for accelerated graphics port bus bridges
US6182201B1 (en) * 1997-04-14 2001-01-30 International Business Machines Corporation Demand-based issuance of cache operations to a system bus
US6192457B1 (en) 1997-07-02 2001-02-20 Micron Technology, Inc. Method for implementing a graphic address remapping table as a virtual register file in system memory
US6195734B1 (en) 1997-07-02 2001-02-27 Micron Technology, Inc. System for implementing a graphic address remapping table as a virtual register file in system memory
US6205530B1 (en) * 1997-05-08 2001-03-20 Hyundai Electronics Industries Co., Ltd. Address translation unit supporting variable page sizes
US6223239B1 (en) 1998-08-12 2001-04-24 Compaq Computer Corporation Dual purpose apparatus, method and system for accelerated graphics port or system area network interface
US6230223B1 (en) 1998-06-01 2001-05-08 Compaq Computer Corporation Dual purpose apparatus method and system for accelerated graphics or second memory interface
US6249853B1 (en) 1997-06-25 2001-06-19 Micron Electronics, Inc. GART and PTES defined by configuration registers
US6252612B1 (en) 1997-12-30 2001-06-26 Micron Electronics, Inc. Accelerated graphics port for multiple memory controller computer system
US6282625B1 (en) 1997-06-25 2001-08-28 Micron Electronics, Inc. GART and PTES defined by configuration registers
US6321276B1 (en) * 1998-08-04 2001-11-20 Microsoft Corporation Recoverable methods and systems for processing input/output requests including virtual memory addresses
US6336180B1 (en) * 1997-04-30 2002-01-01 Canon Kabushiki Kaisha Method, apparatus and system for managing virtual memory with virtual-physical mapping
US6457068B1 (en) * 1999-08-30 2002-09-24 Intel Corporation Graphics address relocation table (GART) stored entirely in a local memory of an expansion bridge for address translation
US6549997B2 (en) * 2001-03-16 2003-04-15 Fujitsu Limited Dynamic variable page size translation of addresses
US6594701B1 (en) 1998-08-04 2003-07-15 Microsoft Corporation Credit-based methods and systems for controlling data flow between a sender and a receiver with reduced copying of data
US6625715B1 (en) * 1999-12-30 2003-09-23 Intel Corporation System and method for translation buffer accommodating multiple page sizes
US20040160448A1 (en) * 1997-12-30 2004-08-19 Joseph Jeddeloh Accelerated graphics port for a multiple memory controller computer system
US6807622B1 (en) 2000-08-09 2004-10-19 Advanced Micro Devices, Inc. Processor which overrides default operand size for implicit stack pointer references and near branches
US6807617B2 (en) 2001-04-02 2004-10-19 Advanced Micro Devices, Inc. Processor, method and apparatus with descriptor table storing segment descriptors of varying size
US20040250053A1 (en) * 2000-08-09 2004-12-09 Mcgrath Kevin J. Multiple entry points for system call instructions
US6839813B2 (en) * 2000-08-21 2005-01-04 Texas Instruments Incorporated TLB operations based on shared bit
US6857058B1 (en) * 1999-10-04 2005-02-15 Intel Corporation Apparatus to map pages of disparate sizes and associated methods
US20050231515A1 (en) * 1999-10-04 2005-10-20 Intel Corporation Apparatus to map virtual pages to disparate-sized, non-contiguous real pages
US6973562B1 (en) * 2000-01-14 2005-12-06 Advanced Micro Devices, Inc. Establishing an operating mode in a processor
US20060004983A1 (en) * 2004-06-30 2006-01-05 Tsao Gary Y Method, system, and program for managing memory options for devices
US7058791B1 (en) 2000-08-09 2006-06-06 Advanced Micro Devices, Inc. Establishing a mode indication responsive to two or more indications
US20060129785A1 (en) * 2004-12-15 2006-06-15 International Business Machines (Ibm) Corporation Storage of data blocks of logical volumes in a virtual disk storage subsystem
US20060149919A1 (en) * 2005-01-05 2006-07-06 Arizpe Arturo L Method, system, and program for addressing pages of memory by an I/O device
US20060224857A1 (en) * 2005-03-29 2006-10-05 O'connor Dennis M Locking entries into translation lookaside buffers
US7124286B2 (en) 2000-01-14 2006-10-17 Advanced Micro Devices, Inc. Establishing an operating mode in a processor
US20060235999A1 (en) * 2005-04-15 2006-10-19 Shah Hemal V Doorbell mechanism
US7278008B1 (en) 2004-01-30 2007-10-02 Nvidia Corporation Virtual address translation system with caching of variable-range translation clusters
US7290116B1 (en) 2004-06-30 2007-10-30 Sun Microsystems, Inc. Level 2 cache index hashing to avoid hot spots
US7296139B1 (en) 2004-01-30 2007-11-13 Nvidia Corporation In-memory table structure for virtual address translation system with translation units of variable range size
US20080022040A1 (en) * 2006-07-19 2008-01-24 Seimizu Joukan Cache memory system
US20080028181A1 (en) * 2006-07-31 2008-01-31 Nvidia Corporation Dedicated mechanism for page mapping in a gpu
US7334108B1 (en) 2004-01-30 2008-02-19 Nvidia Corporation Multi-client virtual address translation system with translation units of variable-range size
US7366829B1 (en) 2004-06-30 2008-04-29 Sun Microsystems, Inc. TLB tag parity checking without CAM read
US7418582B1 (en) 2004-05-13 2008-08-26 Sun Microsystems, Inc. Versatile register file design for a multi-threaded processor utilizing different modes and register windows
US20080276067A1 (en) * 2007-05-01 2008-11-06 Via Technologies, Inc. Method and Apparatus for Page Table Pre-Fetching in Zero Frame Display Channel
US20080294867A1 (en) * 2005-10-31 2008-11-27 Fujitsu Limited Arithmetic processor, information procesing apparatus and memory access method in arithmetic processor
US20090049273A1 (en) * 2004-04-05 2009-02-19 Marvell International Ltd. Physically-tagged cache with virtual fill buffers
US7509484B1 (en) 2004-06-30 2009-03-24 Sun Microsystems, Inc. Handling cache misses by selectively flushing the pipeline
US7543132B1 (en) 2004-06-30 2009-06-02 Sun Microsystems, Inc. Optimizing hardware TLB reload performance in a highly-threaded processor with multiple page sizes
US7571284B1 (en) 2004-06-30 2009-08-04 Sun Microsystems, Inc. Out-of-order memory transactions in a fine-grain multithreaded/multi-core processor
US20100030997A1 (en) * 2008-07-29 2010-02-04 International Business Machines Corporation Virtual memory management
WO2012015766A2 (en) * 2010-07-28 2012-02-02 Rambus Inc. Cache memory that supports tagless addressing
US8296518B2 (en) 2007-06-19 2012-10-23 Fujitsu Limited Arithmetic processing apparatus and method
US20130080735A1 (en) * 2011-09-26 2013-03-28 Fujitsu Limited Address translation device, processing device and control method of processing device
US20130227248A1 (en) * 2012-02-27 2013-08-29 Vmware, Inc. System and method for supporting finer-grained copy-on-write page sizes
US20130339654A1 (en) * 2012-06-14 2013-12-19 International Business Machines Corporation Radix Table Translation of Memory
US20140281353A1 (en) * 2013-03-15 2014-09-18 International Business Machines Corporation Hardware-based pre-page walk virtual address transformation
US8924648B1 (en) 2003-07-03 2014-12-30 Intellectual Venture Funding Llc Method and system for caching attribute data for matching attributes with physical addresses
US20150149743A1 (en) * 2013-11-27 2015-05-28 Realtek Semiconductor Corp. Management method of virtual-to-physical address translation system using part of bits of virtual address as index
US9086988B2 (en) 2012-06-14 2015-07-21 International Business Machines Corporation Identification and consolidation of page table entries
US9424155B1 (en) 2016-01-27 2016-08-23 International Business Machines Corporation Use efficiency of platform memory resources through firmware managed I/O translation table paging
US20160314074A1 (en) * 2015-04-24 2016-10-27 Optimum Semiconductor Technologies, Inc. Computer processor that implements pre-translation of virtual addresses
US9495303B2 (en) * 2015-02-03 2016-11-15 Intel Corporation Fine grained address remapping for virtualization
US9740628B2 (en) 2012-06-14 2017-08-22 International Business Machines Corporation Page table entry consolidation
CN108139981A (en) * 2016-08-11 2018-06-08 华为技术有限公司 The access method and processing chip of list item in a kind of page table cache TLB
WO2018196839A1 (en) * 2017-04-27 2018-11-01 华为技术有限公司 Internal memory access method and computer system
US10642728B2 (en) * 2018-02-28 2020-05-05 Micron Technology, Inc. Storage class memory status
US10719451B2 (en) * 2017-01-13 2020-07-21 Optimum Semiconductor Technologies Inc. Variable translation-lookaside buffer (TLB) indexing
US10769076B2 (en) 2018-11-21 2020-09-08 Nvidia Corporation Distributed address translation in a multi-node interconnect fabric
US10970390B2 (en) * 2018-02-15 2021-04-06 Intel Corporation Mechanism to prevent software side channels
EP4191420A1 (en) * 2021-12-01 2023-06-07 Samsung Electronics Co., Ltd. Address translation cache and system including the same

Families Citing this family (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19524925A1 (en) * 1994-07-09 1996-02-01 Gmd Gmbh Address conversion system for memory management unit
US6175906B1 (en) * 1996-12-06 2001-01-16 Advanced Micro Devices, Inc. Mechanism for fast revalidation of virtual tags
US6289431B1 (en) * 1998-01-26 2001-09-11 Intel Corporation Method and apparatus for accessing more than 4 Gigabytes of physical memory with 4-byte table entries
US6625718B1 (en) * 1998-02-05 2003-09-23 Avaya Technology Corp. Pointers that are relative to their own present locations
JP2000057054A (en) * 1998-08-12 2000-02-25 Fujitsu Ltd High speed address translation system
JP2001184870A (en) * 1999-12-27 2001-07-06 Mitsubishi Electric Corp Associative memory and variable length encoder/decoder using the same
US6598050B1 (en) * 2000-02-11 2003-07-22 Integrated Device Technology, Inc. Apparatus and method for limited data sharing in a multi-tasking system
US7133951B1 (en) 2000-02-29 2006-11-07 Bourekas Philip A Alternate set of registers to service critical interrupts and operating system traps
US6802022B1 (en) 2000-04-14 2004-10-05 Stratus Technologies Bermuda Ltd. Maintenance of consistent, redundant mass storage images
US6901481B2 (en) 2000-04-14 2005-05-31 Stratus Technologies Bermuda Ltd. Method and apparatus for storing transactional information in persistent memory
US6742103B2 (en) 2000-08-21 2004-05-25 Texas Instruments Incorporated Processing system with shared translation lookaside buffer
US6742104B2 (en) 2000-08-21 2004-05-25 Texas Instruments Incorporated Master/slave processing system with shared translation lookaside buffer
EP1262875A1 (en) * 2001-05-28 2002-12-04 Texas Instruments Incorporated Master/slave processing system with shared translation lookaside buffer
US20020133742A1 (en) * 2001-01-16 2002-09-19 Hsiu-Ying Hsu DRAM memory page operation method and its structure
US6886171B2 (en) * 2001-02-20 2005-04-26 Stratus Technologies Bermuda Ltd. Caching for I/O virtual address translation and validation using device drivers
US6766413B2 (en) 2001-03-01 2004-07-20 Stratus Technologies Bermuda Ltd. Systems and methods for caching with file-level granularity
US6938144B2 (en) * 2001-03-22 2005-08-30 Matsushita Electric Industrial Co., Ltd. Address conversion unit for memory device
ATE545909T1 (en) * 2001-05-28 2012-03-15 Texas Instruments Inc MULTIPROCESSOR SYSTEM WITH COMMON ADDRESS TRANSLATION BUFFER MEMORY
GB2395307A (en) * 2002-11-15 2004-05-19 Quadrics Ltd Virtual to physical memory mapping in network interfaces
US7900017B2 (en) 2002-12-27 2011-03-01 Intel Corporation Mechanism for remapping post virtual machine memory pages
US7188229B2 (en) * 2004-01-17 2007-03-06 Sun Microsystems, Inc. Method and apparatus for memory management in a multi-processor computer system
US7390551B2 (en) * 2004-07-02 2008-06-24 Caterpillar Inc. System and method for encapsulation and protection of components
US7418572B2 (en) * 2004-08-18 2008-08-26 International Business Machines Corporation Pretranslating input/output buffers in environments with multiple page sizes
US7386669B2 (en) * 2005-03-31 2008-06-10 International Business Machines Corporation System and method of improving task switching and page translation performance utilizing a multilevel translation lookaside buffer
US7957951B2 (en) * 2007-03-16 2011-06-07 Robert Bosch Gmbh Address translation system for use in a simulation environment
US9153211B1 (en) * 2007-12-03 2015-10-06 Nvidia Corporation Method and system for tracking accesses to virtual addresses in graphics contexts
WO2012095901A1 (en) * 2011-01-12 2012-07-19 パナソニック株式会社 Program execution device and compiler system
DE102011052510A1 (en) * 2011-08-09 2013-02-14 Dspace Digital Signal Processing And Control Engineering Gmbh Method for processing data of a control device in a data communication device
WO2014016650A1 (en) * 2012-07-27 2014-01-30 Freescale Semiconductor, Inc. Circuitry for a computing system and computing system
US9612970B2 (en) 2014-07-17 2017-04-04 Qualcomm Incorporated Method and apparatus for flexible cache partitioning by sets and ways into component caches
US10089238B2 (en) 2014-07-17 2018-10-02 Qualcomm Incorporated Method and apparatus for a shared cache with dynamic partitioning
US10180908B2 (en) 2015-05-13 2019-01-15 Qualcomm Incorporated Method and apparatus for virtualized control of a shared system cache
US11082231B2 (en) * 2017-12-29 2021-08-03 Intel Corporation Indirection directories for cryptographic memory protection
US11088846B2 (en) * 2019-03-28 2021-08-10 Intel Corporation Key rotating trees with split counters for efficient hardware replay protection

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3675215A (en) * 1970-06-29 1972-07-04 Ibm Pseudo-random code implemented variable block-size storage mapping device and method
US4285040A (en) * 1977-11-04 1981-08-18 Sperry Corporation Dual mode virtual-to-real address translation mechanism
US4763250A (en) * 1985-04-01 1988-08-09 Motorola, Inc. Paged memory management unit having variable number of translation table levels
US5058003A (en) * 1988-12-15 1991-10-15 International Business Machines Corporation Virtual storage dynamic address translation mechanism for multiple-sized pages
EP0496228A1 (en) * 1991-01-24 1992-07-29 ILKA MASCHINENFABRIK HALLE GmbH Safete device for a refrigerator utilising ammonia as refrigerant
EP0508577A1 (en) * 1991-03-13 1992-10-14 International Business Machines Corporation Address translation mechanism
US5263140A (en) * 1991-01-23 1993-11-16 Silicon Graphics, Inc. Variable page size per entry translation look-aside buffer
US5327372A (en) * 1992-01-17 1994-07-05 Matsushita Electric Industrial Co., Ltd. Semiconductor memory device

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS62237547A (en) * 1986-04-09 1987-10-17 Hitachi Ltd Address conversion system
US4914577A (en) * 1987-07-16 1990-04-03 Icon International, Inc. Dynamic memory management system and method
EP0506236A1 (en) * 1991-03-13 1992-09-30 International Business Machines Corporation Address translation mechanism
US5465337A (en) * 1992-08-13 1995-11-07 Sun Microsystems, Inc. Method and apparatus for a memory management unit supporting multiple page sizes
US5526504A (en) * 1993-12-15 1996-06-11 Silicon Graphics, Inc. Variable page size translation lookaside buffer

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3675215A (en) * 1970-06-29 1972-07-04 Ibm Pseudo-random code implemented variable block-size storage mapping device and method
US4285040A (en) * 1977-11-04 1981-08-18 Sperry Corporation Dual mode virtual-to-real address translation mechanism
US4763250A (en) * 1985-04-01 1988-08-09 Motorola, Inc. Paged memory management unit having variable number of translation table levels
US5058003A (en) * 1988-12-15 1991-10-15 International Business Machines Corporation Virtual storage dynamic address translation mechanism for multiple-sized pages
US5263140A (en) * 1991-01-23 1993-11-16 Silicon Graphics, Inc. Variable page size per entry translation look-aside buffer
EP0496228A1 (en) * 1991-01-24 1992-07-29 ILKA MASCHINENFABRIK HALLE GmbH Safete device for a refrigerator utilising ammonia as refrigerant
EP0508577A1 (en) * 1991-03-13 1992-10-14 International Business Machines Corporation Address translation mechanism
US5327372A (en) * 1992-01-17 1994-07-05 Matsushita Electric Industrial Co., Ltd. Semiconductor memory device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Tallure et al., "Tradeoffs in Supporting Two Page Sizes"; Computer News; May 1992, No. 2; New York; pp. 415-424.
Tallure et al., Tradeoffs in Supporting Two Page Sizes ; Computer News; May 1992, No. 2; New York; pp. 415 424. *

Cited By (152)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6393545B1 (en) 1919-04-30 2002-05-21 Canon Kabushiki Kaisha Method apparatus and system for managing virtual memory with virtual-physical mapping
US5784707A (en) * 1994-01-12 1998-07-21 Sun Microsystems, Inc. Method and apparatus for managing virtual computer memory with multiple page sizes
US5652872A (en) * 1994-03-08 1997-07-29 Exponential Technology, Inc. Translator having segment bounds encoding for storage in a TLB
US5751990A (en) * 1994-04-26 1998-05-12 International Business Machines Corporation Abridged virtual address cache directory
US6047354A (en) * 1994-09-09 2000-04-04 Hitachi, Ltd. Data processor for implementing virtual pages using a cache and register
US5694567A (en) * 1995-02-09 1997-12-02 Integrated Device Technology, Inc. Direct-mapped cache with cache locking allowing expanded contiguous memory storage by swapping one or more tag bits with one or more index bits
US5680566A (en) * 1995-03-03 1997-10-21 Hal Computer Systems, Inc. Lookaside buffer for inputting multiple address translations in a computer system
US5893931A (en) * 1995-03-03 1999-04-13 Fujitsu Limited Lookaside buffer for address translation in a computer system
US5897660A (en) * 1995-04-07 1999-04-27 Intel Corporation Method for managing free physical pages that reduces trashing to improve system performance
US5740464A (en) * 1995-05-15 1998-04-14 Nvidia Corporation Architecture for providing input/output operations in a computer system
US5623692A (en) * 1995-05-15 1997-04-22 Nvidia Corporation Architecture for providing input/output operations in a computer system
US5640591A (en) * 1995-05-15 1997-06-17 Nvidia Corporation Method and apparatus for naming input/output devices in a computer system
US5732404A (en) * 1996-03-29 1998-03-24 Unisys Corporation Flexible expansion of virtual memory addressing
US5809562A (en) * 1996-05-20 1998-09-15 Integrated Device Technology, Inc. Cache array select logic allowing cache array size to differ from physical page size
US5860147A (en) * 1996-09-16 1999-01-12 Intel Corporation Method and apparatus for replacement of entries in a translation look-aside buffer
US5897666A (en) * 1996-12-09 1999-04-27 International Business Machines Corporation Generation of unique address alias for memory disambiguation buffer to avoid false collisions
US5918251A (en) * 1996-12-23 1999-06-29 Intel Corporation Method and apparatus for preloading different default address translation attributes
US5822576A (en) * 1997-03-26 1998-10-13 International Business Machines Corporation Branch history table with branch pattern field
US6088780A (en) * 1997-03-31 2000-07-11 Institute For The Development Of Emerging Architecture, L.L.C. Page table walker that uses at least one of a default page size and a page size selected for a virtual address space to position a sliding field in a virtual address
US6012132A (en) * 1997-03-31 2000-01-04 Intel Corporation Method and apparatus for implementing a page table walker that uses a sliding field in the virtual addresses to identify entries in a page table
US6182201B1 (en) * 1997-04-14 2001-01-30 International Business Machines Corporation Demand-based issuance of cache operations to a system bus
US6336180B1 (en) * 1997-04-30 2002-01-01 Canon Kabushiki Kaisha Method, apparatus and system for managing virtual memory with virtual-physical mapping
US6205530B1 (en) * 1997-05-08 2001-03-20 Hyundai Electronics Industries Co., Ltd. Address translation unit supporting variable page sizes
US6065010A (en) * 1997-06-10 2000-05-16 Daikin Us Corporation Computer implemented method of generating virtual files for sharing information of physical information file
US6418523B2 (en) 1997-06-25 2002-07-09 Micron Electronics, Inc. Apparatus comprising a translation lookaside buffer for graphics address remapping of virtual addresses
US6069638A (en) * 1997-06-25 2000-05-30 Micron Electronics, Inc. System for accelerated graphics port address remapping interface to main memory
US6282625B1 (en) 1997-06-25 2001-08-28 Micron Electronics, Inc. GART and PTES defined by configuration registers
US6249853B1 (en) 1997-06-25 2001-06-19 Micron Electronics, Inc. GART and PTES defined by configuration registers
US5899994A (en) * 1997-06-26 1999-05-04 Sun Microsystems, Inc. Flexible translation storage buffers for virtual address translation
EP0887736A1 (en) * 1997-06-26 1998-12-30 Sun Microsystems, Inc. Flexible translation storage buffers for virtual address translation caching
US6128684A (en) * 1997-06-30 2000-10-03 Nec Corporation Bus bridge
US6593932B2 (en) 1997-07-02 2003-07-15 Micron Technology, Inc. System for implementing a graphic address remapping table as a virtual register file in system memory
US6192457B1 (en) 1997-07-02 2001-02-20 Micron Technology, Inc. Method for implementing a graphic address remapping table as a virtual register file in system memory
US6195734B1 (en) 1997-07-02 2001-02-27 Micron Technology, Inc. System for implementing a graphic address remapping table as a virtual register file in system memory
US6055617A (en) * 1997-08-29 2000-04-25 Sequent Computer Systems, Inc. Virtual address window for accessing physical memory in a computer system
US5999743A (en) * 1997-09-09 1999-12-07 Compaq Computer Corporation System and method for dynamically allocating accelerated graphics port memory space
US5986677A (en) * 1997-09-30 1999-11-16 Compaq Computer Corporation Accelerated graphics port read transaction merging
US5949436A (en) * 1997-09-30 1999-09-07 Compaq Computer Corporation Accelerated graphics port multiple entry gart cache allocation system and method
US5936640A (en) * 1997-09-30 1999-08-10 Compaq Computer Corporation Accelerated graphics port memory mapped status and control registers
US5905509A (en) * 1997-09-30 1999-05-18 Compaq Computer Corp. Accelerated Graphics Port two level Gart cache having distributed first level caches
US6057863A (en) * 1997-10-31 2000-05-02 Compaq Computer Corporation Dual purpose apparatus, method and system for accelerated graphics port and fibre channel arbitrated loop interfaces
US6157398A (en) * 1997-12-30 2000-12-05 Micron Technology, Inc. Method of implementing an accelerated graphics port for a multiple memory controller computer system
US20050001847A1 (en) * 1997-12-30 2005-01-06 Joseph Jeddeloh Method of implementing an accelerated graphics/port for a multiple memory controller computer system
US6252612B1 (en) 1997-12-30 2001-06-26 Micron Electronics, Inc. Accelerated graphics port for multiple memory controller computer system
US20040160448A1 (en) * 1997-12-30 2004-08-19 Joseph Jeddeloh Accelerated graphics port for a multiple memory controller computer system
US6741254B1 (en) 1997-12-30 2004-05-25 Micron Technology, Inc. Method of implementing an accelerated graphics port for a multiple memory controller computer system
US6717582B2 (en) 1997-12-30 2004-04-06 Micron Technology, Inc. Accelerated graphics port for a multiple memory controller computer system
US6947050B2 (en) 1997-12-30 2005-09-20 Micron Technology Inc. Method of implementing an accelerated graphics/port for a multiple memory controller computer system
US7777752B2 (en) 1997-12-30 2010-08-17 Round Rock Research, Llc Method of implementing an accelerated graphics port for a multiple memory controller computer system
US7071946B2 (en) 1997-12-30 2006-07-04 Micron Technology, Inc. Accelerated graphics port for a multiple memory controller computer system
US8564602B2 (en) 1997-12-30 2013-10-22 Round Rock Research, Llc Method of implementing an accelerated graphics port for a multiple memory controller computer system
US6081881A (en) * 1998-02-20 2000-06-27 Unisys Corporation Method of and apparatus for speeding up the execution of normal extended mode transfer instructions
US6108761A (en) * 1998-02-20 2000-08-22 Unisys Corporation Method of and apparatus for saving time performing certain transfer instructions
US6078338A (en) * 1998-03-11 2000-06-20 Compaq Computer Corporation Accelerated graphics port programmable memory access arbiter
US6230223B1 (en) 1998-06-01 2001-05-08 Compaq Computer Corporation Dual purpose apparatus method and system for accelerated graphics or second memory interface
US6594701B1 (en) 1998-08-04 2003-07-15 Microsoft Corporation Credit-based methods and systems for controlling data flow between a sender and a receiver with reduced copying of data
US6321276B1 (en) * 1998-08-04 2001-11-20 Microsoft Corporation Recoverable methods and systems for processing input/output requests including virtual memory addresses
US6760787B2 (en) 1998-08-04 2004-07-06 Miscrosoft Corporation Recoverable methods and systems for processing input/output requests including virtual memory addresses
US6223239B1 (en) 1998-08-12 2001-04-24 Compaq Computer Corporation Dual purpose apparatus, method and system for accelerated graphics port or system area network interface
US6167476A (en) * 1998-09-24 2000-12-26 Compaq Computer Corporation Apparatus, method and system for accelerated graphics port bus bridges
US20040068602A1 (en) * 1998-09-24 2004-04-08 Olarig Sompong Paul Apparatus, method and system for accelerated graphics port bus bridges
US6457068B1 (en) * 1999-08-30 2002-09-24 Intel Corporation Graphics address relocation table (GART) stored entirely in a local memory of an expansion bridge for address translation
US6618770B2 (en) 1999-08-30 2003-09-09 Intel Corporation Graphics address relocation table (GART) stored entirely in a local memory of an input/output expansion bridge for input/output (I/O) address translation
US7117339B2 (en) 1999-10-04 2006-10-03 Intel Corporation Apparatus to map virtual pages to disparate-sized, non-contiguous real pages
US6857058B1 (en) * 1999-10-04 2005-02-15 Intel Corporation Apparatus to map pages of disparate sizes and associated methods
US20050231515A1 (en) * 1999-10-04 2005-10-20 Intel Corporation Apparatus to map virtual pages to disparate-sized, non-contiguous real pages
US6970992B2 (en) * 1999-10-04 2005-11-29 Intel Corporation Apparatus to map virtual pages to disparate-sized, non-contiguous real pages and methods relating thereto
US20030196066A1 (en) * 1999-12-30 2003-10-16 Intel Corporation System and method for translation buffer accommodating multiple page sizes
US6625715B1 (en) * 1999-12-30 2003-09-23 Intel Corporation System and method for translation buffer accommodating multiple page sizes
US7124286B2 (en) 2000-01-14 2006-10-17 Advanced Micro Devices, Inc. Establishing an operating mode in a processor
US6973562B1 (en) * 2000-01-14 2005-12-06 Advanced Micro Devices, Inc. Establishing an operating mode in a processor
US7100028B2 (en) 2000-08-09 2006-08-29 Advanced Micro Devices, Inc. Multiple entry points for system call instructions
US7058791B1 (en) 2000-08-09 2006-06-06 Advanced Micro Devices, Inc. Establishing a mode indication responsive to two or more indications
US6807622B1 (en) 2000-08-09 2004-10-19 Advanced Micro Devices, Inc. Processor which overrides default operand size for implicit stack pointer references and near branches
US20050033940A1 (en) * 2000-08-09 2005-02-10 Advanced Micro Devices, Inc. Processor which overrides default operand size for implicit stack pointer references and near branches
US20040250053A1 (en) * 2000-08-09 2004-12-09 Mcgrath Kevin J. Multiple entry points for system call instructions
US7284115B2 (en) 2000-08-09 2007-10-16 Advanced Micro Devices, Inc. Processor which overrides default operand size for implicit stack pointer references and near branches
US6839813B2 (en) * 2000-08-21 2005-01-04 Texas Instruments Incorporated TLB operations based on shared bit
US6549997B2 (en) * 2001-03-16 2003-04-15 Fujitsu Limited Dynamic variable page size translation of addresses
US6807617B2 (en) 2001-04-02 2004-10-19 Advanced Micro Devices, Inc. Processor, method and apparatus with descriptor table storing segment descriptors of varying size
US8924648B1 (en) 2003-07-03 2014-12-30 Intellectual Venture Funding Llc Method and system for caching attribute data for matching attributes with physical addresses
US7386697B1 (en) 2004-01-30 2008-06-10 Nvidia Corporation Memory management for virtual address space with translation units of variable range size
US7278008B1 (en) 2004-01-30 2007-10-02 Nvidia Corporation Virtual address translation system with caching of variable-range translation clusters
US7797510B1 (en) 2004-01-30 2010-09-14 Nvidia Corporation Memory management for virtual address space with translation units of variable range size
US7296139B1 (en) 2004-01-30 2007-11-13 Nvidia Corporation In-memory table structure for virtual address translation system with translation units of variable range size
US7334108B1 (en) 2004-01-30 2008-02-19 Nvidia Corporation Multi-client virtual address translation system with translation units of variable-range size
US20090049273A1 (en) * 2004-04-05 2009-02-19 Marvell International Ltd. Physically-tagged cache with virtual fill buffers
US7761661B2 (en) * 2004-04-05 2010-07-20 Marvell International Ltd. Physically-tagged cache with virtual fill buffers
US7418582B1 (en) 2004-05-13 2008-08-26 Sun Microsystems, Inc. Versatile register file design for a multi-threaded processor utilizing different modes and register windows
US7509484B1 (en) 2004-06-30 2009-03-24 Sun Microsystems, Inc. Handling cache misses by selectively flushing the pipeline
US7366829B1 (en) 2004-06-30 2008-04-29 Sun Microsystems, Inc. TLB tag parity checking without CAM read
US20060004983A1 (en) * 2004-06-30 2006-01-05 Tsao Gary Y Method, system, and program for managing memory options for devices
US7290116B1 (en) 2004-06-30 2007-10-30 Sun Microsystems, Inc. Level 2 cache index hashing to avoid hot spots
US7543132B1 (en) 2004-06-30 2009-06-02 Sun Microsystems, Inc. Optimizing hardware TLB reload performance in a highly-threaded processor with multiple page sizes
US7571284B1 (en) 2004-06-30 2009-08-04 Sun Microsystems, Inc. Out-of-order memory transactions in a fine-grain multithreaded/multi-core processor
US20060129785A1 (en) * 2004-12-15 2006-06-15 International Business Machines (Ibm) Corporation Storage of data blocks of logical volumes in a virtual disk storage subsystem
US7685400B2 (en) * 2004-12-15 2010-03-23 International Business Machines Corporation Storage of data blocks of logical volumes in a virtual disk storage subsystem
US7370174B2 (en) * 2005-01-05 2008-05-06 Intel Corporation Method, system, and program for addressing pages of memory by an I/O device
US20060149919A1 (en) * 2005-01-05 2006-07-06 Arizpe Arturo L Method, system, and program for addressing pages of memory by an I/O device
US20060224857A1 (en) * 2005-03-29 2006-10-05 O'connor Dennis M Locking entries into translation lookaside buffers
US9842058B2 (en) 2005-03-29 2017-12-12 Micron Technology, Inc. Locking entries into translation lookaside buffers
US7853957B2 (en) 2005-04-15 2010-12-14 Intel Corporation Doorbell mechanism using protection domains
US20060235999A1 (en) * 2005-04-15 2006-10-19 Shah Hemal V Doorbell mechanism
US20080294867A1 (en) * 2005-10-31 2008-11-27 Fujitsu Limited Arithmetic processor, information procesing apparatus and memory access method in arithmetic processor
US7797494B2 (en) 2005-10-31 2010-09-14 Fujitsu Limited Arithmetic processor, information processing apparatus and memory access method in arithmetic processor
US7870325B2 (en) 2006-07-19 2011-01-11 Panasonic Corporation Cache memory system
US20080022040A1 (en) * 2006-07-19 2008-01-24 Seimizu Joukan Cache memory system
US20080028181A1 (en) * 2006-07-31 2008-01-31 Nvidia Corporation Dedicated mechanism for page mapping in a gpu
US20080276067A1 (en) * 2007-05-01 2008-11-06 Via Technologies, Inc. Method and Apparatus for Page Table Pre-Fetching in Zero Frame Display Channel
US8296518B2 (en) 2007-06-19 2012-10-23 Fujitsu Limited Arithmetic processing apparatus and method
US7930515B2 (en) 2008-07-29 2011-04-19 International Business Machines Corporation Virtual memory management
US20100030997A1 (en) * 2008-07-29 2010-02-04 International Business Machines Corporation Virtual memory management
WO2012015766A3 (en) * 2010-07-28 2012-05-03 Rambus Inc. Cache memory that supports tagless addressing
WO2012015766A2 (en) * 2010-07-28 2012-02-02 Rambus Inc. Cache memory that supports tagless addressing
US11537531B2 (en) 2010-07-28 2022-12-27 Rambus Inc. Cache memory that supports tagless addressing
US10891241B2 (en) 2010-07-28 2021-01-12 Rambus Inc. Cache memory that supports tagless addressing
US20130080735A1 (en) * 2011-09-26 2013-03-28 Fujitsu Limited Address translation device, processing device and control method of processing device
US8914611B2 (en) * 2011-09-26 2014-12-16 Fujitsu Limited Address translation device, processing device and control method of processing device
US9152570B2 (en) * 2012-02-27 2015-10-06 Vmware, Inc. System and method for supporting finer-grained copy-on-write page sizes
US20130227248A1 (en) * 2012-02-27 2013-08-29 Vmware, Inc. System and method for supporting finer-grained copy-on-write page sizes
US9740628B2 (en) 2012-06-14 2017-08-22 International Business Machines Corporation Page table entry consolidation
US9753860B2 (en) 2012-06-14 2017-09-05 International Business Machines Corporation Page table entry consolidation
US9092359B2 (en) 2012-06-14 2015-07-28 International Business Machines Corporation Identification and consolidation of page table entries
US9086988B2 (en) 2012-06-14 2015-07-21 International Business Machines Corporation Identification and consolidation of page table entries
US20130339652A1 (en) * 2012-06-14 2013-12-19 International Business Machines Corporation Radix Table Translation of Memory
US9811472B2 (en) * 2012-06-14 2017-11-07 International Business Machines Corporation Radix table translation of memory
US9785569B2 (en) * 2012-06-14 2017-10-10 International Business Machines Corporation Radix table translation of memory
US20130339654A1 (en) * 2012-06-14 2013-12-19 International Business Machines Corporation Radix Table Translation of Memory
US20140281209A1 (en) * 2013-03-15 2014-09-18 International Business Machines Corporation Hardware-based pre-page walk virtual address transformation
US20140281353A1 (en) * 2013-03-15 2014-09-18 International Business Machines Corporation Hardware-based pre-page walk virtual address transformation
US10956340B2 (en) 2013-03-15 2021-03-23 International Business Machines Corporation Hardware-based pre-page walk virtual address transformation independent of page size utilizing bit shifting based on page size
US10216642B2 (en) * 2013-03-15 2019-02-26 International Business Machines Corporation Hardware-based pre-page walk virtual address transformation where the virtual address is shifted by current page size and a minimum page size
US9824023B2 (en) * 2013-11-27 2017-11-21 Realtek Semiconductor Corp. Management method of virtual-to-physical address translation system using part of bits of virtual address as index
US20150149743A1 (en) * 2013-11-27 2015-05-28 Realtek Semiconductor Corp. Management method of virtual-to-physical address translation system using part of bits of virtual address as index
US9495303B2 (en) * 2015-02-03 2016-11-15 Intel Corporation Fine grained address remapping for virtualization
US20160314074A1 (en) * 2015-04-24 2016-10-27 Optimum Semiconductor Technologies, Inc. Computer processor that implements pre-translation of virtual addresses
US10169039B2 (en) * 2015-04-24 2019-01-01 Optimum Semiconductor Technologies, Inc. Computer processor that implements pre-translation of virtual addresses
US10514915B2 (en) 2015-04-24 2019-12-24 Optimum Semiconductor Technologies Inc. Computer processor with address register file
US9424155B1 (en) 2016-01-27 2016-08-23 International Business Machines Corporation Use efficiency of platform memory resources through firmware managed I/O translation table paging
US10310759B2 (en) 2016-01-27 2019-06-04 International Business Machines Corporation Use efficiency of platform memory resources through firmware managed I/O translation table paging
CN108139981A (en) * 2016-08-11 2018-06-08 华为技术有限公司 The access method and processing chip of list item in a kind of page table cache TLB
US10740247B2 (en) * 2016-08-11 2020-08-11 Huawei Technologies Co., Ltd. Method for accessing entry in translation lookaside buffer TLB and processing chip
US10719451B2 (en) * 2017-01-13 2020-07-21 Optimum Semiconductor Technologies Inc. Variable translation-lookaside buffer (TLB) indexing
CN108804350B (en) * 2017-04-27 2020-02-21 华为技术有限公司 Memory access method and computer system
CN108804350A (en) * 2017-04-27 2018-11-13 华为技术有限公司 A kind of memory pool access method and computer system
WO2018196839A1 (en) * 2017-04-27 2018-11-01 华为技术有限公司 Internal memory access method and computer system
US10970390B2 (en) * 2018-02-15 2021-04-06 Intel Corporation Mechanism to prevent software side channels
US10642728B2 (en) * 2018-02-28 2020-05-05 Micron Technology, Inc. Storage class memory status
US10891223B2 (en) 2018-02-28 2021-01-12 Micron Technology, Inc. Storage class memory status
US10769076B2 (en) 2018-11-21 2020-09-08 Nvidia Corporation Distributed address translation in a multi-node interconnect fabric
US11327900B2 (en) 2018-11-21 2022-05-10 Nvidia Corporation Securing memory accesses in a virtualized environment
EP4191420A1 (en) * 2021-12-01 2023-06-07 Samsung Electronics Co., Ltd. Address translation cache and system including the same

Also Published As

Publication number Publication date
EP0642086A1 (en) 1995-03-08
DE69408922D1 (en) 1998-04-16
JPH07200409A (en) 1995-08-04
EP0642086B1 (en) 1998-03-11
US5956756A (en) 1999-09-21
DE69408922T2 (en) 1998-11-26

Similar Documents

Publication Publication Date Title
US5479627A (en) Virtual address to physical address translation cache that supports multiple page sizes
JP3577575B2 (en) Apparatus and method for performing address translation
US5426750A (en) Translation lookaside buffer apparatus and method with input/output entries, page table entries and page table pointers
US8296547B2 (en) Loading entries into a TLB in hardware via indirect TLB entries
US6014732A (en) Cache memory with reduced access time
US5918250A (en) Method and apparatus for preloading default address translation attributes
EP1096385B1 (en) A method and apparatus for forming an entry address
US5493660A (en) Software assisted hardware TLB miss handler
EP0797149A2 (en) Architecture and method for sharing tlb entries
KR20080041707A (en) Tlb lock indicator
JP2005538465A (en) Conversion index buffer with partial tags hashed based on page size index
US11403222B2 (en) Cache structure using a logical directory
US11775445B2 (en) Translation support for a virtual cache
US10606762B2 (en) Sharing virtual and real translations in a virtual cache
US6298411B1 (en) Method and apparatus to share instruction images in a virtual cache
TWI590053B (en) Selective prefetching of physically sequential cache line to cache line that includes loaded page table
US7562204B1 (en) Identifying and relocating relocatable kernel memory allocations in kernel non-relocatable memory
US6990551B2 (en) System and method for employing a process identifier to minimize aliasing in a linear-addressed cache
JP2008512758A (en) Virtual address cache and method for sharing data stored in virtual address cache
US7519791B2 (en) Address conversion technique in a context switching environment
JP2008511882A (en) Virtual address cache and method for sharing data using unique task identifiers
US6674441B1 (en) Method and apparatus for improving performance of an accelerated graphics port (AGP) device
EP0206050A2 (en) Virtually addressed cache memory with physical tags
US7181587B1 (en) Mapping an arbitrary number of contiguous memory pages at an arbitrary alignment

Legal Events

Date Code Title Description
AS Assignment

Owner name: SUN MICROSYSTEMS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KHALIDI, YOUSEF A.;ANDERSON, GLEN R.;CHESSIN, STEPHEN A.;AND OTHERS;REEL/FRAME:006693/0467;SIGNING DATES FROM 19930824 TO 19930831

AS Assignment

Owner name: SUN MICROSYSTEMS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KHALIDI, YOUSEF A.;ANDERSON, GLEN R.;CHESSIN, STEPHEN A.;AND OTHERS;REEL/FRAME:006757/0580;SIGNING DATES FROM 19930824 TO 19930831

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 8

REMI Maintenance fee reminder mailed
FPAY Fee payment

Year of fee payment: 12