USRE43798E1 - Configurable cache allowing cache-type and buffer-type access - Google Patents

Configurable cache allowing cache-type and buffer-type access Download PDF

Info

Publication number
USRE43798E1
USRE43798E1 US11/606,321 US60632106A USRE43798E US RE43798 E1 USRE43798 E1 US RE43798E1 US 60632106 A US60632106 A US 60632106A US RE43798 E USRE43798 E US RE43798E
Authority
US
United States
Prior art keywords
cache
address
access
memory
accessed
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US11/606,321
Inventor
Craig C. Hansen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microunity Systems Engineering Inc
Original Assignee
Microunity Systems Engineering Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=24159506&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=USRE43798(E1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Microunity Systems Engineering Inc filed Critical Microunity Systems Engineering Inc
Priority to US11/606,321 priority Critical patent/USRE43798E1/en
Application granted granted Critical
Publication of USRE43798E1 publication Critical patent/USRE43798E1/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0223User address space allocation, e.g. contiguous or non contiguous base addressing
    • G06F12/0284Multiple user address space allocation, e.g. using different base addresses
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/10Address translation
    • G06F12/1027Address translation using associative or pseudo-associative address translation means, e.g. translation look-aside buffer [TLB]
    • G06F12/1045Address translation using associative or pseudo-associative address translation means, e.g. translation look-aside buffer [TLB] associated with a data cache
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/14Protection against unauthorised use of memory or access to memory
    • G06F12/1458Protection against unauthorised use of memory or access to memory by checking the subject access rights
    • G06F12/1491Protection against unauthorised use of memory or access to memory by checking the subject access rights in a hierarchical protection system, e.g. privilege levels, memory rings

Definitions

  • the present invention relates to computer memory systems and particularly to virtual memory systems.
  • Caches may use either a virtual or physical address to index the cache, known as a virtual-index cache or a physical-index cache. Additionally, caches may use either a virtual or physical address stored and compared against in the cache tag storage, known as a virtual-tag cache or a physical-tag cache. Virtual-index and virtual-tag caches are generally able to attain higher peak performance, but add constraint to the mapping of addresses available when sharing data or changing the address mapping. In particular, the problem called aliasing occurs, in which two tasks use different virtual addresses to reference the same physical memory. Aliasing may require that tasks sharing memory space use identical or similar virtual addresses.
  • the present invention is a virtual memory system that performs virtual address-to-physical address translations in a manner that increases the overall efficiency and flexibility of the virtual memory system.
  • a virtual memory system that functions to translate a task specific virtual address (referred to as a local virtual address) into a virtual address that is generalized to all or a group of tasks (referred to as a global virtual address) and then translates the global virtual address into an address which points to a block of physical memory, (referred to as the physical address) is described.
  • a first embodiment of the virtual memory system of the present invention includes a local-to-global virtual address translator for translating the local virtual address into the global virtual address and a global virtual-to-physical address translator for translating the global virtual address into the physical address. In an alternate embodiment, separate local-to-global virtual address translators are used for translating each of the data and instruction access addresses.
  • the local-to-global virtual address translator and the global virtual-to-physical address translator each include a plurality of cells, each cell implementing a single entry in a translation lookaside buffer (TLB) which defines a particular address space mapping.
  • TLB entry includes a match field, a mask field, an XOR field, and a protection field.
  • Each cell includes a first logic means for matching the input address to be translated with the contents of the cell's match field to generate a match indicator output signal, a second logic means for masking the match indicator output signal with the contents of the cell's mask field to generate a masked output signal, a third logic means for generating a select signal if all of the signals mating up the masked output signal are at the same logic level, a fourth logic means for outputting the cell's XOR value if the cell is selected, and a fifth logic means for providing a protection signal when the cell is selected.
  • the local-to-global virtual address translator comprises a single cell. In this case, all of the local virtual addresses are translated by the one cell and thus this embodiment of the local-to-global virtual address translator does not require the first and second multiplexers.
  • the local-to-global virtual address translator, the global virtual-to-physical (GV/P) address translator and a cache tag storage provide protection information for each instruction or data access.
  • the protection information includes the task privilege for the given access.
  • the address's protection information is passed to the processor and is checked with the protection information associated with the corresponding local virtual, global virtual, or physical memory area that the address is attempting to access.
  • the virtual cache structure also employs a means for determining whether a local virtual address cache hit or miss has occurred. If a cache hit occurs, the data in the virtual cache is accessed and the access operation is complete. If a cache miss occurs, the global virtual address is passed to the global virtual-to-physical address translator where it is translated into a physical address. The physical address is then used to access the data from either main memory or secondary memory.
  • FIGS. 1 shows one embodiment of the block diagram of the virtual memory system of the present invention.
  • FIG. 2 illustrates an embodiment of the local virtual-to-global translator 110 of the present invention.
  • FIG. 3 illustrates an embodiment of the local TLB of the present invention.
  • FIG. 4 illustrates an embodiment of a protection field utilized to provide protection information in the virtual memory system of the present invention.
  • FIG. 5 illustrates an embodiment of the gateway instruction of the present invention.
  • FIG. 6 illustrates the manner in which gateways are implemented in an embodiment of the present invention.
  • FIG. 7 illustrates another embodiment of the virtual memory system of the present invention having independent data address and instruction address local-to-global virtual translation buffers and independent data and instruction cache systems.
  • FIG. 8 illustrates a configurable cache in accordance with the system and method of the present invention.
  • FIG. 9 shows one embodiment of a logic circuit employed in the system of the present invention to translate cache addresses for accessing the configurable cache of the present invention.
  • a virtual memory system is described.
  • numerous specific details are set forth, such as cache size, address field size and bus widths etc., in order to provide a thorough understanding of the present invention. It will be obvious, however, to one skilled in the art that these specific details need not be employed to practice the present invention. In other instances, well-known memory system structures have not been described in detail in order to avoid unnecessarily obscuring the present invention.
  • FIG. 1 shows a block diagram of one embodiment of the virtual memory system of the present invention.
  • FIG. 1 shows an address represented in three address forms: local virtual address 100 , global virtual address 102 , and physical address 104 .
  • the local virtual address for a given task is a virtual address that is specific to that given task.
  • the global virtual address is also a virtual address.
  • the global virtual address is an address that resides in a virtual memory space that is common to a large group of the tasks. In one embodiment of the present invention, there is only one global virtual memory space. In another embodiment, there are several global virtual memory spaces.
  • the physical address indicates the actual location of the address within the physical memory when it is present in the main memory.
  • FIG. 1 shows a 64-bit local virtual instruction address 100 comprising two portions 100 a and 100 b. Portions 100 a and 100 b correspond to the 16 and 48 bit portions of the address, respectively.
  • FIG. 1 also shows a local-to-global virtual address translator 110 for translating local virtual address 100 .
  • the higher order 16 bits of address 100 i.e. portion 100 a, are coupled to the local-to-global virtual translator.
  • Portion 100 a defines a large area (referred to as a block) of virtual memory space allocated to a specific task.
  • the lower 48 bits of address 100 (portion 100 b) are the offset bits of the local virtual address.
  • Portions 100 b points to a specific area within the block.
  • Local-to-global virtual address translation is performed by mapping the higher order 16 bit portions 100 a of local virtual address 100 to 16 bit portions 102 a of global virtual address 102 .
  • translator 110 outputs the 16 higher order bits ( 102 a) of global virtual address 102 .
  • the lower order 48 bits of address 100 i.e. 100 b
  • are passed directly to the lower order 48 bits of global virtual address 102 designated as portions 102 b and 102 c).
  • Global virtual-to-physical address translation is performed by translator 130 .
  • the higher order 58 bits of global virtual address 102 i.e. portions 102 a and 102 b) are coupled to translator 130 .
  • Translator 130 maps the 58 bits of the selected global virtual address to 58 bits of a physical address.
  • Portion 104 a in physical address 104 corresponds to the translated 58 higher order bits of the global virtual address.
  • the 6 lower bits of address 102 i.e. portion 102 c
  • FIG. 1 also shows a cache storage 111 and its corresponding cache tag 112 .
  • the cache and tags are virtual-indexed and virtual-tagged to facilitate reducing memory access times.
  • the cache storage stores the data corresponding to recently accessed virtual addresses.
  • the cache tag stores global-virtual addresses corresponding to the data stored in the cache storage.
  • the system When an address accesses the virtual memory system shown in FIG. 1 , the system performs three tasks simultaneously.
  • the lower order 48 bits (i.e. the offset) of the local virtual instruction address is coupled to virtual cache 111 to access a block of data in the cache.
  • a 32K byte cache is employed, having 512 blocks with 64 bytes each.
  • the block of data accessed from cache 111 is coupled to the cache's data bus 133 and transmitted to the CPU.
  • portion 100 b is coupled to cache tag 112 to access the index corresponding to the accessed cache data.
  • the accessed index is coupled to a tag compare unit 113 to determine if it matches with the higher order 52 bits of the global virtual instruction address, portion 102 a and 102 b. If it does, a cache hit occurs. This means that the block of data stored in cache 111 is the desired data that corresponds to the original local virtual address.
  • the cache hit indicator is transmitted to the processor on bus 136 and indicates to the processor that the data transmitted on the cache bus is valid. Protection information from cache tag 112 is coupled to the processor on bus 135 . At this point global virtual-to-physical address translation is not necessary since valid data has been transmitted to the CPU.
  • a cache miss is transmitted to the CPU on bus 136 indicating that the data sent on data bus 133 is not valid—at which time the CPU discards the data sent by the cache and the global virtual address is translated into a physical address. If the translation is successful, as indicated by translation hit and protection information 160 , the physical address is then used to access the physical memory and the data return from memory is written into the cache storage 111 and the global-virtual address 102 and protection information 160 is written into cache tag 112 . If the translation is unsuccessful, an exception is indicated to the CPU, and the physical memory access may not occur.
  • the type of addresses stored in the cache tag 112 is a physical address that can be used to access the physical memory.
  • tag compare unit 113 is coupled to physical address 104 instead of global virtual address 102 . Because a virtual-index cache is employed, the portion of the physical address which must be stored in cache tag 112 is larger and therefore a more limited set of protection information is stored in cache tag 112 and transmitted to the processor on bus 135 . This information is supplemented by protection information 160 from the global virtual to physical translator 130 .
  • the memory system employs a single path to perform both data and instruction accesses.
  • both instruction memory accesses and data memory accesses are performed using partially separated paths. These accesses are each separately cached by multiple instances of cache storage, cache tag, and tag compare unit.
  • the instruction cache is virtual-indexed and virtual-tagged
  • the data cache is virtual-indexed and physically-tagged.
  • Local-virtual to global-virtual address translation is accomplished by two instances of the local-virtual to global-virtual translator. These multiple instances may be made to appear as if a single translator is shared by initializing and writing the same data and control information to each instance of the translator.
  • the instruction access path includes a local-to-global virtual address translator 110 , an instruction cache 111 , an instruction cache tag 112 , and cache hit/miss flag 113 and the data access path includes a separate local-to-global virtual address translator 120 , a data cache 121 , a data cache tag 122 , and a cache hit/miss flag 123 .
  • the above-mentioned instruction and data access path elements correspond to and function similarly to the access path elements (i.e. translator 110 , cache 111 , cache tag 112 , and flag 113 ) shown in FIG. 1 .
  • buses 133 and 137 ( FIG. 7 ) correspond to bus 133 ( FIG.
  • buses 134 and 138 correspond to bus 134 ( FIG. 1 ) in that they provide the protection information from the local-to-global virtual address translator
  • buses 135 and 139 correspond to bus 135 ( FIG. 1 ) in that they provide protection information from the cache tag
  • buses 136 and 140 FIG. 7 ) correspond to bus 136 ( FIG. 1 ) in that they provide a signal to the CPU to indicate whether a cache hit or miss has occurred.
  • local-to-global virtual address translation for both of the instruction and data addresses is performed by a single, shared translator.
  • the virtual instruction or data address is referred to as local since it belongs to a virtual address space that is unique to one specific task.
  • each task in the system has a unique local virtual address space.
  • each of the local virtual address spaces are mapped into a common global virtual address space.
  • An alternative embodiment, is that the local virtual address spaces are mapped into more than one global virtual address space.
  • FIG. 2 illustrates the local-to-global virtual address translator 110 of the present invention having four translator cells 1 - 4 .
  • Each translator cell comprises exclusive OR (XOR) logic gate 511 having a first parallel input coupled to the highest order 16 bits of the local virtual address (instruction or data) and a second parallel input coupled to register 402 .
  • Register 402 stores a 16 bit match value.
  • Logic gate 511 is used to perform a bit-by-bit comparison between the higher order 16 bits from the local virtual address and the 16 bit match value. If the match value is the same as the 16 bits from the local virtual address, the XOR gate 511 outputs 16 bits that are all 0's—indicating a match. In the case in which bits differ, gate 511 outputs a “1” for the mis-matched bit locations and “0”'s for the matched bit locations.
  • a non-zero match indicator signal from XOR gate 511 indicates a mis-match.
  • the 16 bit match indicator output signal from XOR gate 511 is coupled to a first parallel input of AND gate 512 .
  • the second parallel input is coupled to 16 bit mask register 401 .
  • the mask register is used to mask out certain of the 16 bits that are non-zero.
  • the two parallel inputs of AND gate 512 are bit-by-bit ANDed together. If all of the 16 bits from XOR gate 511 are zero, then the output of AND gate 16 will also be zero no matter what the mask value is. If any of the bits from XOR gate 511 are non-zero, the mask may be employed to mask these out so that the output of AND gate 512 is 16 bits of all zeros.
  • any of the bits from AND gate 512 are non-zero and they are not masked out by mask 401 then no match occurs and the 16 bit masked output signal of AND gate 512 will contain a “1”.
  • the 16 bits from AND gate 512 are coupled to NOR gate 513 and are NORed together. If all of the 16 bits are zero, NOR gate 513 outputs a “1” indicating that the cell is selected. If any of the 16 bits are non-zero, NOR gate 513 outputs a “0” indicating that the cell is not selected.
  • the select signal is ANDed with each of the values in XOR register 403 .
  • AND gate 514 If the select signal is a 1, indicating that the cell is selected, then AND gate 514 outputs the contents of XOR register 403 . If the select signal is a 0, indicating the cell is not selected, AND gate 514 outputs a 16 bit word of “0”s.
  • FIG. 2 also shows three other cells (cells 2 - 4 )—each cell determining whether a match exists between the 16 bits of local virtual address and match register 402 . Only one of the cells will match the local virtual address at a time. Thus, only one of the cells outputs the 16 bit contents of their corresponding XOR register 403 and the rest output 16 bits of “0”s. Each of the 16 bits of each cell are ORed with the corresponding bits from the other cells (all at a “0” logic level) by OR gate 550 causing gate 550 to output the contents of the XOR value from the selected cell. This value is then XORed with the higher order 16 bits from the original local virtual address being translated by gate 560 to generate the global virtual address.
  • cells 2 - 4 each cell determining whether a match exists between the 16 bits of local virtual address and match register 402 . Only one of the cells will match the local virtual address at a time. Thus, only one of the cells outputs the 16 bit contents of their corresponding XOR register 40
  • the select signal from the output of NOR gate 513 is also coupled to one input of AND gate 515 .
  • the other input of gate 515 is coupled to register 404 which contains the protection information for the global virtual address space defined by cell 1 .
  • AND gate 515 passes the content of the protection register 404 to one input of OR gate 570 . If the select signal is a “0”, then AND gate 515 outputs all “0”s. Since only one cell is selected at a time, gate 570 outputs protection information from only one cell. In other words, OR gate 570 outputs the protection information from the selected cell. This information is then transmitted to the CPU on bus 111 .
  • an alternative space conservative embodiment of the present invention translator 110 comprises a single cell 1 .
  • the match register is all “0”s and the mask register is all “1”s so that they become ineffective.
  • all of the local virtual addresses are translated by the single cell having a mapping defined by the XOR register 403 in that cell.
  • local protection is not required.
  • the local-to-global translator would not need to generate any local protection at this point and the local protection portion of the cell would not be used.
  • Global virtual-to-physical address translator 130 is implement the same as translator 110 except that bus widths are increased to accommodate translation of the higher order 58 bits of the global virtual address. It should be apparent that the match, mask, and XOR values are also increased to 58 bits.
  • translator 110 has multiple cells 1 - 4 each cell having its own unique set of match, mask, XOR and protection registers—each defining a different mapping and consequently a global virtual memory space having its own set of addresses.
  • the match, mask, XOR and protection registers used with translator 110 are stored in a memory buffer referred to as the local TLB.
  • the match, mask, XOR, and protect registers used for translator 130 are stored in a global TLB.
  • FIG. 3 illustrates four entries in a local TLB corresponding to each of the cells in the translator illustrated in FIG. 2 .
  • each entry is a 64 bit word having four fields; the mask field 401 , the match field 402 , the XOR field 403 , and the protect field 404 .
  • a different entry is used for each cell. For example, Mask[ 1 ], Match [ 1 ], XOR[ 1 ], and Protect[ 1 ] are all used by cell 1 ( FIG. 2 ). Similarly, Mask[ 2 ], Match [ 2 ], XOR[ 2 ], and Protect[ 2 ] are all used by cell 2 .
  • the global TLB is similarly structured, having a different field for each of the mask, match, XOR and protect values.
  • the number of entries in the global TLB range from 64 to 256.
  • each of the local and global TLBs generate protection information by passing the contents of the cells protection register 404 ( FIG. 2 ) from the selected cell to OR gate 570 and then to the CPU ( FIGS. 1 and 2 ).
  • a portion or all of the protection information is also concatenated onto the cache tag index for both the data and instruction caches ( FIG. 7 ).
  • the protection information field is stored within each of cache tags 112 and 122 and is concatenated onto the cache tag when these caches are accessed.
  • FIG. 4 illustrates one embodiment of the protection information field of the present invention. As can be seen the 16 bit protection field comprises many sub-fields. The sub-fields define different aspects of a given address's ability to access data and modify it.
  • cc field (cache control) defines the different states (i.e. cached, coherent, noallocate, physical) of the data stored at the address being accessed and the cs field (coherence state) defines what may be done to the data when it is accessed (read, write, or replace).
  • the p field defines the scheduling priority with which the memory access is to be handled, so that high-priority memory accesses may be performed earlier in preference to low-priority memory accesses.
  • the d field defines that a particular exception is to be taken when accesses occur to memory controlled by this field, called a detail exception.
  • the s field defines whether the access is to be strongly or weakly ordered with respect to other memory accesses.
  • the d field when set causes a detail exception to take place.
  • This exception passes control to a software routine which may use a variety of methods to further determine whether an exception should be handled or suppressed for this particular memory reference. For example, it may determine from the virtual address which bytes within a memory block are to be accessed, and consult a bit-map for each byte as to whether the particular type of access is to proceed. In this manner, access may be controlled at the byte level.
  • the software routine sets system state which is preserved only until the target instruction is re-executed upon returning from the exception handler. The system state is then reset so that detail exceptions may occur on the next successive instruction and particularly upon any further re-execution of the same instruction as may occur after a branch operation.
  • the r, w, x, and g fields define the minimum privilege level of the access being performed.
  • the r field indicates the access privilege level for a read access
  • the w field indicates the privilege level for a write access
  • the x field indicates the privilege level for an execute access
  • the g field indicates the privilege level for a gateway access.
  • a higher privilege level can always access a memory space having a lower privilege level, but a lower privilege level cannot access data stored at a higher privilege level unless a gateway access is specified.
  • cache coherency information in addition to access privilege level is provided for at the local virtual, global virtual and physical memory level, cache coherency is maintained throughout the virtual memory system of the present invention at each of these levels.
  • Gateway accesses provide a means of accessing high privileged areas of memory in a very controlled manner. For instance, it may be desired to protect certain types of routines, such as operating system software, in a more robust manner than other software routines stored in memory. To provide this protection, the system software is stored in an area of memory (physical or virtual) having the highest privilege level where accesses only having that privilege level is able to access the system software. Further access control to this area is imposed by ensuring that lower types of accesses are only allowed to access this high privileged area by passing through a gateway. The gateway controls the specific entrance and exit points into the privileged memory area.
  • Gateway accesses are implemented with gateway pointers which are 128 bits of data providing two pointers.
  • the first pointer gives the memory address of where the higher priority area of memory is to be entered.
  • the second pointer is 64 bits of data that can be used for a general purpose, but has the special characteristic that it is loaded as the gateway is entered and that the target procedure may depend upon it being properly loaded into a particular general register of the processor. Because of this characteristic, it is normally loaded with a pointer to a data region to be employed by the target procedure.
  • Gateways are implemented by the gateway instruction of the present invention.
  • FIGS. 5A and 5B illustrate two formats of gateway instructions of the present invention.
  • the B.GATE.I operation provides a secure means to call a procedure, including those at a higher privilege level.
  • the operation code for the gateway instruction format shown in FIG. 5A is shown below:
  • a virtual address of the gateway pointer is computed from the sum of the contents of register ra and the sign-extended value of the 12-bit immediate field.
  • the contents of 16 bytes of memory using the little-endian byte order is fetched.
  • a branch and link occurs to the low-order octlet (i.e. eight bytes) of the memory data, and the successor to the current program counter, catenated with the current execution privilege is placed in register 0 ( FIG. 5A ).
  • the privilege level is set to the contents of the low-order two bits of the memory data.
  • Register 1 is loaded with the high-order octlet of the memory data.
  • An access disallowed exception occurs if the new privilege level is greater than the privilege level required to write the memory data, or if the old privilege level is lower than the privilege required to access the memory data as a gateway. Furthermore, an access disallowed exception occurs if the target virtual address is a higher privilege than the current level and gateway access is not set for the gateway virtual address, or if the access is not aligned on a 16-byte boundary.
  • a reserved instruction exception occurs if the rb field is non-zero.
  • This operation provides a secure means to call a procedure, including those at a higher privilege level.
  • the operation code for the gateway instruction format shown in FIG. 5B is shown below:
  • a virtual address of the gateway pointer is computed from the sum of the contents of register ra and register rb.
  • the contents of 16 bytes of memory using the little-endian byte order is fetched.
  • a branch and link occurs to the low-order octlet of the memory data, and the successor to the current program counter, catenated with the current execution privilege is placed in register 0 .
  • the privilege level is set to the contents of the low-order two bits of the memory data.
  • Register 1 is loaded with the high-order octlet of the memory data.
  • An access disallowed exception occurs if the new privilege level is greater than the privilege level required to write the memory data, or if the old privilege level is lower than the privilege required to access the memory data as a gateway.
  • An access disallowed exception occurs if the target virtual address is a higher privilege than the current level and gateway access is not set for the gateway virtual address, or if the access is not aligned on a 16-byte boundary.
  • a reserved instruction exception occurs if the rb field is non-zero.
  • FIG. 6 illustrates the manner in which gateways are implemented in one embodiment of the present invention.
  • Register 1001 corresponds to the address location of the gateway instruction in virtual memory. The address stored in register 1001 points to the address ( 1002 ) where the 128 bit gateway pointers are stored.
  • Link register 1002 stores index 1011 that indicates the entry point of memory area 1010 being accessed. As shown in FIG. 6 , register 1002 points to address 1011 in memory area 1010 .
  • Data pointer 1003 stores the index into the designated data register 1021 within the memory area 1020 .
  • the gateway operation includes several steps. First, the address stored at the first pointer is loaded into the CPU's program counter, causing the CPU to start fetching instructions at that address, performing the instructions, and incrementing the program counter. Further, the previous address stored in the program counter is stored in the location of the first pointer. In addition, the second pointer indicates the address of a data register that is initialized with any data needed to process the instructions. After the access operation is performed, the value stored in the first register is reloaded into the program counter so that processing can return back to its original processing location.
  • This gateway method i.e. 1) setting the program counter to an entry point address, 2) loading another register with the program counters previous value, and then 3) initiating a data register, reduces processing steps in comparison to prior art methods that utilize CPU status registers.
  • a 64 bit program counter register is employed where the lower order two bits of the register correspond to the privilege level of the access and the rest of the bits make up the instruction address.
  • the new instruction address provided by the 128 bit gateway is loaded into the program counter register along with its corresponding new privilege level.
  • another instruction is executed which causes the program counter to be reloaded with the return address and the old privilege level. In this way, privilege level consistency is maintained.
  • FIG. 7 illustrates one embodiment of the memory system of the present invention in which independent local-to-global data buffers and independent cache and cache tags are employed for each of the instruction and data paths.
  • each of the instruction cache 111 and the data cache 121 may be selectively configured into a cache portion 300 and a buffer portion 301 ( FIG. 8 ).
  • One advantage to configuring and accessing cache 111 / 121 as a buffer is that buffer access times are fixed and typically short, compared to cache accesses which potentially are longer and variable due to cache misses.
  • Each of the cache portions 300 for each of the instruction and data caches utilizes its corresponding cache tag.
  • the cache portion in the information cache 111 utilizes cache tag 112 and the cache portion in data cache 121 utilizes cache tag 122 .
  • cache portion 300 of cache 111 is accessed, a portion of the instruction address that is used to access the cache portion is also coupled to cache tag 112 .
  • Buffer portion 301 does not make use of cache tags 112 or 122 .
  • Caches 111 and 121 are configured with 2 bits each of configuration information: instruction cache configuration (icc) information and data cache configuration (dcc) information, respectively, that are initially set in a status register when the processor is configured by software prior to running tasks.
  • 32K bytes of memory are utilized for each of instruction and data caches 111 and 121 .
  • cache 111 is divided into 4K bytes of a cache portion and 28K bytes of a buffer portion.
  • 8K bytes are utilized as cache and 24K bytes function as buffer
  • a maximum of half of the 32K bytes of cache 111 can may be used as cache and a maximum of 32K bytes of cache 111 may be used as buffer.
  • address bits a 47 -a 0 are used to access data and instruction caches 111 and 121 . Since the size of each of the cache and buffer portions for each of caches 111 and 121 is selectable, address bits a 14 -a 0 must be modified respective of the selected sizes of each of portions 300 and 301 . Address bits a 14 -a 0 are coupled to C 1 ( FIG. 7 ) in the case of the instruction cache 111 and C 2 in the case of the 15 data cache 121 . C 1 and C 2 (shown in detail in FIG. 9 ) translate addresses a 14 -a 0 depending on the configuration indicators icc and dcc, respectively.
  • bits a 14 -a 0 are not modified and are always directly coupled to buffer portion 301 no matter what icc is equal to.
  • cache 111 or 121 is configured such that there are zero bytes of cache portion 300 and 32K bytes of buffer portion 301 .
  • a 14 -a 0 are again directly coupled to cache 111 or 121 without modification.
  • a 14 is set to 1 before coupling bits a 14 -a 0 to the cache.
  • the cache portion is divided into blocks of 64 bytes.
  • 16K bytes of the cache portion is divided into 256 blocks of 64 bytes each.
  • instruction address bits a 13 -a 6 are coupled to the cache tag when the cache portion is being accessed.
  • the present invention also stores static information indicating the minimum page size of the memory.
  • a 5-bit field of data designates the page size.
  • the page size is set by system software and is determined by the number of bits that are always masked in the global TLB when performing address translation. For instance, if the lower order ten bits are masked in each cell of the global TLB, the page size is 1024 words of data.
  • portions of the local-to-global virtual address translator is shared between the instruction path and the data path.
  • the main translator is located in the data path as block 120 , and the instruction translator 110 is loaded upon demand from the data translator 120 . Consequently, the number of entries in instruction translator 110 need not be as large as that of data translator 120 .
  • the instruction translator 110 holds a single entry. The single entry is loaded from the data translator 120 whenever either a branch operation occurs or a page boundary is crossed. In one embodiment, the single entry further may eliminate the masking and matching portions of the entry, thus consisting only of XOR data or the global virtual address which is coupled to address 102 .
  • Boundary crossing detector 151 functions to evaluate whether a page boundary is potentially crossed by detecting a carry in certain bits in the instruction address, (bits a 31 -a 6 in this embodiment) as controlled by the minimum page size.
  • OR gate 152 outputs a reload buffer signal indicating that the protection information provided on bus 153 from the global TLB for that particular instruction is to be loaded into instruction protection buffer 150 .
  • the virtual memory of the present invention may be implemented in a computer system having multiple simultaneous threads of execution—referred to as a superthreaded or multi-threaded computer system.
  • the computer system is designed such that it employs a different copy of the register file, the local-to-global TLB, the data and instruction caches and the protection information buffer for each thread and only one data cache, instruction cache and global-to-physical TLB for all threads. Consequently, each of the threads of the machine may correspond to different tasks that operate independently and may be protected from each other.
  • Data may be shared or maintained independently between each thread using the virtual memory system of the present invention since each thread has a distinct version of the local-to-global TLB so that they may use the same address to mean different things, or may use the different addresses to mean the same thing, or may use the same address to reference the same memory, depending on the settings of the local-to-global translators.

Abstract

A virtual memory system including a local-to-global virtual address translator for translating local virtual addresses having associated task specific address spaces into global virtual addresses corresponding to an address space associated with multiple tasks, and a global virtual-to-physical address translator for translating global virtual addresses to physical addresses. Protection information is provided by each of the local virtual-to-global virtual address translator, the global virtual-to-physical address translator, the cache tag storage, or a protection information buffer depending on whether a cache bit or miss occurs during a given data or instruction access. The cache is configurable such that it can be configured into a buffer portion or a cache portion for faster cache accesses.

Description

This application is a continuation of application Ser. No. 08/541,416, filed Oct. 10, 1995, now U.S. Pat. No. 6,101,590.Multiple reissue applications have been filed for U.S. Pat. No. 6,427,190. This reissue is a continuation of Reissue application Ser. No. 10/901,482, filed Jul. 29, 2004 now U.S. Pat. No. Re. 39,500. The original U.S. Pat. No. 6,427,190 is a continuation of application Ser. No. 08/541,416, filed Oct. 10, 1995, now U.S. Pat. No. 6,101,590.
FIELD OF THE INVENTION
The present invention relates to computer memory systems and particularly to virtual memory systems.
BACKGROUND OF THE INVENTION
In order to enhance performance and utility in a computer system a technique called virtual memory is frequently used. One motivation for using virtual memory is to allow multiple programs to simultaneously share a computer system's main memory. This is achieved by allocating individual portions (referred to as blocks or segments) of the main memory to each of the programs being run (also referred to as a tasks). Virtual memory systems are also used in cases when a single program is too large to fit into main memory. In this case, portions of the program are stored in secondary memory and the virtual memory system assists in retrieving these portions from the secondary memory.
Virtual memory is implemented by using virtual addresses at the task or program level—each task having its own set of independent addresses. When a program performs a memory access, the virtual addresses are translated into physical addresses that may or may not be the same as other physical addresses for other tasks. The translation may be successful, leading to an access to main memory using that physical address. The translation may be unsuccessful, indicating that physical, or main memory has not been allocated for that virtual address, leading to a processor exception, from which the program may be aborted or physical memory may be allocated and the task restarted. To enhance the translation performance, virtual addresses are translated to physical addresses using information stored in a translation lookaside buffer (TLB), also known as a translation cache. The TLB provides the information that defines the mapping for each of the virtual addresses.
There are basically two categories of virtual memory systems presently utilized: paging and segmentation. Paging systems typically use fixed size blocks for allocating memory to processes. Segmentation, in contrast uses variable size blocks which may range from a value as small as one byte. Paging suffers from the disadvantage that sections of contiguous physical memory become unused because the page size is fixed: this is known as internal fragmentation. Segmentation, on the other hand, has disadvantages that the variable-sized segments may create unused regions of memory as segments are allocated, deallocated, and reallocated in arbitrary order, leaving holes in the consecutive memory allocation which become unused or unusable because they are not of a suitable size. A hybrid of the two categories has been employed in prior art systems in which segmentation and paging are both employed together.
Virtual memory systems may also employ a memory cache system to minimize virtual memory misses which includes a cache data storage and its corresponding cache tag storage. The cache stores recently accessed data and the tag storage stores a portion of the virtual address or physical address, providing the means by which it can be determined whether the cache contains the requested address. Only a portion of the address is usually required because the remaining portion of the address is used to locate (index) a reference within the cache data and tag storage, and so need not be checked again.
Caches may use either a virtual or physical address to index the cache, known as a virtual-index cache or a physical-index cache. Additionally, caches may use either a virtual or physical address stored and compared against in the cache tag storage, known as a virtual-tag cache or a physical-tag cache. Virtual-index and virtual-tag caches are generally able to attain higher peak performance, but add constraint to the mapping of addresses available when sharing data or changing the address mapping. In particular, the problem called aliasing occurs, in which two tasks use different virtual addresses to reference the same physical memory. Aliasing may require that tasks sharing memory space use identical or similar virtual addresses.
Since virtual memory allows two processes to share the same portion of physical memory such that each of the processes' virtual memory are mapped to different addresses, it is necessary to implement a protection scheme that prevents one task (i.e. a set of program instructions) from modifying a portion of memory, unless specifically allowed. Typically, tasks are assigned privilege levels which indicate the task's its ability to modify areas within physical memory and establish a control hierarchy, where higher privileged tasks are able to manipulate the storage of lower privileged tasks, including the possibility of higher privileged tasks manipulating the state of the virtual memory system itself.
One implementation of a protection scheme presently employed by virtual memory systems are “gateways” or “call gates” that function to provide a given task limited access privilege to areas in the physical memory having higher privilege than the task. The disadvantages of this prior art gateway implementation is that they utilize the CPU's status register requiring additional instructions in order to modify the status registers, and fail to provide securely initialized machine state, requiring additional instructions to initialize CPU registers used to access privileged memory registions. As a result, prior art gateway methods tend to reduce overall system performance by increasing execution times.
The present invention is a virtual memory system that performs virtual address-to-physical address translations in a manner that increases the overall efficiency and flexibility of the virtual memory system.
SUMMARY OF THE INVENTION
A virtual memory system that functions to translate a task specific virtual address (referred to as a local virtual address) into a virtual address that is generalized to all or a group of tasks (referred to as a global virtual address) and then translates the global virtual address into an address which points to a block of physical memory, (referred to as the physical address) is described. A first embodiment of the virtual memory system of the present invention includes a local-to-global virtual address translator for translating the local virtual address into the global virtual address and a global virtual-to-physical address translator for translating the global virtual address into the physical address. In an alternate embodiment, separate local-to-global virtual address translators are used for translating each of the data and instruction access addresses.
In one embodiment of the present invention, the local-to-global virtual address translator and the global virtual-to-physical address translator each include a plurality of cells, each cell implementing a single entry in a translation lookaside buffer (TLB) which defines a particular address space mapping. The TLB entry includes a match field, a mask field, an XOR field, and a protection field. Each cell includes a first logic means for matching the input address to be translated with the contents of the cell's match field to generate a match indicator output signal, a second logic means for masking the match indicator output signal with the contents of the cell's mask field to generate a masked output signal, a third logic means for generating a select signal if all of the signals mating up the masked output signal are at the same logic level, a fourth logic means for outputting the cell's XOR value if the cell is selected, and a fifth logic means for providing a protection signal when the cell is selected. Each of the translators also includes a means for multiplexing all of the XOR values from each cell and outputting the XOR value of the selected cell and a second means for multiplexing all of the protection information from each cell and outputting the protection information of the selected cell. Further, each of the translators includes a logic means for combining the XOR value from the selected cell with the address to be translated, using a bitwise exclusive-or operation to generate the translated address (either global virtual address or physical address).
In another embodiment, the local-to-global virtual address translator comprises a single cell. In this case, all of the local virtual addresses are translated by the one cell and thus this embodiment of the local-to-global virtual address translator does not require the first and second multiplexers.
The local-to-global virtual address translator, the global virtual-to-physical (GV/P) address translator and a cache tag storage provide protection information for each instruction or data access. The protection information includes the task privilege for the given access. The address's protection information is passed to the processor and is checked with the protection information associated with the corresponding local virtual, global virtual, or physical memory area that the address is attempting to access.
The present invention utilizes gateways to allow a given address to access high privilege areas of memory at certain entry points. A gateway instruction includes a gateway operation code and the gateway operation's privilege level in its protection information. The gateway instruction also provides an address that points to the gateway for the gateway operation. The gateway is a register that stores the gateway entry pointer and a data register pointer. The gateway operation comprises several steps which include storing the address of the gateway entry pointer in the processor program counter, storing the contents of the previous value stored in the program counter in a first register, initializing a second register with data to be used in the operation, performing the gateway operation, reloading the contents of the first register into the program counter.
The virtual memory system also includes a cache and a tag structure. The cache stores the accessed data or instruction that was recently retrieved by a previously translated virtual address. In one embodiment of the present invention a separate cache and tag system is used for each of the data and instruction accesses. In this embodiment, if an instruction access is being performed, the instruction cache tag utilizes the virtual address's corresponding global virtual address to tag the instruction stored in the virtual instruction cache. If a data access is being performed, the data cache tag utilizes the virtual address's corresponding physical address to tag the data stored in the virtual data cache.
The virtual cache structure also employs a means for determining whether a local virtual address cache hit or miss has occurred. If a cache hit occurs, the data in the virtual cache is accessed and the access operation is complete. If a cache miss occurs, the global virtual address is passed to the global virtual-to-physical address translator where it is translated into a physical address. The physical address is then used to access the data from either main memory or secondary memory.
BRIEF DESCRIPTION OF THE DRAWINGS
FIGS. 1 shows one embodiment of the block diagram of the virtual memory system of the present invention.
FIG. 2 illustrates an embodiment of the local virtual-to-global translator 110 of the present invention.
FIG. 3 illustrates an embodiment of the local TLB of the present invention.
FIG. 4 illustrates an embodiment of a protection field utilized to provide protection information in the virtual memory system of the present invention.
FIG. 5 illustrates an embodiment of the gateway instruction of the present invention.
FIG. 6 illustrates the manner in which gateways are implemented in an embodiment of the present invention.
FIG. 7 illustrates another embodiment of the virtual memory system of the present invention having independent data address and instruction address local-to-global virtual translation buffers and independent data and instruction cache systems.
FIG. 8 illustrates a configurable cache in accordance with the system and method of the present invention.
FIG. 9 shows one embodiment of a logic circuit employed in the system of the present invention to translate cache addresses for accessing the configurable cache of the present invention.
DETAILED DESCRIPTION
A virtual memory system is described. In the following description, numerous specific details are set forth, such as cache size, address field size and bus widths etc., in order to provide a thorough understanding of the present invention. It will be obvious, however, to one skilled in the art that these specific details need not be employed to practice the present invention. In other instances, well-known memory system structures have not been described in detail in order to avoid unnecessarily obscuring the present invention.
Overview of the Virtual Memory System
FIG. 1 shows a block diagram of one embodiment of the virtual memory system of the present invention. FIG. 1 shows an address represented in three address forms: local virtual address 100, global virtual address 102, and physical address 104. The local virtual address for a given task is a virtual address that is specific to that given task. The global virtual address is also a virtual address. However, the global virtual address is an address that resides in a virtual memory space that is common to a large group of the tasks. In one embodiment of the present invention, there is only one global virtual memory space. In another embodiment, there are several global virtual memory spaces. The physical address indicates the actual location of the address within the physical memory when it is present in the main memory.
FIG. 1 shows a 64-bit local virtual instruction address 100 comprising two portions 100a and 100b. Portions 100a and 100b correspond to the 16 and 48 bit portions of the address, respectively.
FIG. 1 also shows a local-to-global virtual address translator 110 for translating local virtual address 100. As illustrated, the higher order 16 bits of address 100, i.e. portion 100a, are coupled to the local-to-global virtual translator. Portion 100a defines a large area (referred to as a block) of virtual memory space allocated to a specific task. The lower 48 bits of address 100, (portion 100b) are the offset bits of the local virtual address. Portions 100b points to a specific area within the block.
Local-to-global virtual address translation is performed by mapping the higher order 16 bit portions 100a of local virtual address 100 to 16 bit portions 102a of global virtual address 102. As shown in FIG. 1, translator 110 outputs the 16 higher order bits (102a) of global virtual address 102. The lower order 48 bits of address 100 (i.e. 100b) are passed directly to the lower order 48 bits of global virtual address 102 (designated as portions 102b and 102c).
Global virtual-to-physical address translation is performed by translator 130. The higher order 58 bits of global virtual address 102 (i.e. portions 102a and 102b) are coupled to translator 130. Translator 130 maps the 58 bits of the selected global virtual address to 58 bits of a physical address. Portion 104a in physical address 104 corresponds to the translated 58 higher order bits of the global virtual address. The 6 lower bits of address 102 (i.e. portion 102c) are coupled to the 6 lower order bits of address 104 (i.e. portion 104b).
FIG. 1 also shows a cache storage 111 and its corresponding cache tag 112. The cache and tags are virtual-indexed and virtual-tagged to facilitate reducing memory access times. The cache storage stores the data corresponding to recently accessed virtual addresses. The cache tag stores global-virtual addresses corresponding to the data stored in the cache storage.
When an address accesses the virtual memory system shown in FIG. 1, the system performs three tasks simultaneously. First, the lower order 48 bits (i.e. the offset) of the local virtual instruction address is coupled to virtual cache 111 to access a block of data in the cache. In one embodiment of the present invention, a 32K byte cache is employed, having 512 blocks with 64 bytes each. The block of data accessed from cache 111 is coupled to the cache's data bus 133 and transmitted to the CPU. Second, while the cache is being accessed, local virtual address 100 is translated into a global virtual address 102. This is performed as described above. Third, portion 100b is coupled to cache tag 112 to access the index corresponding to the accessed cache data. The accessed index is coupled to a tag compare unit 113 to determine if it matches with the higher order 52 bits of the global virtual instruction address, portion 102a and 102b. If it does, a cache hit occurs. This means that the block of data stored in cache 111 is the desired data that corresponds to the original local virtual address. The cache hit indicator is transmitted to the processor on bus 136 and indicates to the processor that the data transmitted on the cache bus is valid. Protection information from cache tag 112 is coupled to the processor on bus 135. At this point global virtual-to-physical address translation is not necessary since valid data has been transmitted to the CPU.
If a cache miss occurs, (i.e. the accessed index provided by the cache tag does not match the higher order 52 bits in the global virtual instruction address), a cache miss is transmitted to the CPU on bus 136 indicating that the data sent on data bus 133 is not valid—at which time the CPU discards the data sent by the cache and the global virtual address is translated into a physical address. If the translation is successful, as indicated by translation hit and protection information 160, the physical address is then used to access the physical memory and the data return from memory is written into the cache storage 111 and the global-virtual address 102 and protection information 160 is written into cache tag 112. If the translation is unsuccessful, an exception is indicated to the CPU, and the physical memory access may not occur.
Physical-Tagged Caches
It should be noted that in one embodiment of the present invention, the type of addresses stored in the cache tag 112 is a physical address that can be used to access the physical memory. In such an embodiment, tag compare unit 113 is coupled to physical address 104 instead of global virtual address 102. Because a virtual-index cache is employed, the portion of the physical address which must be stored in cache tag 112 is larger and therefore a more limited set of protection information is stored in cache tag 112 and transmitted to the processor on bus 135. This information is supplemented by protection information 160 from the global virtual to physical translator 130.
Employing physical address indexes in the cache tag facilitates memory operations such as writing data out of the cache storage 111 and into physical memory when data is modified in the cache in order to retain data consistency. Another reason for using physical address indexes in the cache tag 112 is to facilitate purging data from the cache in order to make room for recently accessed memory storage. In both of these situations, since the cache tag stores physical address indexes, the data can be written into the physical memory without going through a global virtual-to-physical address translation. It should be apparent that the type of address (virtual or physical) stored in the cache tag is chosen so as to optimize the overall accessing operations in the memory system.
Separate Instruction and Data Caches
In the embodiment shown in FIG. 1, the memory system employs a single path to perform both data and instruction accesses. In another embodiment (shown in FIG. 7), both instruction memory accesses and data memory accesses are performed using partially separated paths. These accesses are each separately cached by multiple instances of cache storage, cache tag, and tag compare unit. In this embodiment, the instruction cache is virtual-indexed and virtual-tagged, and the data cache is virtual-indexed and physically-tagged. Local-virtual to global-virtual address translation is accomplished by two instances of the local-virtual to global-virtual translator. These multiple instances may be made to appear as if a single translator is shared by initializing and writing the same data and control information to each instance of the translator.
For instance, referring to FIG. 7, the instruction access path includes a local-to-global virtual address translator 110, an instruction cache 111, an instruction cache tag 112, and cache hit/miss flag 113 and the data access path includes a separate local-to-global virtual address translator 120, a data cache 121, a data cache tag 122, and a cache hit/miss flag 123. It should be understood that the above-mentioned instruction and data access path elements correspond to and function similarly to the access path elements (i.e. translator 110, cache 111, cache tag 112, and flag 113) shown in FIG. 1. Furthermore, buses 133 and 137 (FIG. 7) correspond to bus 133 (FIG. 1) in that they provide the cached instruction or data to the CPU, buses 134 and 138 correspond to bus 134 (FIG. 1) in that they provide the protection information from the local-to-global virtual address translator, buses 135 and 139 (FIG. 7) correspond to bus 135 (FIG. 1) in that they provide protection information from the cache tag, and buses 136 and 140 (FIG. 7) correspond to bus 136 (FIG. 1) in that they provide a signal to the CPU to indicate whether a cache hit or miss has occurred.
It should also be understood that since the global virtual-to-physical address translator 130 is shared between paths multiplexers 131 and 132, controlled by path select signals, are employed to route the global virtual instruction or data addresses through the shared path.
In an alternate embodiment of the present invention, local-to-global virtual address translation for both of the instruction and data addresses is performed by a single, shared translator.
Address Translation
The virtual instruction or data address is referred to as local since it belongs to a virtual address space that is unique to one specific task. Generally, in a computer system that performs multiple tasks, each task in the system has a unique local virtual address space. Accordingly, in one method of the present invention for performing local-to-global virtual translation, each of the local virtual address spaces are mapped into a common global virtual address space. An alternative embodiment, is that the local virtual address spaces are mapped into more than one global virtual address space.
FIG. 2 illustrates the local-to-global virtual address translator 110 of the present invention having four translator cells 1-4. Each translator cell comprises exclusive OR (XOR) logic gate 511 having a first parallel input coupled to the highest order 16 bits of the local virtual address (instruction or data) and a second parallel input coupled to register 402. Register 402 stores a 16 bit match value. Logic gate 511 is used to perform a bit-by-bit comparison between the higher order 16 bits from the local virtual address and the 16 bit match value. If the match value is the same as the 16 bits from the local virtual address, the XOR gate 511 outputs 16 bits that are all 0's—indicating a match. In the case in which bits differ, gate 511 outputs a “1” for the mis-matched bit locations and “0”'s for the matched bit locations. A non-zero match indicator signal from XOR gate 511 indicates a mis-match.
The 16 bit match indicator output signal from XOR gate 511 is coupled to a first parallel input of AND gate 512. The second parallel input is coupled to 16 bit mask register 401. The mask register is used to mask out certain of the 16 bits that are non-zero. The two parallel inputs of AND gate 512 are bit-by-bit ANDed together. If all of the 16 bits from XOR gate 511 are zero, then the output of AND gate 16 will also be zero no matter what the mask value is. If any of the bits from XOR gate 511 are non-zero, the mask may be employed to mask these out so that the output of AND gate 512 is 16 bits of all zeros. However, if any of the bits from AND gate 512 are non-zero and they are not masked out by mask 401 then no match occurs and the 16 bit masked output signal of AND gate 512 will contain a “1”. The 16 bits from AND gate 512 are coupled to NOR gate 513 and are NORed together. If all of the 16 bits are zero, NOR gate 513 outputs a “1” indicating that the cell is selected. If any of the 16 bits are non-zero, NOR gate 513 outputs a “0” indicating that the cell is not selected. The select signal is ANDed with each of the values in XOR register 403. If the select signal is a 1, indicating that the cell is selected, then AND gate 514 outputs the contents of XOR register 403. If the select signal is a 0, indicating the cell is not selected, AND gate 514 outputs a 16 bit word of “0”s.
FIG. 2 also shows three other cells (cells 2-4)—each cell determining whether a match exists between the 16 bits of local virtual address and match register 402. Only one of the cells will match the local virtual address at a time. Thus, only one of the cells outputs the 16 bit contents of their corresponding XOR register 403 and the rest output 16 bits of “0”s. Each of the 16 bits of each cell are ORed with the corresponding bits from the other cells (all at a “0” logic level) by OR gate 550 causing gate 550 to output the contents of the XOR value from the selected cell. This value is then XORed with the higher order 16 bits from the original local virtual address being translated by gate 560 to generate the global virtual address.
The select signal from the output of NOR gate 513 is also coupled to one input of AND gate 515. The other input of gate 515 is coupled to register 404 which contains the protection information for the global virtual address space defined by cell 1. In the event the select signal is a “1”, AND gate 515 passes the content of the protection register 404 to one input of OR gate 570. If the select signal is a “0”, then AND gate 515 outputs all “0”s. Since only one cell is selected at a time, gate 570 outputs protection information from only one cell. In other words, OR gate 570 outputs the protection information from the selected cell. This information is then transmitted to the CPU on bus 111.
It should be noted that an alternative space conservative embodiment of the present invention translator 110 comprises a single cell 1. In this embodiment, the match register is all “0”s and the mask register is all “1”s so that they become ineffective. As a result, all of the local virtual addresses are translated by the single cell having a mapping defined by the XOR register 403 in that cell. Also, since only one cell (i.e. only one global virtual address space) is used local protection is not required. Thus, the local-to-global translator would not need to generate any local protection at this point and the local protection portion of the cell would not be used.
Global virtual-to-physical address translator 130 is implement the same as translator 110 except that bus widths are increased to accommodate translation of the higher order 58 bits of the global virtual address. It should be apparent that the match, mask, and XOR values are also increased to 58 bits.
As shown in FIG. 2, translator 110 has multiple cells 1-4 each cell having its own unique set of match, mask, XOR and protection registers—each defining a different mapping and consequently a global virtual memory space having its own set of addresses. The match, mask, XOR and protection registers used with translator 110 are stored in a memory buffer referred to as the local TLB. Similarly, the match, mask, XOR, and protect registers used for translator 130 are stored in a global TLB.
FIG. 3 illustrates four entries in a local TLB corresponding to each of the cells in the translator illustrated in FIG. 2. In one embodiment, each entry is a 64 bit word having four fields; the mask field 401, the match field 402, the XOR field 403, and the protect field 404. As can be seen in FIG. 3, a different entry is used for each cell. For example, Mask[1], Match [1], XOR[1], and Protect[1] are all used by cell 1 (FIG. 2). Similarly, Mask[2], Match [2], XOR[2], and Protect[2] are all used by cell 2.
The global TLB is similarly structured, having a different field for each of the mask, match, XOR and protect values. In other embodiments of the present invention, the number of entries in the global TLB range from 64 to 256.
Protection Information
As described above, each of the local and global TLBs generate protection information by passing the contents of the cells protection register 404 (FIG. 2) from the selected cell to OR gate 570 and then to the CPU (FIGS. 1 and 2). A portion or all of the protection information is also concatenated onto the cache tag index for both the data and instruction caches (FIG. 7). The protection information field is stored within each of cache tags 112 and 122 and is concatenated onto the cache tag when these caches are accessed. FIG. 4 illustrates one embodiment of the protection information field of the present invention. As can be seen the 16 bit protection field comprises many sub-fields. The sub-fields define different aspects of a given address's ability to access data and modify it. Since the present invention's virtual memory system is particularly useful when utilized in a computer system employing multiple processors and multiple caches it is necessary to provide information to ensure that the data stored in each of the caches and the main memory are consistent. This is also referred to as maintaining cache coherency. Fields cs and cc are used to ensure cache coherency. The cc field (cache control) defines the different states (i.e. cached, coherent, noallocate, physical) of the data stored at the address being accessed and the cs field (coherence state) defines what may be done to the data when it is accessed (read, write, or replace). The p field defines the scheduling priority with which the memory access is to be handled, so that high-priority memory accesses may be performed earlier in preference to low-priority memory accesses. The d field defines that a particular exception is to be taken when accesses occur to memory controlled by this field, called a detail exception. The s field defines whether the access is to be strongly or weakly ordered with respect to other memory accesses.
The d field, when set causes a detail exception to take place. This exception passes control to a software routine which may use a variety of methods to further determine whether an exception should be handled or suppressed for this particular memory reference. For example, it may determine from the virtual address which bytes within a memory block are to be accessed, and consult a bit-map for each byte as to whether the particular type of access is to proceed. In this manner, access may be controlled at the byte level. In order to suppress the handling of the exception, the software routine sets system state which is preserved only until the target instruction is re-executed upon returning from the exception handler. The system state is then reset so that detail exceptions may occur on the next successive instruction and particularly upon any further re-execution of the same instruction as may occur after a branch operation.
The r, w, x, and g fields define the minimum privilege level of the access being performed. In one embodiment of the present invention, each field is two bits wide thus allowing for four levels of privilege for each type of access (i.e. minimum privilege=3 and maximum privilege=0). The r field indicates the access privilege level for a read access, the w field indicates the privilege level for a write access, the x field indicates the privilege level for an execute access and the g field indicates the privilege level for a gateway access. A higher privilege level can always access a memory space having a lower privilege level, but a lower privilege level cannot access data stored at a higher privilege level unless a gateway access is specified.
It should be noted that since the cache coherency information in addition to access privilege level is provided for at the local virtual, global virtual and physical memory level, cache coherency is maintained throughout the virtual memory system of the present invention at each of these levels.
Gateway Accesses
Gateway accesses provide a means of accessing high privileged areas of memory in a very controlled manner. For instance, it may be desired to protect certain types of routines, such as operating system software, in a more robust manner than other software routines stored in memory. To provide this protection, the system software is stored in an area of memory (physical or virtual) having the highest privilege level where accesses only having that privilege level is able to access the system software. Further access control to this area is imposed by ensuring that lower types of accesses are only allowed to access this high privileged area by passing through a gateway. The gateway controls the specific entrance and exit points into the privileged memory area.
Gateway accesses are implemented with gateway pointers which are 128 bits of data providing two pointers. The first pointer gives the memory address of where the higher priority area of memory is to be entered. The second pointer is 64 bits of data that can be used for a general purpose, but has the special characteristic that it is loaded as the gateway is entered and that the target procedure may depend upon it being properly loaded into a particular general register of the processor. Because of this characteristic, it is normally loaded with a pointer to a data region to be employed by the target procedure.
Gateways are implemented by the gateway instruction of the present invention. FIGS. 5A and 5B illustrate two formats of gateway instructions of the present invention.
The Branch: Gateway: Immediate (B.GATE.I) Operation
The B.GATE.I operation, provides a secure means to call a procedure, including those at a higher privilege level. The operation code for the gateway instruction format shown in FIG. 5A is shown below:
Operation Code
B.GATE.I Branch gateway immediate
For the instruction format shown in FIG. 5A, a virtual address of the gateway pointer is computed from the sum of the contents of register ra and the sign-extended value of the 12-bit immediate field. The contents of 16 bytes of memory using the little-endian byte order is fetched. A branch and link occurs to the low-order octlet (i.e. eight bytes) of the memory data, and the successor to the current program counter, catenated with the current execution privilege is placed in register 0 (FIG. 5A). The privilege level is set to the contents of the low-order two bits of the memory data. Register 1 is loaded with the high-order octlet of the memory data.
An access disallowed exception occurs if the new privilege level is greater than the privilege level required to write the memory data, or if the old privilege level is lower than the privilege required to access the memory data as a gateway. Furthermore, an access disallowed exception occurs if the target virtual address is a higher privilege than the current level and gateway access is not set for the gateway virtual address, or if the access is not aligned on a 16-byte boundary. A reserved instruction exception occurs if the rb field is non-zero.
The following is a definition in terms of an algorithm showing the operations performed to implement the B.GATE.I instruction of the present invention.
Definition
  • def BranchGatewayImmediate(ra,rb,imm) as
    • a←RegRead(ra, 64)
    • VirtAddr←a+(imm11 52 ∥imm)
    • if VirtAddr3..0≠0 then
      • raise AccessDisallowedByVirtualAddress
    • endif
    • if rb≠0 then
      • raise ReservedInstruction
    • endif
    • b←LoadMemory(VirtAddr,128,L)
    • bx←b127..64 ∥ ProgramCounter63..2+1 ∥ PrivilegeLevel
    • ProgramCounter←b63..2 ∥ 02
    • PrivilegeLevel←b1..0
    • RegWrite(rb, 128, bx)
  • enddef
  • Exceptions
  • Reserved Instruction
  • Access disallowed by virtual address
  • Access disallowed by tag
  • Access disallowed by global TLB
  • Access disallowed by local TLB
  • Access detail required by tag
  • Access detail required by local TLB
  • Access detail required by global TLB
  • Cache coherence intervention required by tag
  • Cache coherence intervention required by local TLB
  • Cache coherence intervention required by global TLB
  • Local TLB miss
  • Global TLB miss
  • Branh Gateway Operation
This operation provides a secure means to call a procedure, including those at a higher privilege level. The operation code for the gateway instruction format shown in FIG. 5B is shown below:
Operation Codes
B.GATE Branch gateway
For the instruction format shown in FIG. 5B, a virtual address of the gateway pointer is computed from the sum of the contents of register ra and register rb. The contents of 16 bytes of memory using the little-endian byte order is fetched. A branch and link occurs to the low-order octlet of the memory data, and the successor to the current program counter, catenated with the current execution privilege is placed in register 0. The privilege level is set to the contents of the low-order two bits of the memory data. Register 1 is loaded with the high-order octlet of the memory data.
An access disallowed exception occurs if the new privilege level is greater than the privilege level required to write the memory data, or if the old privilege level is lower than the privilege required to access the memory data as a gateway. An access disallowed exception occurs if the target virtual address is a higher privilege than the current level and gateway access is not set for the gateway virtual address, or if the access is not aligned on a 16-byte boundary. A reserved instruction exception occurs if the rb field is non-zero.
The following is a definition in terms of an algorithm showing the operations performed to implement the B.GATE instruction of the present invention.
Definition
  • def BranchGateway(ra,rb,rc) as
    • a←RegRead(ra, 64)
    • b←RegRead(rb, 64)
    • VirtAddr←a+b
    • if VirtAddr3.0≠0 then
      • raise AccessDisallowedByVirtualAddress
    • endif
    • if rc≠0 then
      • raise ReservedInstruction
    • endif
    • c←LoadMemory(VirtAddr,128,L)
    • cx←c127..64 ∥ ProgramCounter63..2+1 ∥ PrivilegeLevel
    • ProgramCounter←c63.22 ∥0
    • PrivilegeLevel←c1..0
    • RegWrite(rc, 128, cx)
  • enddef
  • Exceptions
  • Reserved Instruction
  • Access disallowed by virtual address
  • Access disallowed by tag
  • Access disallowed by global TLB
  • Access disallowed by local TLB
  • Access detail required by tag
  • Access detail required by local TLB
  • Access detail required by global TLB
  • Cache coherence intervention required by tag
  • Cache coherence intervention required by local TLB
  • Cache coherence intervention required by global TLB
  • Local TLB miss
  • Global TLB miss
FIG. 6 illustrates the manner in which gateways are implemented in one embodiment of the present invention. Register 1001 corresponds to the address location of the gateway instruction in virtual memory. The address stored in register 1001 points to the address (1002) where the 128 bit gateway pointers are stored. Link register 1002 stores index 1011 that indicates the entry point of memory area 1010 being accessed. As shown in FIG. 6, register 1002 points to address 1011 in memory area 1010. Data pointer 1003 stores the index into the designated data register 1021 within the memory area 1020.
The gateway operation includes several steps. First, the address stored at the first pointer is loaded into the CPU's program counter, causing the CPU to start fetching instructions at that address, performing the instructions, and incrementing the program counter. Further, the previous address stored in the program counter is stored in the location of the first pointer. In addition, the second pointer indicates the address of a data register that is initialized with any data needed to process the instructions. After the access operation is performed, the value stored in the first register is reloaded into the program counter so that processing can return back to its original processing location. This gateway method, i.e. 1) setting the program counter to an entry point address, 2) loading another register with the program counters previous value, and then 3) initiating a data register, reduces processing steps in comparison to prior art methods that utilize CPU status registers.
In one embodiment of the present invention a 64 bit program counter register is employed where the lower order two bits of the register correspond to the privilege level of the access and the rest of the bits make up the instruction address. At the beginning of the gateway operation, the new instruction address provided by the 128 bit gateway is loaded into the program counter register along with its corresponding new privilege level. At the end of the gateway procedure, another instruction is executed which causes the program counter to be reloaded with the return address and the old privilege level. In this way, privilege level consistency is maintained.
Cache Partitioning
FIG. 7 illustrates one embodiment of the memory system of the present invention in which independent local-to-global data buffers and independent cache and cache tags are employed for each of the instruction and data paths. In this embodiment, each of the instruction cache 111 and the data cache 121 may be selectively configured into a cache portion 300 and a buffer portion 301 (FIG. 8). One advantage to configuring and accessing cache 111/121 as a buffer is that buffer access times are fixed and typically short, compared to cache accesses which potentially are longer and variable due to cache misses.
Each of the cache portions 300 for each of the instruction and data caches utilizes its corresponding cache tag. For instance, the cache portion in the information cache 111 utilizes cache tag 112 and the cache portion in data cache 121 utilizes cache tag 122. When cache portion 300 of cache 111 is accessed, a portion of the instruction address that is used to access the cache portion is also coupled to cache tag 112. Buffer portion 301, on the other hand, does not make use of cache tags 112 or 122. Caches 111 and 121 are configured with 2 bits each of configuration information: instruction cache configuration (icc) information and data cache configuration (dcc) information, respectively, that are initially set in a status register when the processor is configured by software prior to running tasks.
In one embodiment of the present invention, 32K bytes of memory are utilized for each of instruction and data caches 111 and 121. In the case of instruction cache 111, when icc=0, zero bytes of cache 111 function as a cache portion and 32K bytes of cache 111 function as a buffer portion. For icc=1, cache 111 is divided into 4K bytes of a cache portion and 28K bytes of a buffer portion. For icc=2, 8K bytes are utilized as cache and 24K bytes function as buffer, and finally for icc=3, 16K bytes are cache and 16K bytes are buffer. Thus, for this embodiment, a maximum of half of the 32K bytes of cache 111 can may be used as cache and a maximum of 32K bytes of cache 111 may be used as buffer.
In addition to the configuration information, a flag bit in the instruction address (a47, for this embodiment) is used to indicate whether the access is being performed with cache portion 300 or buffer portion 301 of either of caches 111 or 121. If a47=1, then buffer portion 301 is accessed with the address in the instruction register and if a47=0, cache 300 is accessed.
In the embodiment shown in FIG. 7, address bits a47-a0 are used to access data and instruction caches 111 and 121. Since the size of each of the cache and buffer portions for each of caches 111 and 121 is selectable, address bits a14-a0 must be modified respective of the selected sizes of each of portions 300 and 301. Address bits a14-a0 are coupled to C1 (FIG. 7) in the case of the instruction cache 111 and C2 in the case of the 15 data cache 121. C1 and C2 (shown in detail in FIG. 9) translate addresses a14-a0 depending on the configuration indicators icc and dcc, respectively. In the case in which a47=1, bits a14-a0 are not modified and are always directly coupled to buffer portion 301 no matter what icc is equal to. Similarly, for a47=0 and icc=0 cache 111 or 121 is configured such that there are zero bytes of cache portion 300 and 32K bytes of buffer portion 301. As a result, a14-a0 are again directly coupled to cache 111 or 121 without modification. In the case in which a47=0 and icc=3, 16K bytes of cache portion 300 are being accessed. As a result, a14 is set to 1 before coupling bits a14-a0 to the cache. When a47=0 and icc=2, 8K bytes of cache portion 300 are being accessed and a14 and a13 are set to 1 before coupling bits a14-a0 to the cache. Finally, if a47=0 and icc=1, 4K bytes of cache portion 300 are being accessed and a14, a13, a12 are set to 1 before coupling bits a14-a0 to the cache 111 or 121.
In one embodiment of the present invention the cache portion is divided into blocks of 64 bytes. Thus, in the case of the maximum cache portion size condition (i.e. icc=3) 16K bytes of the cache portion, is divided into 256 blocks of 64 bytes each. In order to access each of the 256 blocks, eight address bits are needed to access each cache entry (i.e. 28=256). In the embodiment shown in FIG. 7, instruction address bits a13-a6 are coupled to the cache tag when the cache portion is being accessed.
Boundary Crossing Detector
The present invention also stores static information indicating the minimum page size of the memory. For the embodiment shown in FIG. 7, a 5-bit field of data designates the page size. The page size is set by system software and is determined by the number of bits that are always masked in the global TLB when performing address translation. For instance, if the lower order ten bits are masked in each cell of the global TLB, the page size is 1024 words of data.
In one embodiment of the present invention, portions of the local-to-global virtual address translator is shared between the instruction path and the data path. The main translator is located in the data path as block 120, and the instruction translator 110 is loaded upon demand from the data translator 120. Consequently, the number of entries in instruction translator 110 need not be as large as that of data translator 120. In one embodiment, the instruction translator 110 holds a single entry. The single entry is loaded from the data translator 120 whenever either a branch operation occurs or a page boundary is crossed. In one embodiment, the single entry further may eliminate the masking and matching portions of the entry, thus consisting only of XOR data or the global virtual address which is coupled to address 102.
When the information cache 111 is accessed as a cache portion (a47=0), cache tag 112 is simultaneously accessed and provides previously cached protection information to the processor on bus 135 to multiplexer 154, (FIG. 7). In the case when cache 111 is accessed as a buffer portion (a47=1) and the cache tag is not used, protection information buffer 150 provides the appropriate protection information on bus 156 to multiplexer 154, (FIG.7). Instruction address bit a47 which indicates whether instruction cache 111 is configured as a buffer or a cache controls multiplexer 154 so that it passes the correct protection information to the CPU. Protection information 160 (FIG. 7) is loaded into the protection information buffer from the global translator 130 whenever either a branch operation occurs or a page boundary is crossed.
Boundary crossing detector 151 functions to evaluate whether a page boundary is potentially crossed by detecting a carry in certain bits in the instruction address, (bits a31-a6 in this embodiment) as controlled by the minimum page size. When the instruction address indicates that a boundary is potentially being crossed or the instruction corresponds to a branch operation, OR gate 152 outputs a reload buffer signal indicating that the protection information provided on bus 153 from the global TLB for that particular instruction is to be loaded into instruction protection buffer 150.
Protection information buffer 150 is used when translating consecutive addresses, such as instruction addresses. For instance, when an instruction address is first translated by local-to-global TLB 110 and then translated by global-to-physical TLB 130, protection information is provided to the processor by the global TLB. For the previously described embodiment shown in FIG. 1, if the next instruction address to be translated is consecutive with respect to the previous address it will also be translated through the local and global TLBs. Since the consecutive address has all of the same higher order address bits as the previously translated address the global TLB would provide the same protection information as that of the previously translated address.
In the embodiment of the present invention shown in FIG. 7, protection information buffer 150 provides a means of by-passing the step of translating consecutive addresses through the global TLB to obtain the protection information as is done in the FIG. 1 embodiment of the present invention. Instead, protection information 160 from the previously translated address is loaded in protection information buffer 150 when a boundary is detected and it is transmitted to the CPU when the consecutive address is being translated if 1) a47=1, i.e. the buffer portion of instruction cache 111 is being accessed and 2) a given number of higher order bits (as specified by the minimum page size) in the local virtual address match the same given number of higher order bits in the local virtual address of the previously translated address.
The virtual memory of the present invention may be implemented in a computer system having multiple simultaneous threads of execution—referred to as a superthreaded or multi-threaded computer system. The computer system is designed such that it employs a different copy of the register file, the local-to-global TLB, the data and instruction caches and the protection information buffer for each thread and only one data cache, instruction cache and global-to-physical TLB for all threads. Consequently, each of the threads of the machine may correspond to different tasks that operate independently and may be protected from each other. Data may be shared or maintained independently between each thread using the virtual memory system of the present invention since each thread has a distinct version of the local-to-global TLB so that they may use the same address to mean different things, or may use the different addresses to mean the same thing, or may use the same address to reference the same memory, depending on the settings of the local-to-global translators.
Thus, a virtual memory system utilizing local virtual and global virtual addresses is described. Although the elements of the present invention have been described in conjunction with a certain embodiment, it is appreciated that the invention may be implemented in a variety of other ways. Consequently, it is to be understood that the particular embodiment shown and described by way of illustration are in no way intended to be considered limiting. Reference to the details of these embodiments is not intended to limit the scope of the claims which themselves recite only those features regarded as essential to the invention.

Claims (38)

1. A memory storage system for storing recently accessed data from a main memory in a computer system, said memory storage system comprising:
a memory storage area which is configurable into a cache portion and a buffer portion; and
means for storing indices corresponding to data stored in said memory storage area;
wherein when said cache portion is accessed by a given address, said index storage means is also accessed by said address to check if said data accessed from said cache portion is valid and when said buffer portion is accessed by said given address, said index storage means is not checked.
2. The memory system as described in claim 1 wherein said index storage means further includes protection information indicating the access privilege of said given address into said main memory and when said index storage means is accessed by said given address, said protection information is concatenated onto an address accessed from said index storage means by said given address and provided to a CPU.
3. The memory system as described in claim 2, wherein said protection information includes a field that defines a coherence state of data stored at said given address, wherein said coherence state indicates whether data stored at said given address may be read, written into, or replaced.
4. The memory system as described in claim 2 wherein said protection information includes a field that defines an access priority of said given address, wherein said access priority indicates the order at which said given address is accessed with respect to other accesses in said memory system.
5. The memory system as described in claim 2 wherein said protection information includes a field for indicating when a detail exception should occur.
6. The memory system as described in claim 2 wherein said protection information includes a field that defines a cache control condition of said given address, wherein said cache control condition indicates states of said data stored in said given address including a cache coherent state, a non-allocated state, and a physical state.
7. The memory system as described in claim 1, wherein said given address includes an indication of whether an access is being performed on at least one of said cache portion and said buffer portion.
8. A method for storing recently accessed data from a main memory in a computer system, comprising the steps of:
configuring a memory storage area into a cache portion and a buffer portion; and
storing indices corresponding to data stored in said memory storage area;
wherein when said cache portion is accessed by a given address, said stored indices are also accessed by said given address to check if said data accessed from said cache portion is valid and when said buffer portion is accessed by said given address, said stored indices are not checked.
9. The method of claim 8, wherein said given address includes an indication of whether an access is being performed on at least one of said cache portion and said buffer portion.
10. The method of claim 9, wherein said protection information includes a field that defines a coherence state of data stored at said given address, wherein said coherence state indicates whether data stored at said given address may be read, written into, or replaced.
11. The method of claim 9, wherein said protection information includes a field that defines an access priority of said given address, wherein said access priority indicates the order at which said given address is accessed with respect to other accesses in said memory system.
12. The method of claim 9, wherein said protection information includes a field for indicating when a detail exception should occur.
13. The method of claim 9, wherein said protection information includes a field that defines a cache control condition of said given address, wherein said cache control condition indicates states of said data stored in said given address including a cache coherent state, a non-allocated state, and a physical state.
14. The method of claim 8, wherein said stored indices further include protection information indicating the access privilege of said given address into said main memory and when said stored indices are accessed by said given address, said protection information is concatenated onto an address accessed from said stored indices by said given address and provided to a CPU.
15. A computer-readable medium containing a program that performs the steps of:
receiving an indication that a memory storage area has been configured into a cache portion and a buffer portion; and
storing indices corresponding to data stored in said memory storage area;
wherein when said cache portion is accessed by a given address, said stored indices are also accessed by said given address to check if said data accessed from said cache portion is valid and when said buffer portion is accessed by said given address, said stored indices are not checked.
16. The method of claim 15, wherein said given address includes an indication of whether an access is being performed on at least one of said cache portion and said buffer portion.
17. The computer-readable medium of claim 16, wherein said protection information includes a field that defines a coherence state of data stored at said given address, wherein said coherence state indicates whether data stored at said given address may be read, written into, or replaced.
18. The computer-readable medium of claim 16, wherein said protection information includes a field that defines an access priority of said given address, wherein said access priority indicates the order at which said given address is accessed with respect to other accesses in said memory system.
19. The computer-readable medium of claim 16, wherein said protection information includes a field for indicating when a detail exception should occur.
20. The computer-readable medium of claim 16, wherein said protection information includes a field that defines a cache control condition of said given address, wherein said cache control condition indicates states of said data stored in said given address including a cache coherent state, a non-allocated state, and a physical state.
21. The computer-readable medium of claim 15, wherein said stored indices further include protection information indicating the access privilege of said given address into said main memory and when said stored indices are accessed by said given address, said protection information is concatenated onto an address accessed from said stored indices by said given address and provided to a CPU.
22. A method of storing data accessed by a processor, the method comprising:
receiving configuration information;
accessing a main memory;
partitioning a memory storage area into one of a predetermined number of combinations of a cache portion size and a buffer portion size in accordance with the received configuration information;
receiving an address within an address space divided into a cache address space portion and a buffer address space portion, wherein the address includes a field for indicating whether an access includes a cache access or a buffer access;
modifying at least a portion of the address using the configuration information, to produce a modified address; and
accessing data in said memory storage area by means of the modified address.
23. The method of claim 22 further comprising:
if a cache miss is detected in an access to the cache portion, accessing the main memory for transferring data to or from the main memory,
wherein an access to the buffer portion does not result in an access to the main memory.
24. The method of claim 22 further comprising:
storing indices corresponding to data stored in the memory storage area in a cache tag memory;
upon accessing the cache portion, checking the cache tag memory to determine whether the cache tag memory contains an index corresponding to the cache access; and
generating a cache hit if a corresponding index is found, and generating a cache miss if a corresponding index is not found.
25. The method of claim 22 wherein, in an access to the cache portion, the memory storage system determines whether an access to the main memory is required, and in an access to the buffer portion, no access to the main memory is required.
26. The method of claim 22 wherein buffer accesses have a fixed access time and cache accesses have a variable access time due to cache misses.
27. The method of claim 22, wherein an address bit in said address indicates whether the access is a cache access or a buffer access.
28. The method of claim 22 further comprising:
partitioning the memory storage area between the cache portion and the buffer portion in accordance with configuration information received by the memory storage system.
29. The method of claim 22 further comprising:
setting the configuration information in a status register when the computer system is configured by software.
30. The method of claim 22 further comprising:
allocating address ranges to the cache portion and the buffer portion in accordance with the partitioning of the memory storage area between the cache portion and the buffer portion.
31. The method of claim 22 wherein the address space is divided into a cache address space portion and a buffer address space portion in accordance with at least the state of an address bit within the address space.
32. The method of claim 22 further comprising:
delivering a portion of the address to the memory storage area without modification.
33. The method of claim 22 further comprising:
delivering the modified address to the cache portion.
34. The method of claim 22 wherein the modified address is used to access the entire memory storage area for accesses within the buffer address space portion.
35. A method of storing data accessed by a processor, the method comprising:
accessing a main memory;
receiving configuration information;
configuring a memory storage area into a cache portion and a buffer portion by partitioning the memory storage area between the cache portion and the buffer portion in accordance with the configuration information into one of a predetermined number of combinations of cache portion size and buffer portion size in accordance with configuration information received by the memory storage system;
receiving an address within an address space divided into a cache address space portion and a buffer address space portion;
modifying at least a portion of the address using the configuration information, to produce a modified address; and
accessing data in said memory storage area by means of an address that includes a field for indicating whether the access includes a cache access or a buffer access.
36. The method of claim 35 further comprising:
delivering a portion of the address to the memory storage area without modification.
37. The method of claim 35 further comprising:
delivering the modified address to the cache portion.
38. The method of claim 35 wherein the modified address is used to access the entire memory storage area for accesses within the buffer address space portion.
US11/606,321 1995-10-10 2006-11-30 Configurable cache allowing cache-type and buffer-type access Expired - Lifetime USRE43798E1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/606,321 USRE43798E1 (en) 1995-10-10 2006-11-30 Configurable cache allowing cache-type and buffer-type access

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US08/541,416 US6101590A (en) 1995-10-10 1995-10-10 Virtual memory system with local and global virtual address translation
US09/570,402 US6427190B1 (en) 1995-10-10 2000-05-12 Configurable cache allowing cache-type and buffer-type access
US10/901,482 USRE39500E1 (en) 1995-10-10 2004-07-29 Configurable cache allowing cache-type and buffer-type access
US11/606,321 USRE43798E1 (en) 1995-10-10 2006-11-30 Configurable cache allowing cache-type and buffer-type access

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09/570,402 Reissue US6427190B1 (en) 1995-10-10 2000-05-12 Configurable cache allowing cache-type and buffer-type access

Publications (1)

Publication Number Publication Date
USRE43798E1 true USRE43798E1 (en) 2012-11-06

Family

ID=24159506

Family Applications (5)

Application Number Title Priority Date Filing Date
US08/541,416 Expired - Lifetime US6101590A (en) 1995-10-10 1995-10-10 Virtual memory system with local and global virtual address translation
US09/570,397 Expired - Lifetime US6256715B1 (en) 1995-10-10 2000-05-12 System and method of performing gateway access
US09/570,402 Ceased US6427190B1 (en) 1995-10-10 2000-05-12 Configurable cache allowing cache-type and buffer-type access
US10/901,482 Expired - Lifetime USRE39500E1 (en) 1995-10-10 2004-07-29 Configurable cache allowing cache-type and buffer-type access
US11/606,321 Expired - Lifetime USRE43798E1 (en) 1995-10-10 2006-11-30 Configurable cache allowing cache-type and buffer-type access

Family Applications Before (4)

Application Number Title Priority Date Filing Date
US08/541,416 Expired - Lifetime US6101590A (en) 1995-10-10 1995-10-10 Virtual memory system with local and global virtual address translation
US09/570,397 Expired - Lifetime US6256715B1 (en) 1995-10-10 2000-05-12 System and method of performing gateway access
US09/570,402 Ceased US6427190B1 (en) 1995-10-10 2000-05-12 Configurable cache allowing cache-type and buffer-type access
US10/901,482 Expired - Lifetime USRE39500E1 (en) 1995-10-10 2004-07-29 Configurable cache allowing cache-type and buffer-type access

Country Status (4)

Country Link
US (5) US6101590A (en)
AU (1) AU7595596A (en)
TW (1) TW376488B (en)
WO (1) WO1997014084A2 (en)

Families Citing this family (72)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6101590A (en) * 1995-10-10 2000-08-08 Micro Unity Systems Engineering, Inc. Virtual memory system with local and global virtual address translation
DE19735948C1 (en) * 1997-08-19 1998-10-01 Siemens Nixdorf Inf Syst Method for improving controllability in data processing equipment with translation-look-aside-buffer (TLB)
JPH11238016A (en) * 1998-02-23 1999-08-31 Nec Corp Method and device for memory protection
JP2000057039A (en) * 1998-08-03 2000-02-25 Canon Inc Method and device for controlling access, file system and information processor
US6574721B1 (en) * 1999-08-31 2003-06-03 International Business Machines Corporation Apparatus and method for providing simultaneous local and global addressing using software to distinguish between local and global addresses
US6948011B1 (en) * 1999-12-07 2005-09-20 Advanced Micro Devices Alternate Register Mapping
US6594750B1 (en) * 1999-12-08 2003-07-15 Ati International Srl Method and apparatus for handling an accessed bit in a page table entry
US6704851B2 (en) * 2000-04-26 2004-03-09 Aicas Gmbh Method of dynamically allocating a memory
US7043633B1 (en) * 2000-08-28 2006-05-09 Verizon Corporation Services Group Inc. Method and apparatus for providing adaptive self-synchronized dynamic address translation
US8037530B1 (en) 2000-08-28 2011-10-11 Verizon Corporate Services Group Inc. Method and apparatus for providing adaptive self-synchronized dynamic address translation as an intrusion detection sensor
US6826684B1 (en) * 2000-08-28 2004-11-30 Verizon Corporate Services Group Inc. Sliding scale adaptive self-synchronized dynamic address translation
US7739497B1 (en) * 2001-03-21 2010-06-15 Verizon Corporate Services Group Inc. Method and apparatus for anonymous IP datagram exchange using dynamic network address translation
US6779099B2 (en) * 2001-07-20 2004-08-17 Chien-Tzu Hou Operation method for controlling access attributes of a memorized page of a memory unit and its structure
US6516387B1 (en) * 2001-07-30 2003-02-04 Lsi Logic Corporation Set-associative cache having a configurable split and unified mode
US20030046492A1 (en) * 2001-08-28 2003-03-06 International Business Machines Corporation, Armonk, New York Configurable memory array
US6839826B2 (en) * 2002-02-06 2005-01-04 Sandisk Corporation Memory device with pointer structure to map logical to physical addresses
US7055015B2 (en) * 2002-02-25 2006-05-30 Ricoh Company, Ltd. Information processing apparatus in which processes can reduce overhead of memory access and efficiently share memory
US7543087B2 (en) * 2002-04-22 2009-06-02 Alacritech, Inc. Freeing transmit memory on a network interface device prior to receiving an acknowledgement that transmit data has been received by a remote device
US7308580B2 (en) * 2002-04-23 2007-12-11 International Business Machines Corporation System and method for ensuring security with multiple authentication schemes
US7272550B2 (en) * 2002-04-23 2007-09-18 International Business Machines Corporation System and method for configurable binding of access control lists in a content management system
US7284265B2 (en) * 2002-04-23 2007-10-16 International Business Machines Corporation System and method for incremental refresh of a compiled access control table in a content management system
US6976023B2 (en) * 2002-04-23 2005-12-13 International Business Machines Corporation System and method for managing application specific privileges in a content management system
US6912643B2 (en) * 2002-08-19 2005-06-28 Aristos Logic Corporation Method of flexibly mapping a number of storage elements into a virtual storage element
US7529906B2 (en) * 2003-08-18 2009-05-05 Cray Incorporated Sharing memory within an application using scalable hardware resources
US7577816B2 (en) * 2003-08-18 2009-08-18 Cray Inc. Remote translation mechanism for a multinode system
US7266786B2 (en) * 2002-11-05 2007-09-04 Sonics, Inc. Method and apparatus for configurable address mapping and protection architecture and hardware for on-chip systems
US7020760B2 (en) * 2002-12-16 2006-03-28 International Business Machines Corporation Hybrid logical block virtualization system for a storage area network
US7269825B1 (en) * 2002-12-27 2007-09-11 Unisys Corporation Method and system for relative address translation
US6981125B2 (en) * 2003-04-22 2005-12-27 International Business Machines Corporation Method and apparatus for managing shared virtual storage in an information handling system
US7334110B1 (en) 2003-08-18 2008-02-19 Cray Inc. Decoupled scalar/vector computer architecture system and method
US7366873B1 (en) 2003-08-18 2008-04-29 Cray, Inc. Indirectly addressed vector load-operate-store method and apparatus
US8307194B1 (en) 2003-08-18 2012-11-06 Cray Inc. Relaxed memory consistency model
US7743223B2 (en) * 2003-08-18 2010-06-22 Cray Inc. Decoupling of write address from its associated write data in a store to a shared memory in a multiprocessor system
US7437521B1 (en) 2003-08-18 2008-10-14 Cray Inc. Multistream processing memory-and barrier-synchronization method and apparatus
US7421565B1 (en) 2003-08-18 2008-09-02 Cray Inc. Method and apparatus for indirectly addressed vector load-add -store across multi-processors
US7543133B1 (en) * 2003-08-18 2009-06-02 Cray Inc. Latency tolerant distributed shared memory multiprocessor computer
US7735088B1 (en) 2003-08-18 2010-06-08 Cray Inc. Scheduling synchronization of programs running as streams on multiple processors
US7519771B1 (en) 2003-08-18 2009-04-14 Cray Inc. System and method for processing memory instructions using a forced order queue
US7503048B1 (en) 2003-08-18 2009-03-10 Cray Incorporated Scheduling synchronization of programs running as streams on multiple processors
US7647358B2 (en) * 2004-03-22 2010-01-12 Microsoft Corporation Computing device with relatively limited storage space and operating/file system thereof
US8069192B2 (en) * 2004-03-22 2011-11-29 Microsoft Corporation Computing device with relatively limited storage space and operating / file system thereof
US7370171B1 (en) * 2004-04-26 2008-05-06 Sun Microsystems, Inc. Scalable buffer control for a tracing framework
US7206915B2 (en) * 2004-06-03 2007-04-17 Emc Corp Virtual space manager for computer having a physical address extension feature
US20060004941A1 (en) * 2004-06-30 2006-01-05 Shah Hemal V Method, system, and program for accessesing a virtualized data structure table in cache
US8250296B2 (en) * 2004-12-01 2012-08-21 Dell Products L.P. System and method for information handling system memory page mapping optimization
US7890721B2 (en) * 2005-02-16 2011-02-15 Atmel Corporation Implementation of integrated status of a protection register word in a protection register array
US7478769B1 (en) 2005-03-09 2009-01-20 Cray Inc. Method and apparatus for cooling electronic components
US20060230198A1 (en) * 2005-03-30 2006-10-12 Kabushiki Kaisha Toshiba Data processing circuit, data processing method and image forming apparatus
US20090043956A1 (en) * 2005-04-20 2009-02-12 Simon Ford Mapping an input data value to a resultant data value
US7616210B2 (en) * 2005-08-23 2009-11-10 Canon Kabushiki Kaisha Memory apparatus and memory control method
US8065499B2 (en) * 2006-02-22 2011-11-22 Oracle America, Inc. Methods and apparatus to implement parallel transactions
US8028133B2 (en) * 2006-02-22 2011-09-27 Oracle America, Inc. Globally incremented variable or clock based methods and apparatus to implement parallel transactions
US7809922B2 (en) * 2007-10-21 2010-10-05 International Business Machines Corporation Translation lookaside buffer snooping within memory coherent system
US7979681B2 (en) * 2007-11-20 2011-07-12 Qualcomm Incorporated System and method of selectively accessing a register file
US8239879B2 (en) * 2008-02-01 2012-08-07 International Business Machines Corporation Notification by task of completion of GSM operations at target node
US8484307B2 (en) * 2008-02-01 2013-07-09 International Business Machines Corporation Host fabric interface (HFI) to perform global shared memory (GSM) operations
US8214604B2 (en) * 2008-02-01 2012-07-03 International Business Machines Corporation Mechanisms to order global shared memory operations
US8146094B2 (en) * 2008-02-01 2012-03-27 International Business Machines Corporation Guaranteeing delivery of multi-packet GSM messages
US7873879B2 (en) * 2008-02-01 2011-01-18 International Business Machines Corporation Mechanism to perform debugging of global shared memory (GSM) operations
US8255913B2 (en) * 2008-02-01 2012-08-28 International Business Machines Corporation Notification to task of completion of GSM operations by initiator node
US8200910B2 (en) * 2008-02-01 2012-06-12 International Business Machines Corporation Generating and issuing global shared memory operations via a send FIFO
US8275947B2 (en) * 2008-02-01 2012-09-25 International Business Machines Corporation Mechanism to prevent illegal access to task address space by unauthorized tasks
JP2009205445A (en) * 2008-02-28 2009-09-10 Sony Corp Address conversion circuit
US8312219B2 (en) 2009-03-02 2012-11-13 International Business Machines Corporation Hybrid caching techniques and garbage collection using hybrid caching techniques
US9116845B2 (en) 2011-02-23 2015-08-25 Freescale Semiconductor, Inc. Remote permissions provisioning for storage in a cache and device therefor
US8949551B2 (en) 2011-02-23 2015-02-03 Freescale Semiconductor, Inc. Memory protection unit (MPU) having a shared portion and method of operation
US9767268B2 (en) 2011-04-20 2017-09-19 International Business Machines Corporation Optimizing a compiled access control table in a content management system
US8639895B2 (en) 2011-07-14 2014-01-28 Freescale Semiconductor, Inc. Systems and methods for memory region descriptor attribute override
US9940069B1 (en) * 2013-02-27 2018-04-10 EMC IP Holding Company LLC Paging cache for storage system
US10353818B1 (en) 2013-02-27 2019-07-16 EMC IP Holding Company LLC Dataset paging cache for storage system
US9530000B2 (en) * 2013-06-14 2016-12-27 Microsoft Technology Licensing, Llc Secure privilege level execution and access protection
CN109376123B (en) * 2014-08-12 2022-08-19 华为技术有限公司 Method for managing files, distributed storage system and management node

Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3938097A (en) * 1974-04-01 1976-02-10 Xerox Corporation Memory and buffer arrangement for digital computers
US4442484A (en) 1980-10-14 1984-04-10 Intel Corporation Microprocessor memory management and protection mechanism
US4580240A (en) 1981-12-15 1986-04-01 Nippon Electric Co., Ltd. Memory arrangement operable as a cache and a local memory
US4985825A (en) 1989-02-03 1991-01-15 Digital Equipment Corporation System for delaying processing of memory access exceptions until the execution stage of an instruction pipeline of a virtual memory system based digital computer
US5067078A (en) 1989-04-17 1991-11-19 Motorola, Inc. Cache which provides status information
US5091846A (en) 1986-10-03 1992-02-25 Intergraph Corporation Cache providing caching/non-caching write-through and copyback modes for virtual addresses and including bus snooping to maintain coherency
US5132927A (en) 1990-10-09 1992-07-21 Tandem Computers Incorporated System for cache space allocation using selective addressing
EP0497600A2 (en) 1991-02-01 1992-08-05 Canon Kabushiki Kaisha Memory access method and apparatus
US5179675A (en) 1988-09-30 1993-01-12 International Computers Limited Data processing system with cache memory addressable by virtual and physical address
US5222223A (en) 1989-02-03 1993-06-22 Digital Equipment Corporation Method and apparatus for ordering and queueing multiple memory requests
US5226133A (en) 1989-12-01 1993-07-06 Silicon Graphics, Inc. Two-level translation look-aside buffer using partial addresses for enhanced speed
US5319760A (en) 1991-06-28 1994-06-07 Digital Equipment Corporation Translation buffer for virtual machines with address space match
US5390310A (en) 1991-09-30 1995-02-14 Apple Computer, Inc. Memory management unit having cross-domain control
US5410669A (en) 1993-04-05 1995-04-25 Motorola, Inc. Data processor having a cache memory capable of being used as a linear ram bank
US5420998A (en) 1992-04-10 1995-05-30 Fujitsu Limited Dual memory disk drive
US5423019A (en) 1991-12-19 1995-06-06 Opti Inc. Automatic cache flush with readable and writable cache tag memory
US5430888A (en) 1988-07-25 1995-07-04 Digital Equipment Corporation Pipeline utilizing an integral cache for transferring data to and from a register
US5446861A (en) 1979-06-04 1995-08-29 Unisys Corporation Variable data rate improvement of disc cache subsystem
US5557769A (en) * 1994-06-17 1996-09-17 Advanced Micro Devices Mechanism and protocol for maintaining cache coherency within an integrated processor
US5586293A (en) 1991-08-24 1996-12-17 Motorola, Inc. Real time cache implemented by on-chip memory having standard and cache operating modes
US5619673A (en) 1994-06-29 1997-04-08 Intel Corporation Virtual access cache protection bits handling method and apparatus
US5822762A (en) 1994-12-12 1998-10-13 Fujitsu Limited Information processing device with decision circuits and partitioned address areas
US6000012A (en) * 1996-12-09 1999-12-07 International Business Machines Corporation Method and apparatus for prioritizing and routing commands from a command source to a command sink
US6101590A (en) * 1995-10-10 2000-08-08 Micro Unity Systems Engineering, Inc. Virtual memory system with local and global virtual address translation
US6484237B1 (en) 1999-07-15 2002-11-19 Texas Instruments Incorporated Unified multilevel memory system architecture which supports both cache and addressable SRAM
US6606686B1 (en) 1999-07-15 2003-08-12 Texas Instruments Incorporated Unified memory system architecture including cache and directly addressable static random access memory

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4315310A (en) * 1979-09-28 1982-02-09 Intel Corporation Input/output data processing system
US4500952A (en) * 1980-05-23 1985-02-19 International Business Machines Corporation Mechanism for control of address translation by a program using a plurality of translation tables
US4539637A (en) * 1982-08-26 1985-09-03 At&T Bell Laboratories Method and apparatus for handling interprocessor calls in a multiprocessor system
US4710868A (en) * 1984-06-29 1987-12-01 International Business Machines Corporation Interconnect scheme for shared memory local networks
US4792895A (en) * 1984-07-30 1988-12-20 International Business Machines Corp. Instruction processing in higher level virtual machines by a real machine
US4695948A (en) * 1985-02-28 1987-09-22 International Business Machines Corporation Bus to bus converter using a RAM for multiple address mapping
US4809160A (en) * 1985-10-28 1989-02-28 Hewlett-Packard Company Privilege level checking instruction for implementing a secure hierarchical computer system
IT1228728B (en) * 1989-03-15 1991-07-03 Bull Hn Information Syst MULTIPROCESSOR SYSTEM WITH GLOBAL DATA REPLICATION AND TWO LEVELS OF ADDRESS TRANSLATION UNIT.
US5303378A (en) * 1991-05-21 1994-04-12 Compaq Computer Corporation Reentrant protected mode kernel using virtual 8086 mode interrupt service routines
US5430850A (en) * 1991-07-22 1995-07-04 Massachusetts Institute Of Technology Data processing system with synchronization coprocessor for multiple threads
US6006318A (en) * 1995-08-16 1999-12-21 Microunity Systems Engineering, Inc. General purpose, dynamic partitioning, programmable media processor
US5742840A (en) * 1995-08-16 1998-04-21 Microunity Systems Engineering, Inc. General purpose, multiple precision parallel operation, programmable media processor

Patent Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3938097A (en) * 1974-04-01 1976-02-10 Xerox Corporation Memory and buffer arrangement for digital computers
US5446861A (en) 1979-06-04 1995-08-29 Unisys Corporation Variable data rate improvement of disc cache subsystem
US4442484A (en) 1980-10-14 1984-04-10 Intel Corporation Microprocessor memory management and protection mechanism
US4580240A (en) 1981-12-15 1986-04-01 Nippon Electric Co., Ltd. Memory arrangement operable as a cache and a local memory
US5091846A (en) 1986-10-03 1992-02-25 Intergraph Corporation Cache providing caching/non-caching write-through and copyback modes for virtual addresses and including bus snooping to maintain coherency
US5430888A (en) 1988-07-25 1995-07-04 Digital Equipment Corporation Pipeline utilizing an integral cache for transferring data to and from a register
US5179675A (en) 1988-09-30 1993-01-12 International Computers Limited Data processing system with cache memory addressable by virtual and physical address
US5222223A (en) 1989-02-03 1993-06-22 Digital Equipment Corporation Method and apparatus for ordering and queueing multiple memory requests
US4985825A (en) 1989-02-03 1991-01-15 Digital Equipment Corporation System for delaying processing of memory access exceptions until the execution stage of an instruction pipeline of a virtual memory system based digital computer
US5067078A (en) 1989-04-17 1991-11-19 Motorola, Inc. Cache which provides status information
US5226133A (en) 1989-12-01 1993-07-06 Silicon Graphics, Inc. Two-level translation look-aside buffer using partial addresses for enhanced speed
US5132927A (en) 1990-10-09 1992-07-21 Tandem Computers Incorporated System for cache space allocation using selective addressing
EP0497600A2 (en) 1991-02-01 1992-08-05 Canon Kabushiki Kaisha Memory access method and apparatus
US5319760A (en) 1991-06-28 1994-06-07 Digital Equipment Corporation Translation buffer for virtual machines with address space match
US5586293A (en) 1991-08-24 1996-12-17 Motorola, Inc. Real time cache implemented by on-chip memory having standard and cache operating modes
US5390310A (en) 1991-09-30 1995-02-14 Apple Computer, Inc. Memory management unit having cross-domain control
US5423019A (en) 1991-12-19 1995-06-06 Opti Inc. Automatic cache flush with readable and writable cache tag memory
US5420998A (en) 1992-04-10 1995-05-30 Fujitsu Limited Dual memory disk drive
US5410669A (en) 1993-04-05 1995-04-25 Motorola, Inc. Data processor having a cache memory capable of being used as a linear ram bank
US5557769A (en) * 1994-06-17 1996-09-17 Advanced Micro Devices Mechanism and protocol for maintaining cache coherency within an integrated processor
US5619673A (en) 1994-06-29 1997-04-08 Intel Corporation Virtual access cache protection bits handling method and apparatus
US5822762A (en) 1994-12-12 1998-10-13 Fujitsu Limited Information processing device with decision circuits and partitioned address areas
US6101590A (en) * 1995-10-10 2000-08-08 Micro Unity Systems Engineering, Inc. Virtual memory system with local and global virtual address translation
US6000012A (en) * 1996-12-09 1999-12-07 International Business Machines Corporation Method and apparatus for prioritizing and routing commands from a command source to a command sink
US6484237B1 (en) 1999-07-15 2002-11-19 Texas Instruments Incorporated Unified multilevel memory system architecture which supports both cache and addressable SRAM
US6606686B1 (en) 1999-07-15 2003-08-12 Texas Instruments Incorporated Unified memory system architecture including cache and directly addressable static random access memory

Non-Patent Citations (18)

* Cited by examiner, † Cited by third party
Title
"Organization of the Motorola 88110 Superscalar RISC Microprocessor", Diefendorff et al, 1992, p. 40-63. *
Amendment dated Dec. 1, 2008 in a Reexamination No. 90/007,563.
Amendment dated Feb. 22, 2007 in a Reexamination No. 90/007,563.
Amendment dated Jun. 26, 2009 in a Reexamination No. 90/007,563.
Amendment dated Nov. 1, 2007 in a Reexamination No. 90/007,563.
David A. Wood et al., "Supporting Reference and Dirty Bits in Spur's Virtual Address Cache," Computer Science Division, Electrical Engineering and Computer Science Dept., 1989, pp. 122-130.
Defendants' Exhibit Q Chart from MicroUnity Systems Engineering, Inc. v. Acer, Inc. et al., Case No. 2:10-CV-91 filed Mar. 16, 2010, Eastern District of Texas, Marshall Division.
Diefendorff et al., "Organization of the Motorola 88110 Superscalar RISC Microprocessor", IEEE Micro, Apr. 1992.
Final Office Action dated May 22, 2009 in a Reexamination No. 90/007,563.
Final Office Action dated Sep. 1, 2007 in a Reexamination No. 90/007,563.
Interview Summary dated Jul. 31, 2008 in a Reexamination No. 90/007,563.
Interview Summary dated Jun. 22, 2007 in a Reexamination No. 90/007,563.
MC88110 Second Generation RISC Microprocessor User's Manual, Motorola Inc., 1991. p. 1-626. *
Motorola, Inc., MC88200 Cache/Memory Management Unit User's Manual, 1988, 261 pages.
Non-Final Office Action dated Dec. 22, 2006 in a Reexamination No. 90/007,563.
Notice of Intent to Issue a Reexam Certificate dated Aug. 21, 2009 in a Reexamination No. 90/007,563.
R. H. Katz et al., "Implementing a Cache Consistency Protocol," Computer Science Division, Electrical Engineering and Computer Science Dept., 1985, pp. 1-31.
Schroder et al., "A Hardware for Implementing Protection Rings", Communications of the ACM, Mar. 1972, vol. 15, No. 3.

Also Published As

Publication number Publication date
US6427190B1 (en) 2002-07-30
US6256715B1 (en) 2001-07-03
TW376488B (en) 1999-12-11
WO1997014084A3 (en) 1997-10-23
USRE39500E1 (en) 2007-02-27
US6101590A (en) 2000-08-08
WO1997014084A2 (en) 1997-04-17
AU7595596A (en) 1997-04-30

Similar Documents

Publication Publication Date Title
USRE43798E1 (en) Configurable cache allowing cache-type and buffer-type access
US5075848A (en) Object lifetime control in an object-oriented memory protection mechanism
CA1313424C (en) Nonhierarchical program authorization mechanism
US6266755B1 (en) Translation lookaside buffer with virtual address conflict prevention
US5075845A (en) Type management and control in an object oriented memory protection mechanism
US4979098A (en) Multiple address space token designation, protection controls, designation translation and lookaside
US4277826A (en) Synchronizing mechanism for page replacement control
US5265227A (en) Parallel protection checking in an address translation look-aside buffer
US4136385A (en) Synonym control means for multiple virtual storage systems
US5963984A (en) Address translation unit employing programmable page size
US6823433B1 (en) Memory management system and method for providing physical address based memory access security
US5835925A (en) Using external registers to extend memory reference capabilities of a microprocessor
US7068545B1 (en) Data processing apparatus having memory protection unit
US5481734A (en) Data processor having 2n bits width data bus for context switching function
US5680564A (en) Pipelined processor with two tier prefetch buffer structure and method with bypass
US7721066B2 (en) Efficient encoding for detecting load dependency on store with misalignment
US5418927A (en) I/O cache controller containing a buffer memory partitioned into lines accessible by corresponding I/O devices and a directory to track the lines
US5758120A (en) Method and system for increased system memory concurrency in a multi-processor computer system utilizing concurrent access of reference and change bits
GB2367648A (en) Avoidance of repeated trap handling
US10474575B2 (en) Cache-based communication between execution threads of a data processing system
US6751700B2 (en) Date processor and storage system including a set associative cache with memory aliasing
US6519684B1 (en) Low overhead method for selecting and updating an entry in a cache memory
US6757809B1 (en) Data processor having 2n bits width data bus for context switching functions
US6598050B1 (en) Apparatus and method for limited data sharing in a multi-tasking system
KR100218617B1 (en) Method and system for efficient memory management in a data processing system utilizing a dual mode translation lookaside buffer

Legal Events

Date Code Title Description
FPAY Fee payment

Year of fee payment: 12

SULP Surcharge for late payment

Year of fee payment: 11