WO1998028714A1 - Enhanced texture map data fetching circuit and method - Google Patents

Enhanced texture map data fetching circuit and method Download PDF

Info

Publication number
WO1998028714A1
WO1998028714A1 PCT/US1997/023982 US9723982W WO9828714A1 WO 1998028714 A1 WO1998028714 A1 WO 1998028714A1 US 9723982 W US9723982 W US 9723982W WO 9828714 A1 WO9828714 A1 WO 9828714A1
Authority
WO
WIPO (PCT)
Prior art keywords
texture map
texture
memory
cache
address
Prior art date
Application number
PCT/US1997/023982
Other languages
French (fr)
Inventor
Vincent W. Wang
Jih-Hsien Soong
Hongjun Shu
Tzoyao Chan
Original Assignee
Cirrus Logic, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Cirrus Logic, Inc. filed Critical Cirrus Logic, Inc.
Priority to JP52908498A priority Critical patent/JP4110239B2/en
Priority to EP97953482A priority patent/EP0946929A1/en
Priority to CA002275727A priority patent/CA2275727A1/en
Publication of WO1998028714A1 publication Critical patent/WO1998028714A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0875Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches with dedicated cache, e.g. instruction or stack
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/04Texture mapping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects

Definitions

  • the present invention relates to the field of computer controlled graphics display systems. Specifically, the present invention relates to a system and method for enhancing data throughput in a texture map data retrieval subsystem.
  • the computer controlled graphics display systems typically provide data and control signals to graphics hardware units (e.g., "graphics cards") which contain specialized circuits and encoded procedures for processing graphics instructions at high speeds.
  • graphics instructions are typically stored in a "display list" within computer memory.
  • the instructions define the rendering of several graphic primitives, e.g., individual points, lines, polygons, fills, BIT BLTs (bit block transfers), textures, etc., and graphics commands. Collections of graphics primitives can be used to render a two dimensional image on a display screen of an object that is represented in three dimensional space. Rendering involves translating the above graphics primitives and graphics instructions into raster encoded data that is then loaded into a frame buffer memory for display (“refresh") on the display screen.
  • Some polygon graphics primitives include specifications of texture data, representative of graphic images, to be displayed within the polygon.
  • Texture mapping refers to techniques for adding surface detail to areas or surfaces of the polygons displayed on the two dimensional display screen. Since the original graphics object is three dimensional, texture mapping often involves maintaining certain perspective attributes with respect to the surface detail added to the primitive.
  • a typical texture map includes point elements ("texels") which reside in a (u, v) texture coordinate space.
  • a texture image is represented in computer memory as a bitmap or other raster-based encoded format.
  • the display screen includes point elements (pixels) which reside in an (x, y) display coordinate space.
  • the process of texture mapping occurs by accessing encoded surface detail points or "texels" from a memory unit that stores the surface detail (e.g., an image) and transferring the surface detail texels to predetermined points of the graphics primitive to be texture mapped.
  • the individual texels of the texture map data are read out of memory and applied within their polygon in particular fashions depending on the placement and perspective of their associated polygon.
  • color values for pixels in (x, y) display coordinate space are determined based on sampled texture map values.
  • the process of texture mapping operates by applying color or visual attributes of texels of the (u, v) texture map to corresponding pixels of the graphics primitive on the display screen. After texture mapping, a version of the texture image is visible on surfaces of the graphics primitive, with the proper perspective, if any.
  • texture mapping requires a great demand on the memory capacity of the graphics display system because a lot of texture maps are accessed from memory during a typical display screen update cycle. Since the frequency of the screen update cycles is rapid, the individual polygons of the screen (and related texture map data per polygon) need to be accessed and updated at an extremely rapid frequency requiring great data throughput capacities.
  • high performance graphics hardware units typically contain low access time cache memory units and cache memory controller units for storing and retrieving texture mapped data at high speeds.
  • the cache controller unit of the prior art system stalls while the desired texture data is obtained from external memory. Usually, there is a long latency (stall) from the cache controller unit sending out the external memory request until the texture data is actually fetched from the external memory.
  • the present invention provides a system and method for providing efficient texture map data retrieval in a computer controlled graphics display system.
  • the present invention provides a system as above that provides a texture data cache controller unit that can perform useful texture data processing while waiting for fetched texture data associated with a texture cache miss.
  • Figure 1 is a block diagram of a computer controlled graphics display system in accordance with the present invention.
  • FIG. 2 is a block diagram of the 3D graphics subunit of the computer controlled graphics display system in accordance with the present invention.
  • FIG. 3 is a block diagram of the texture map data access (TDA) circuit of the graphics unit in accordance with the present invention.
  • TDA texture map data access
  • Figure 4A and Figure 4B illustrate, respectively, a texture map data request miss address and a texture map data request hit address being placed on the top entry of the texture map address (TMA) FLFO memory unit of the present invention.
  • TMA texture map address
  • Figure 5A and Figure 5B illustrate, respectively, a texture map data request hit address being placed on the top entry of the texture map address FIFO memory unit after the FIFO configurations of Figure 4 A and Figure 4B.
  • Figure 6A illustrates a first texture engine stall condition whereby an (n+l) m unprocessed miss address is received by the cache controller of the present invention.
  • Figure 6B illustrates a second texture engine stall condition whereby a miss address reaches the last entry of the texture map address FIFO memory unit but the cache controller has not yet finished its data fetch operation.
  • Figure 7A is an illustration of a FIFO configuration of the texture map address FIFO memory unit wherein fetched texture data is supplied for miss address that is the oldest address in the texture map address FIFO memory unit.
  • Figure 7B is an illustration of a FIFO configuration of the texture map address FIFO memory unit as shown in Figure 7 A with an (n+l) m miss address simultaneously entering into the texture map address FIFO memory unit.
  • Figure 7C, Figure 7D, Figure 7E, and Figure 7F illustrate, respectively, texture data stored in the memory cache for a first, second, third, and fourth hit address being supplied simultaneously with the processing of a data fetch operation for a miss address stored in an upper entry of the texture map address FIFO memory unit.
  • Figure 7G illustrates that during the processing of a data fetch operation for the miss address of Figures 7B-7F, subsequently received hit addresses were placed within the upper portion of the texture map address FIFO memory unit.
  • Figure 8 illustrates steps within a flow chart diagram of the operation of the present invention for placing texture data hit and miss addresses onto the texture map address FIFO memory unit.
  • Figure 9 illustrates steps within a flow chart diagram of the operation of the present invention for removing texture data hit and miss addresses from the texture map address FIFO memory unit.
  • a circuit and method are described for increasing the processing efficiency of texture map data requests within a computer controlled graphics display system.
  • a 3D graphics display subunit is included in the graphics display system and this subunit includes a polygon engine, a texture map engine, and pixel pipeline.
  • the texture map engine contains a texture map data access (TDA) circuit having a cache controller with a computer readable cache memory for containing recently used texture maps stored in (u,v) coordinate space.
  • the cache controller is limited in handling only n cache miss operations simultaneously. In one embodiment, n is 1.
  • the TDA circuit also contains a texture map address FIFO memory unit for storing texture map addresses associated with texture data requests that hit or missed in the cache memory unit.
  • the texture engine stalls when the (n+l) tn unprocessed texture request miss is encountered. Therefore, the TMA FIFO at any time contains at most n miss addresses therein. Processing efficiency is increased when a miss is encountered but the TMA FLFO contains unprocessed hit addresses.
  • it can also advantageously: (1) supply data from the cache memory for the previously encountered and stored hit addresses; and (2) accept new hit addresses into the TMA FLFO thereby effectively avoiding a texture engine stall. This is quite unlike the prior art systems which process no hit addresses upon a texture miss but rather stall the texture engine.
  • embodiments of the present invention include a computer controlled graphics display system having a host processor for executing instructions and processing data, a bus coupled to the host processor, a display screen for displaying graphics images, and a circuit for retrieving texture map data, the circuit including: main memory (or a local frame buffer) for containing texture map data used for texture mapping operations of graphics primitives; a texture map cache memory for containing recently accessed sub-texture map data of the texture map data of the main memory (or local frame buffer memory); a cache lookup circuit coupled to a cache controller circuit for determining if a received texture map address is a hit address or a miss address with respect to contents of the cache memory; a FLFO memory coupled to the cache lookup circuit for storing a plurality of received texture map addresses; and a cache controller circuit coupled to the main memory, the FLFO memory, and the texture map cache memory, the cache controller circuit (1) for fetching, during a fetch interval, first texture map data from the main memory for storage into the texture map cache memory, the first texture map data corresponding
  • Embodiments include the above and wherein the cache controller is also for providing, during the fetch interval, third texture map data stored within the texture map cache memory, the third texture map data corresponding to a second texture map hit address stored in a third location of the FLFO memory, the second texture map hit address received by the FLFO memory before the texture map miss address and wherein the FIFO memory is also for receiving and storing texture map hit addresses during the fetch interval.
  • the random access memory unit coupled with the bus 100 for storing information and instructions for the host processor 101, a computer readable non-volatile memory unit 103 (e.g., read only memory unit) coupled with the bus 100 for storing static information and instructions for the host processor 101, a computer readable data storage device 104 such as a magnetic or optical disk and disk drive (e.g., hard drive or floppy diskette) coupled with the bus 100 for storing information and instructions, and a display device 105 coupled to the bus 100 for displaying information to the computer user.
  • the display device 105 utilized with the computer system 112 of the present invention can be a liquid crystal device, cathode ray tube, or other display device suitable for creating graphic images and alphanumeric characters recognizable to the user.
  • the host computer system 112 provides data and control signals via bus 100 to a graphics hardware unit or system, e.g., "graphics card” 108.
  • the graphics hardware system 108 contains a 3D graphics subunit 109 for executing a series of display instructions found within a display list stored in computer memory.
  • the display list generally contains instructions regarding the rendering of several graphic primitives, e.g., individual points, lines, polygons, fills, BIT BLTs (bit block transfers), textures, etc. Many of the polygon display instructions include texture data to be displayed within the polygon.
  • Texture data is stored in computer readable (e.g., volatile) memory units of system 112, or local frame buffer 110) in the form of raster based data (e.g., in one form its bit mapped) stored in (u,v) coordinates.
  • the individual components (e.g., "texels") of the texture data are read out of memory and applied within their polygon in particular fashions depending on the placement and perspective of their associated polygon.
  • the process of rendering a polygon with associated texture data is called "texture mapping.” Texture mapping requires a great demand on the memory capacity of the computer system 112 because many texture maps are accessed from memory to construct a displayed frame. Since screen updates need to be performed rapidly, polygons need to be updated very rapidly and further texture maps need to be accessed and applied in extremely rapid fashion, increasing memory demands.
  • FIG. 1 illustrates a portion of the circuitry of the graphics subunit 109 including a texture engine 10, a polygon engine 12 and a pixel pipeline 16.
  • the texture engine 10 receives polygon vertex data over bus 5 that corresponds to respective polygons to be rendered.
  • the polygon vertex data includes data points for each vertex of the polygon.
  • each of the three vertexes contains: its own position coordinate values (x, y, z); its own color values (red, green, blue); its own texture map coordinate values (u, v), its own perspective value (w), and other required values including an identification of the texture map data for the polygon, if any.
  • the texture engine 10 is responsible for retrieving the texture map data for the polygon and mapping the texels of the texture data onto the pixels of the polygon. Once the texture engine 10 is given the texture map coordinates (u,v) for each vertex of a triangle, it can go to the texture cache controller 250 and access the matching texels for placement into the triangle. During this process, the texture engine 10 maintains the three dimensional perspective of the surface of the polygon.
  • Texture map data retrieval (TDA) circuit 200 performs texture map data retrieval processes in accordance with the present invention. Aside from the texture retrieval system of texture engine 10, a number of well known procedures and circuits can be used to maintain the perspective and perform the texture mapping operations implemented within texture engine 10. Texture map pixel data is supplied from the texture engine 10 to the pixel pipeline 16 over bus 14a.
  • the polygon engine 12 of Figure 2 receives the polygon data over bus 5 and performs well known polygon rendering functions with respect to the position, color, and perspective of the polygon primitive. Essentially, polygon engine uses interpolation to compute the pixel positions and colors of the pixels within the polygon primitive based on the polygon vertex data. Pixel information resulting from the polygon engine 12 is forwarded to the pixel pipeline over bus 14b.
  • the pixel pipeline 16 blends the texture data (texels) from the texture engine 10 with the pixel data from the polygon engine 12 to form a composite polygon image.
  • the data (pixels) of the composite image are forwarded in raster encoded format over bus 18, stored in a raster encoded frame buffer (situated within graphics subunit 109, but not shown in Figure 2), and eventually displayed on display screen 105 ( Figure 1).
  • the above operation is performed individually for each received polygon primitive.
  • the pixel pipeline 16, in one embodiment, contains a latency of approximately 5-7 clock cycles, depending on programmable image features.
  • FIG. 3 illustrates the components of the texture map data access (TDA) circuit 200 in accordance with the present invention.
  • TDA circuit 200 provides an efficient mechanism for processing texture map data requests (in the form of addresses) whereby useful texture map data can be supplied from a cache memory 251 to the filter 260 simultaneously during a fetch interval wherein other texture data is being fetched from main memory 102 or from a local frame buffer 110 ( Figure 1). This operation is unlike the prior art where during a fetch interval, the texture engine stalls and no useful texture map data is provided until the fetch completes. Further, during the fetch interval, later received hit addresses can be accepted by the TDA circuit 200 thereby preventing the stall out of circuits that supply addresses to the TDA circuit 200.
  • the polygon data received by the texture engine 10 contains an indication (e.g., a pointer) of the texture map (e.g., a base address of the texture map) used by the currently processed polygon.
  • the TDA circuit 200 contains an address control unit 210 which receives, over bus 205, the above referenced texture map base address.
  • the texture map base address information is also forwarded over bus 207 to an address generator unit 220.
  • the address generator unit 220 is controlled by control signals of line 237 that originate from the address control unit 210.
  • Address generator unit 220 also receives (u,v) texture coordinates corresponding to the current polygon primitive.
  • the address generator unit 220 generates the (u,v) texel addresses and level of detail (LOD) information for each corresponding pixel of a polygon primitive. These newly “received” texture map addresses (“texel addresses”) are individually forwarded from unit 220 to a cache lookup unit 230.
  • address generator unit 220 inputs the u_main, v_main, du_main, du_ortho, dv_main, and dv_ortho to compute the required texel addresses.
  • these terms are defined in copending patent application serial number , filed , and entitled "Non-Homogenous Second Order
  • the cache lookup unit 230 of Figure 3 uses control and address bus 235 to examine the contents of a texture map cache memory circuit ("cache memory") 251 to determine if each received texture map address corresponds to texture map data stored in the cache memory 251.
  • the cache memory 251 is situated within a cache controller circuit 250 and in one embodiment is a 16-bit fully-associated cache of a predetermined size (e.g., Ik bytes). In one embodiment, the cache memory 251 is divided into 16 sets (64 bytes each).
  • a received texture map address corresponds to texture map data stored in the cache circuit 251
  • it is a texture map hit address (“hit address”
  • a received texture map address does not correspond to texture map data stored in the cache circuit 251
  • it is a texture map miss address ("miss address”).
  • a fetch interval is required for the cache controller circuit 250 causing it to fetch the required texture map data from main memory 102 or local frame buffer 110 for storage into the cache memory 251.
  • a block of texture data representing portions of several screen lines e.g., 64 bytes corresponding to a single set size
  • the fetched texture map data is fetched in 64 byte blocks (corresponding to a set) each comprising an 8x8 matrix of texture map data, in one embodiment of the present invention.
  • the fetch operation uses bus 202 which is communicatively interfaced with bus 100 of system 112 to receive texture map data from main memory 102 or bus 100" from local frame buffer 110.
  • bus 202 which is communicatively interfaced with bus 100 of system 112 to receive texture map data from main memory 102 or bus 100" from local frame buffer 110.
  • a number of well known circuits and techniques can be used by the cache controller circuit 250 within the scope of the present invention for implementing the texture map data fetch operation.
  • cache controller circuit 250 and cache memory 251 can be used to supply texture map data, stored within cache memory 251, that corresponds to texture map hit addresses that were received before the commencement of the fetch interval.
  • a first-in-first-out (FLFO) memory circuit 240 of Figure 3 is coupled to input received texture map addresses from cache lookup unit 230.
  • FLFO memory 240 also called TMA FIFO
  • Texture map addresses that are at the bottom entry (physical or logical) of the FLFO memory unit 240 are processed such that their corresponding texture map data is retrieved from the cache memory unit 251 and supplied over bus 253 if the texel is available in the texture cache 251.
  • the cache memory unit 251 is addressed by the bottom entry of FIFO memory 240 over address and control bus 238. If the texture map data corresponding to the bottom entry address is not yet available in the cache memory 251 , FLFO memory 240 is stalled until the data becomes available. When the corresponding texture map data is supplied from cache memory 251 , the bottom entry texture map address is removed from FLFO memory 240.
  • Filter unit 260 receives texture map data over bus 253 from the cache controller circuit 250 and performs a number of well known data filtering operations, including performing linear filtering, bi-linear filtering, and tri-linear filtering. The filter unit 260 then outputs the texture map data for the polygon to an optional output FIFO 270 which is coupled to external bus 273. Bus 273 is coupled to bus 14a ( Figure 2) to the pixel pipeline circuit 16.
  • Figure 4A illustrates a FIFO configuration wherein the FIFO memory 240 is empty and a miss address is pushed into the top entry (1).
  • cache controller 250 commences a fetch interval to obtain the texture map data for the miss address.
  • Figure 4B illustrates a FLFO configuration wherein the FIFO memory 240 is empty and a hit address is pushed into the top entry (1).
  • Figure 5A illustrates the FIFO configuration of Figure 4A, but a later received hit address is next pushed onto the top entry (1) and the miss address is then copied into entry (2). It is appreciated that in this configuration, the fetch interval for the miss address of entry (2) is not yet complete.
  • Figure 5A illustrates that multiple later received hit addresses can be pushed onto the
  • FIG. 5B illustrates the FIFO configuration of Figure 4B, but a later received hit address is next pushed onto the top entry (1) and the first hit address is popped out of the FIFO 240.
  • Figure 6A and Figure 6B illustrate two FLFO configurations that cause a FLFO memory 240 stall condition.
  • the FLFO memory 240 and the TDA circuit 200 does not accept any more texture map addresses.
  • Figure 6A illustrates a FLFO stall configuration where an early received miss address is near the bottom of the FIFO memory 240 in entry (m'), a number of hit addresses were received and stored in entries (1) through (m'-l), and then another miss address 310a is received.
  • the miss address in entry (m') is still pending, e.g., its fetch interval is still open.
  • this condition causes a FLFO stall because the cache controller circuit 250 can only process one fetch interval at a time. Therefore, the FIFO memory 240 stalls and does not accept new texture map addresses until it removes the miss address in entry
  • Figure 6B illustrates a second FIFO stall configuration where a pending miss address reaches the bottom entry, (m), of the of the FIFO memory 240 and the remainder of the FLFO memory 240, entries (1) - (m-1), are filled with hit addresses.
  • the texture map data corresponding to the miss address in entry (m) is not yet available because its fetch interval is not yet complete.
  • the FLFO memory 240 stalls until the texture map data for the miss address in entry (m) becomes available, at which time this miss address is removed from the FLFO memory 240.
  • Figure 7A illustrates that the "bottom" entry of the FIFO memory 240 can be viewed as the actual last physical entry (m) or as a logical bottom entry defined as the entry that contains the oldest texture map address.
  • the last physical entry (m) of the FIFO memory 240 can be viewed as the actual last physical entry (m) or as a logical bottom entry defined as the entry that contains the oldest texture map address.
  • FLFO memory 240 is used to address the cache memory 251 to retrieve the corresponding texture map data.
  • the logical bottom entry (n ⁇ ) can also be used in another embodiment of the present invention.
  • a miss address is at position (m') being the oldest address within the FIFO memory 240 and other subsequently received hit addresses are stored in entries (1) - (m'-l).
  • the fetch interval for the miss address at entry (m') completes causing its corresponding texture data to be supplied from cache memory 251.
  • the miss address in entry (m') is then removed from the FIFO memory 240.
  • Figure 7B- Figure7G illustrate an effective use of the TDA circuit 200 of the present invention for texture map data retrieval wherein texture map data is supplied to the texture engine 10 from the TDA circuit 200 during a fetch interval.
  • Figure 7B illustrates a FLFO configuration where a first miss address is pending and stored in the (m 1 ) entry, several hit addresses are stored in entries (1) through (m'-l) of FLFO memory 240, and then a second miss address 320a is received.
  • the cache controller circuit 250 commences a fetch interval to retrieve the texture map data from main memory 102 for the second miss address. Simultaneous with this fetch interval, the cache controller circuit 250 also accesses the cache memory 251 with the hit address (address 1) located at the bottom entry (m') of the FIFO memory 240 (either physical bottom or logical bottom). The cache controller circuit 250 then supplies the texture map data corresponding to this hit address 1 to the texture filter 260. Therefore, the present invention TDA circuit 200 advantageously allows texture data to be supplied during a texture data fetch interval.
  • FIFO memory 240 to advantageously accept new hit addresses.
  • the cache controller circuit 250 also accesses the cache memory 251 with the hit address (address2) located at the bottom entry (m') of the FLFO memory 240 (either physical bottom or logical bottom). The cache controller 250 then supplies the texture map data corresponding to this hit address2 to the texture filter 260.
  • the cache controller circuit 250 also accesses the cache memory 251 with the hit address (address3) located at the bottom entry (m') of the FIFO memory 240 (either physical bottom or logical bottom). The cache controller 250 then supplies the texture map data corresponding to this hit address3 to the texture filter 260.
  • the cache controller circuit 250 also accesses the cache memory 251 with the hit address (address4) located at the bottom entry (m') of the FLFO memory 240.
  • the cache controller 250 then supplies the texture map data corresponding to this hit address4 to the texture filter 260. This process can continue for each next clock cycle provided: (1) there are previously received hit addresses, relative to the pending miss address; and (2) the fetch interval is still pending. It is appreciated that with reference to Figure7C through Figure 7F, after each hit address is processed, it is popped of off the FIFO memory 240.
  • Figure 7G illustrates the FIFO configuration at the completion of the pending fetch interval (t > 4) when the requested texture data for the miss address becomes stored in the cache memory 251.
  • the cache controller circuit 250 accesses the cache memory 251 with the miss address located at the bottom entry (m') of the FIFO memory 240.
  • the cache controller 250 then supplies the texture map data corresponding to this miss address to the texture filter 260.
  • the hit addresses located in entries (1) - (4) are processed during the next four clock cycles as new texture map addresses are received at the top of FLFO memory 240.
  • the TDA circuit 200 of the present invention advantageously processes hit addresses to supply texture map data during a fetch interval. Also during this interval, new hit addresses are accepted by the FIFO memory unit 240.
  • the entry size of the FIFO memory 240 should be roughly equal to the number of clock cycles required to perform a data fetch by the cache controller 250. In one embodiment of the present invention this value is 20 cycles, e.g., (m) should be roughly 20 entries.
  • each texture map data fetch interval involves the fetching of a block of texture map data, e.g., 64 bytes each corresponding to either four lines of 16 bytes each or eight lines of 8 bytes each.
  • TDA circuit 200 is very effective because texture mapping and texture filtering procedures often operate on texture data that are closely situated within a given address space neighborhood. Therefore, texture map address misses are not usually received back- to-back in time, but are rather spaced out between hit addresses. In this expected operational environment, FLFO memory 240 does not stall often as a result of encountering two pending miss addresses thus providing good data throughput of the cache memory 251.
  • FIG. 8 illustrates steps of process 400 performed by the TDA circuit 200 of the present invention for placing new texture map addresses (hit/miss) into the entries of FLFO memory 240.
  • Process 400 commences at step 410 where it is checked if a new texture map address is received at address generator unit 220, e.g., new (u,v) coordinates and texture map base address. If so, the address generator circuit 220 computes a new texture map address at step 410.
  • the cache lookup circuit 230 checks if the new address is a hit address or a miss address. If the address is a hit address, then step 430 is entered where the hit address is pushed into the FIFO 240 and step 410 is re-entered. At step 420, if the address is a miss address then step 440 is entered.
  • TDA circuit 200 pushes the miss address into the FIFO 240. This commences a fetch interval for the pending miss address.
  • step 450 when another newly received texture map address is received by the TMA circuit 200, then step 455 is entered.
  • the cache lookup circuit 230 checks if the new address is a hit address or a miss address. If the new address is a hit address then at step 460 the TMA circuit 200 checks if there is vacancy in the FLFO 240. If there is vacancy, then step 440 is entered where the newly received address is placed into the FIFO
  • step 460 if there is no vacancy in the FIFO 240, then step 465 is entered where the TMA circuit 200 stalls and waits until the pending texture data fetching interval is complete so that a new
  • the TMA circuit 200 checks if another miss address is already pending within the TMA FLFO 240. If so, then at step 475 the TMA circuit 200 stalls and waits until the pending texture data fetching interval is complete so that the newly received miss address can be supplied to the cache controller
  • step 440 is then entered to push the new miss address into the FLFO 240.
  • a fetch interval is commenced causing the cache controller circuit 250 to access main memory 102 (or the local frame buffer 1 10) to retrieve the texture map data for the newly stored pending miss address.
  • step 470 if another miss address is not already pending in FLFO 240, then step 440 is entered to push the new miss address into the FIFO 240.
  • a fetch interval is commenced causing the cache controller circuit 250 to access main memory 102 (or local frame buffer 110) to retrieve the texture map data for the newly stored pending miss address.
  • FIG. 9 illustrates steps of process 500 performed by the TDA circuit 200 of the present invention for removing texture map addresses from the bottom entry of FIFO memory 240.
  • Process 500 commences at step 510 where TDA circuit 200 checks if FIFO 240 is not empty or empty. If not empty, then step 520 is entered.
  • the TMA circuit 200 of the present invention checks if the texture data corresponding to the address stored in the last entry (e.g., logical bottom or physical bottom) of the FLFO is present within the cache memory 251. If not, then step 510 is entered.
  • step 530 is entered.
  • the TMA circuit 200 pops out the last FLFO entry from FIFO 240.
  • the cache controller 250 accesses texture data from the cache memory 251 corresponding to the last entry address and supplies the texture data to the filter unit 260.

Abstract

A circuit and method for increasing the processing efficiency of texture map data requests within a 3D subunit of a computer controlled graphics display system. The 3D graphics display subsystem includes a polygon engine, a texture map engine and a pixel pipeline. The texture map engine contains a texture map data access (TDA) circuit having a cache controller with a computer readable cache memory for containing recently used texture maps stored in (u, v) coordinate space. The cache controller is limited in handling only n cache miss operations simultaneously. In one embodiment, n is 1. The TDA circuit also contains a texture map address (TMA) FIFO memory unit for storing texture map addresses associated with texture data requests that hit or missed in the cache memory unit. Since the cache controller handles up to n misses, the texture engine stalls when the (n+1)th unprocessed texture request miss is encountered. Therefore, the TMA FIFO at any time contains at most n miss addresses therein. Processing efficiency is increased when a miss is encountered but the TMA FIFO contains unprocessed hit addresses. At this time, simultaneously with the cache controller fetching the texture data for the missed address, it can also advantageously: (1) supply data from the cache memory for the previously encountered and stored hit addresses; and (2) accept new hit addresses into the TMA FIFO thereby effectively avoiding a texture engine stall. This is quite unlike the prior art systems which process no hit addresses upon a texture miss but rather stall the texture engine.

Description

ENHANCED TEXTURE MAP DATA FETCHING CIRCUIT AND METHOD
FIELD OF THE INVENTION The present invention relates to the field of computer controlled graphics display systems. Specifically, the present invention relates to a system and method for enhancing data throughput in a texture map data retrieval subsystem.
BACKGROUND OF THE INVENTION The computer controlled graphics display systems typically provide data and control signals to graphics hardware units (e.g., "graphics cards") which contain specialized circuits and encoded procedures for processing graphics instructions at high speeds. The graphics instructions are typically stored in a "display list" within computer memory. The instructions define the rendering of several graphic primitives, e.g., individual points, lines, polygons, fills, BIT BLTs (bit block transfers), textures, etc., and graphics commands. Collections of graphics primitives can be used to render a two dimensional image on a display screen of an object that is represented in three dimensional space. Rendering involves translating the above graphics primitives and graphics instructions into raster encoded data that is then loaded into a frame buffer memory for display ("refresh") on the display screen.
Some polygon graphics primitives include specifications of texture data, representative of graphic images, to be displayed within the polygon. Texture mapping refers to techniques for adding surface detail to areas or surfaces of the polygons displayed on the two dimensional display screen. Since the original graphics object is three dimensional, texture mapping often involves maintaining certain perspective attributes with respect to the surface detail added to the primitive. Stored in memory, a typical texture map includes point elements ("texels") which reside in a (u, v) texture coordinate space. A texture image is represented in computer memory as a bitmap or other raster-based encoded format. Further, the display screen includes point elements (pixels) which reside in an (x, y) display coordinate space.
Generally, the process of texture mapping occurs by accessing encoded surface detail points or "texels" from a memory unit that stores the surface detail (e.g., an image) and transferring the surface detail texels to predetermined points of the graphics primitive to be texture mapped. The individual texels of the texture map data are read out of memory and applied within their polygon in particular fashions depending on the placement and perspective of their associated polygon. Thus, color values for pixels in (x, y) display coordinate space are determined based on sampled texture map values. The process of texture mapping operates by applying color or visual attributes of texels of the (u, v) texture map to corresponding pixels of the graphics primitive on the display screen. After texture mapping, a version of the texture image is visible on surfaces of the graphics primitive, with the proper perspective, if any.
However, the process of texture mapping requires a great demand on the memory capacity of the graphics display system because a lot of texture maps are accessed from memory during a typical display screen update cycle. Since the frequency of the screen update cycles is rapid, the individual polygons of the screen (and related texture map data per polygon) need to be accessed and updated at an extremely rapid frequency requiring great data throughput capacities. In view of the above memory demands, high performance graphics hardware units typically contain low access time cache memory units and cache memory controller units for storing and retrieving texture mapped data at high speeds. With texture caches, as a texture-mapped polygon is processed through the graphics unit, an address check is made by the graphics controller as to whether or not the texture map for the polygon is stored in the texture cache. If the requested memory addresses are not present in the texture cache, the cache controller unit of the prior art system stalls while the desired texture data is obtained from external memory. Usually, there is a long latency (stall) from the cache controller unit sending out the external memory request until the texture data is actually fetched from the external memory.
During the stall period, certain portions of the graphics unit wait for the cache controller to replace the least recently used (LRU) set of cache data with newly fetched data from an external source. Meanwhile, many portions of the graphics unit are momentarily prevented from performing useful work either because the source of their data is stalled (e.g., the texture engine is stalled because the requested texels are not present) or because the unit that receives data from them is busy and will not accept their data. In view of pressing demands for ever increasing graphics data throughputs (e.g., for real-time audio/video multi-media applications), the above latency condition is disadvantageous in a high performance graphics display system. Accordingly, the present invention provides a system and method for providing efficient texture map data retrieval in a computer controlled graphics display system. Further, the present invention provides a system as above that provides a texture data cache controller unit that can perform useful texture data processing while waiting for fetched texture data associated with a texture cache miss. These and other advantageous not specifically recited above will become clear within discussions of the present invention to follow.
BRIEF DESCRIPTION OF THE DRAWINGS
Figure 1 is a block diagram of a computer controlled graphics display system in accordance with the present invention.
Figure 2 is a block diagram of the 3D graphics subunit of the computer controlled graphics display system in accordance with the present invention.
Figure 3 is a block diagram of the texture map data access (TDA) circuit of the graphics unit in accordance with the present invention.
Figure 4A and Figure 4B illustrate, respectively, a texture map data request miss address and a texture map data request hit address being placed on the top entry of the texture map address (TMA) FLFO memory unit of the present invention.
Figure 5A and Figure 5B illustrate, respectively, a texture map data request hit address being placed on the top entry of the texture map address FIFO memory unit after the FIFO configurations of Figure 4 A and Figure 4B.
Figure 6A illustrates a first texture engine stall condition whereby an (n+l)m unprocessed miss address is received by the cache controller of the present invention.
Figure 6B illustrates a second texture engine stall condition whereby a miss address reaches the last entry of the texture map address FIFO memory unit but the cache controller has not yet finished its data fetch operation.
Figure 7A is an illustration of a FIFO configuration of the texture map address FIFO memory unit wherein fetched texture data is supplied for miss address that is the oldest address in the texture map address FIFO memory unit.
Figure 7B is an illustration of a FIFO configuration of the texture map address FIFO memory unit as shown in Figure 7 A with an (n+l)m miss address simultaneously entering into the texture map address FIFO memory unit. Figure 7C, Figure 7D, Figure 7E, and Figure 7F illustrate, respectively, texture data stored in the memory cache for a first, second, third, and fourth hit address being supplied simultaneously with the processing of a data fetch operation for a miss address stored in an upper entry of the texture map address FIFO memory unit. Figure 7G illustrates that during the processing of a data fetch operation for the miss address of Figures 7B-7F, subsequently received hit addresses were placed within the upper portion of the texture map address FIFO memory unit.
Figure 8 illustrates steps within a flow chart diagram of the operation of the present invention for placing texture data hit and miss addresses onto the texture map address FIFO memory unit.
Figure 9 illustrates steps within a flow chart diagram of the operation of the present invention for removing texture data hit and miss addresses from the texture map address FIFO memory unit.
SUMMARY OF THE INVENTION
A circuit and method are described for increasing the processing efficiency of texture map data requests within a computer controlled graphics display system. A 3D graphics display subunit is included in the graphics display system and this subunit includes a polygon engine, a texture map engine, and pixel pipeline. The texture map engine contains a texture map data access (TDA) circuit having a cache controller with a computer readable cache memory for containing recently used texture maps stored in (u,v) coordinate space. The cache controller is limited in handling only n cache miss operations simultaneously. In one embodiment, n is 1. The TDA circuit also contains a texture map address FIFO memory unit for storing texture map addresses associated with texture data requests that hit or missed in the cache memory unit. Since the cache controller handles up to n misses, the texture engine stalls when the (n+l)tn unprocessed texture request miss is encountered. Therefore, the TMA FIFO at any time contains at most n miss addresses therein. Processing efficiency is increased when a miss is encountered but the TMA FLFO contains unprocessed hit addresses. At this time, simultaneously with the cache controller fetching the texture data for the missed address, it can also advantageously: (1) supply data from the cache memory for the previously encountered and stored hit addresses; and (2) accept new hit addresses into the TMA FLFO thereby effectively avoiding a texture engine stall. This is quite unlike the prior art systems which process no hit addresses upon a texture miss but rather stall the texture engine.
Specifically, embodiments of the present invention include a computer controlled graphics display system having a host processor for executing instructions and processing data, a bus coupled to the host processor, a display screen for displaying graphics images, and a circuit for retrieving texture map data, the circuit including: main memory (or a local frame buffer) for containing texture map data used for texture mapping operations of graphics primitives; a texture map cache memory for containing recently accessed sub-texture map data of the texture map data of the main memory (or local frame buffer memory); a cache lookup circuit coupled to a cache controller circuit for determining if a received texture map address is a hit address or a miss address with respect to contents of the cache memory; a FLFO memory coupled to the cache lookup circuit for storing a plurality of received texture map addresses; and a cache controller circuit coupled to the main memory, the FLFO memory, and the texture map cache memory, the cache controller circuit (1) for fetching, during a fetch interval, first texture map data from the main memory for storage into the texture map cache memory, the first texture map data corresponding to a texture map miss address stored in a first location of the FIFO memory, and the cache controller circuit,(2) for also providing, within the fetch interval, second texture map data stored within the texture map cache memory, the second texture map data corresponding to a first texture map hit address stored in a second location of the FIFO memory, the first texture map hit address received by the FLFO memory before the texture map miss address.
Embodiments include the above and wherein the cache controller is also for providing, during the fetch interval, third texture map data stored within the texture map cache memory, the third texture map data corresponding to a second texture map hit address stored in a third location of the FLFO memory, the second texture map hit address received by the FLFO memory before the texture map miss address and wherein the FIFO memory is also for receiving and storing texture map hit addresses during the fetch interval.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
In the following detailed description of the present invention, an efficient texture data retrieval method used in conjunction with a texture data cache memory, numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it will be obvious to one skilled in the art that the present invention may be practiced without these specific details or by using alternate elements or processes. In other instances well known processes, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects of the present invention.
NOTATION AND NOMENCLATURE
Some portions of the detailed descriptions which follow are presented in terms of procedures, logic blocks, processing, and other symbolic representations of operations on data bits within a computer memory. These descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. A procedure, logic block, process, etc., is herein, and generally, conceived to be a self-consistent sequence of steps or instructions leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these physical manipulations take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated in a computer system. Herein, these signals are referred to as bits, values, elements, symbols, characters, terms, numbers, or the like with reference to the present invention.
It should be borne in mind, however, that all of these terms are to be interpreted as referencing physical manipulations and quantities and are merely convenient labels and are to be interpreted further in view of terms commonly used in the art. Unless specifically stated otherwise as apparent from the following discussions, it is understood that throughout discussions of the present invention, discussions utilizing terms such as "processing" or "computing" or "calculating" or "determining" or "displaying" or the like, refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data. The data is represented as physical (electronic) quantities within the computer system's registers and memories and is transformed into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
COMPUTER CONTROLLED GRAPHICS DISPLAY SYSTEM
With reference to Figure 1 , a block diagram is shown of a computer controlled graphics display system 112 used in accordance with the present invention. In general, host computer system 112 used by the an embodiment of the present invention comprises a bus 100 for communicating information, one or more a host processors 101 coupled with the bus 100 for processing information and instructions, a computer readable volatile memory unit 102 (e.g. random access memory unit) coupled with the bus 100 for storing information and instructions for the host processor 101, a computer readable non-volatile memory unit 103 (e.g., read only memory unit) coupled with the bus 100 for storing static information and instructions for the host processor 101, a computer readable data storage device 104 such as a magnetic or optical disk and disk drive (e.g., hard drive or floppy diskette) coupled with the bus 100 for storing information and instructions, and a display device 105 coupled to the bus 100 for displaying information to the computer user. The display device 105 utilized with the computer system 112 of the present invention can be a liquid crystal device, cathode ray tube, or other display device suitable for creating graphic images and alphanumeric characters recognizable to the user. The host computer system 112 provides data and control signals via bus 100 to a graphics hardware unit or system, e.g., "graphics card" 108. The graphics hardware system 108 contains a 3D graphics subunit 109 for executing a series of display instructions found within a display list stored in computer memory. The display list generally contains instructions regarding the rendering of several graphic primitives, e.g., individual points, lines, polygons, fills, BIT BLTs (bit block transfers), textures, etc. Many of the polygon display instructions include texture data to be displayed within the polygon. Texture data is stored in computer readable (e.g., volatile) memory units of system 112, or local frame buffer 110) in the form of raster based data (e.g., in one form its bit mapped) stored in (u,v) coordinates. The individual components (e.g., "texels") of the texture data are read out of memory and applied within their polygon in particular fashions depending on the placement and perspective of their associated polygon. The process of rendering a polygon with associated texture data is called "texture mapping." Texture mapping requires a great demand on the memory capacity of the computer system 112 because many texture maps are accessed from memory to construct a displayed frame. Since screen updates need to be performed rapidly, polygons need to be updated very rapidly and further texture maps need to be accessed and applied in extremely rapid fashion, increasing memory demands. The graphics hardware system
108, over bus 100", supplies data and control signals to a local frame buffer memory 1 10 which refreshes the display device 105 for rendering images (including graphics images) on display device 105. Components of the graphics hardware system 108 (e.g., the 3D graphics subunit 109) are discussed in more detail below. Figure 2 illustrates a portion of the circuitry of the graphics subunit 109 including a texture engine 10, a polygon engine 12 and a pixel pipeline 16. The texture engine 10 receives polygon vertex data over bus 5 that corresponds to respective polygons to be rendered. The polygon vertex data includes data points for each vertex of the polygon. With respect to triangle polygons, each of the three vertexes contains: its own position coordinate values (x, y, z); its own color values (red, green, blue); its own texture map coordinate values (u, v), its own perspective value (w), and other required values including an identification of the texture map data for the polygon, if any. The texture engine 10 is responsible for retrieving the texture map data for the polygon and mapping the texels of the texture data onto the pixels of the polygon. Once the texture engine 10 is given the texture map coordinates (u,v) for each vertex of a triangle, it can go to the texture cache controller 250 and access the matching texels for placement into the triangle. During this process, the texture engine 10 maintains the three dimensional perspective of the surface of the polygon. Texture map data retrieval (TDA) circuit 200 performs texture map data retrieval processes in accordance with the present invention. Aside from the texture retrieval system of texture engine 10, a number of well known procedures and circuits can be used to maintain the perspective and perform the texture mapping operations implemented within texture engine 10. Texture map pixel data is supplied from the texture engine 10 to the pixel pipeline 16 over bus 14a.
The polygon engine 12 of Figure 2 receives the polygon data over bus 5 and performs well known polygon rendering functions with respect to the position, color, and perspective of the polygon primitive. Essentially, polygon engine uses interpolation to compute the pixel positions and colors of the pixels within the polygon primitive based on the polygon vertex data. Pixel information resulting from the polygon engine 12 is forwarded to the pixel pipeline over bus 14b.
The pixel pipeline 16 blends the texture data (texels) from the texture engine 10 with the pixel data from the polygon engine 12 to form a composite polygon image. The data (pixels) of the composite image are forwarded in raster encoded format over bus 18, stored in a raster encoded frame buffer (situated within graphics subunit 109, but not shown in Figure 2), and eventually displayed on display screen 105 (Figure 1). The above operation is performed individually for each received polygon primitive. The pixel pipeline 16, in one embodiment, contains a latency of approximately 5-7 clock cycles, depending on programmable image features.
TEXTURE MAP DATA RETRIEVAL (TDA) CIRCUIT 200
OF THE PRESENT INVENTION
Figure 3 illustrates the components of the texture map data access (TDA) circuit 200 in accordance with the present invention. TDA circuit 200 provides an efficient mechanism for processing texture map data requests (in the form of addresses) whereby useful texture map data can be supplied from a cache memory 251 to the filter 260 simultaneously during a fetch interval wherein other texture data is being fetched from main memory 102 or from a local frame buffer 110 (Figure 1). This operation is unlike the prior art where during a fetch interval, the texture engine stalls and no useful texture map data is provided until the fetch completes. Further, during the fetch interval, later received hit addresses can be accepted by the TDA circuit 200 thereby preventing the stall out of circuits that supply addresses to the TDA circuit 200.
As discussed above, the polygon data received by the texture engine 10 contains an indication (e.g., a pointer) of the texture map (e.g., a base address of the texture map) used by the currently processed polygon. The TDA circuit 200 contains an address control unit 210 which receives, over bus 205, the above referenced texture map base address. The texture map base address information is also forwarded over bus 207 to an address generator unit 220. The address generator unit 220 is controlled by control signals of line 237 that originate from the address control unit 210. Address generator unit 220 also receives (u,v) texture coordinates corresponding to the current polygon primitive. The address generator unit 220 generates the (u,v) texel addresses and level of detail (LOD) information for each corresponding pixel of a polygon primitive. These newly "received" texture map addresses ("texel addresses") are individually forwarded from unit 220 to a cache lookup unit 230.
A number of different texel address generation mechanisms can be used by address generator unit 220 within the present invention. In one embodiment, the address generator unit 220 inputs the u_main, v_main, du_main, du_ortho, dv_main, and dv_ortho to compute the required texel adresses. In this embodiment, these terms are defined in copending patent application serial number , filed , and entitled "Non-Homogenous Second Order
Perspective Texture Mapping Coordinates Using Linear Interpolation," by Vaswani et al., assigned to the assignee of the present invention with attorney docket number CRUS-096-050. The cache lookup unit 230 of Figure 3 uses control and address bus 235 to examine the contents of a texture map cache memory circuit ("cache memory") 251 to determine if each received texture map address corresponds to texture map data stored in the cache memory 251. The cache memory 251 is situated within a cache controller circuit 250 and in one embodiment is a 16-bit fully-associated cache of a predetermined size (e.g., Ik bytes). In one embodiment, the cache memory 251 is divided into 16 sets (64 bytes each). If a received texture map address corresponds to texture map data stored in the cache circuit 251, then it is a texture map hit address ("hit address"). If a received texture map address does not correspond to texture map data stored in the cache circuit 251, then it is a texture map miss address ("miss address"). In the latter case, a fetch interval is required for the cache controller circuit 250 causing it to fetch the required texture map data from main memory 102 or local frame buffer 110 for storage into the cache memory 251. In one embodiment, a block of texture data representing portions of several screen lines (e.g., 64 bytes corresponding to a single set size) is retrieved for each fetch interval. The fetched texture map data is fetched in 64 byte blocks (corresponding to a set) each comprising an 8x8 matrix of texture map data, in one embodiment of the present invention. The cache controller 250 contains circuitry for performing n simultaneous fetch operations wherein texture data is fetched from main memory 102 (or local frame buffer 110) and stored within the cache memory 251. In one embodiment, n=l. When more than n fetch operations are required, the additional fetch operations are delayed until the n fetch operations are complete. In the embodiment described above, when a second fetch operation is required and a pending fetch operation is not yet complete, the second fetch operation is delayed until the completion of the pending fetch operation. The interval of a fetch operation is called the fetch interval. The fetch operation uses bus 202 which is communicatively interfaced with bus 100 of system 112 to receive texture map data from main memory 102 or bus 100" from local frame buffer 110. A number of well known circuits and techniques can be used by the cache controller circuit 250 within the scope of the present invention for implementing the texture map data fetch operation.
In one embodiment of the present invention, in parallel (e.g., simultaneous) with a fetch operation, cache controller circuit 250 and cache memory 251 can be used to supply texture map data, stored within cache memory 251, that corresponds to texture map hit addresses that were received before the commencement of the fetch interval. A first-in-first-out (FLFO) memory circuit 240 of Figure 3 is coupled to input received texture map addresses from cache lookup unit 230. FLFO memory 240 (also called TMA FIFO) contains a number of entries, (1) - (m), for storing hit or miss addresses. Texture map addresses that are at the bottom entry (physical or logical) of the FLFO memory unit 240 are processed such that their corresponding texture map data is retrieved from the cache memory unit 251 and supplied over bus 253 if the texel is available in the texture cache 251. The cache memory unit 251 is addressed by the bottom entry of FIFO memory 240 over address and control bus 238. If the texture map data corresponding to the bottom entry address is not yet available in the cache memory 251 , FLFO memory 240 is stalled until the data becomes available. When the corresponding texture map data is supplied from cache memory 251 , the bottom entry texture map address is removed from FLFO memory 240.
Filter unit 260 receives texture map data over bus 253 from the cache controller circuit 250 and performs a number of well known data filtering operations, including performing linear filtering, bi-linear filtering, and tri-linear filtering. The filter unit 260 then outputs the texture map data for the polygon to an optional output FIFO 270 which is coupled to external bus 273. Bus 273 is coupled to bus 14a (Figure 2) to the pixel pipeline circuit 16.
The operation of the FLFO memory 240 of TDA circuit 200 of the present invention is now explained. Figure 4A illustrates a FIFO configuration wherein the FIFO memory 240 is empty and a miss address is pushed into the top entry (1). Upon this configuration, cache controller 250 commences a fetch interval to obtain the texture map data for the miss address. Figure 4B illustrates a FLFO configuration wherein the FIFO memory 240 is empty and a hit address is pushed into the top entry (1).
Figure 5A illustrates the FIFO configuration of Figure 4A, but a later received hit address is next pushed onto the top entry (1) and the miss address is then copied into entry (2). It is appreciated that in this configuration, the fetch interval for the miss address of entry (2) is not yet complete. Figure 5A illustrates that multiple later received hit addresses can be pushed onto the
FLFO memory 240 during a fetch interval. This later received hit address, although not yet processed by the cache memory unit 251 , remains in the FIFO memory 240. Therefore, the present invention advantageously allows the TDA circuit 200 to accept new texture map addresses during a fetch interval thus preventing the stall out of circuitry that supplies these texture map addresses during the fetch interval. Figure 5B illustrates the FIFO configuration of Figure 4B, but a later received hit address is next pushed onto the top entry (1) and the first hit address is popped out of the FIFO 240.
Figure 6A and Figure 6B illustrate two FLFO configurations that cause a FLFO memory 240 stall condition. During a stall, the FLFO memory 240 and the TDA circuit 200 does not accept any more texture map addresses. Figure 6A illustrates a FLFO stall configuration where an early received miss address is near the bottom of the FIFO memory 240 in entry (m'), a number of hit addresses were received and stored in entries (1) through (m'-l), and then another miss address 310a is received. The miss address in entry (m') is still pending, e.g., its fetch interval is still open. In the embodiment wherein n=l , this condition causes a FLFO stall because the cache controller circuit 250 can only process one fetch interval at a time. Therefore, the FIFO memory 240 stalls and does not accept new texture map addresses until it removes the miss address in entry
( ').
Figure 6B illustrates a second FIFO stall configuration where a pending miss address reaches the bottom entry, (m), of the of the FIFO memory 240 and the remainder of the FLFO memory 240, entries (1) - (m-1), are filled with hit addresses. Although no new texture map address is received, in this configuration the texture map data corresponding to the miss address in entry (m) is not yet available because its fetch interval is not yet complete. In this configuration, the FLFO memory 240 stalls until the texture map data for the miss address in entry (m) becomes available, at which time this miss address is removed from the FLFO memory 240. Figure 7A illustrates that the "bottom" entry of the FIFO memory 240 can be viewed as the actual last physical entry (m) or as a logical bottom entry defined as the entry that contains the oldest texture map address. In one embodiment of the present invention, the last physical entry (m) of the
FLFO memory 240 is used to address the cache memory 251 to retrieve the corresponding texture map data. However, as shown in Figure 7A, the logical bottom entry (nϊ) can also be used in another embodiment of the present invention. In Figure 7A, a miss address is at position (m') being the oldest address within the FIFO memory 240 and other subsequently received hit addresses are stored in entries (1) - (m'-l). At this time, the fetch interval for the miss address at entry (m') completes causing its corresponding texture data to be supplied from cache memory 251. The miss address in entry (m') is then removed from the FIFO memory 240.
Figure 7B-Figure7G illustrate an effective use of the TDA circuit 200 of the present invention for texture map data retrieval wherein texture map data is supplied to the texture engine 10 from the TDA circuit 200 during a fetch interval. Figure 7B illustrates a FLFO configuration where a first miss address is pending and stored in the (m1) entry, several hit addresses are stored in entries (1) through (m'-l) of FLFO memory 240, and then a second miss address 320a is received.
At this moment, t=0, the texture map data for the first miss address becomes available in the cache memory 251, as shown in Figure 7B.
Figure 7C illustrates the FLFO configuration at the next clock cycle, t=l, where the second miss address becomes stored in entry (1) and the hit addresses are each shifted down to occupy entries (2) - (m') of the FIFO memory 240. Also at t=l, the cache controller circuit 250 commences a fetch interval to retrieve the texture map data from main memory 102 for the second miss address. Simultaneous with this fetch interval, the cache controller circuit 250 also accesses the cache memory 251 with the hit address (address 1) located at the bottom entry (m') of the FIFO memory 240 (either physical bottom or logical bottom). The cache controller circuit 250 then supplies the texture map data corresponding to this hit address 1 to the texture filter 260. Therefore, the present invention TDA circuit 200 advantageously allows texture data to be supplied during a texture data fetch interval.
Figure 7D illustrates the FIFO configuration at the next clock cycle, t=2, where the second miss address is stored in entry (2) and the previously received hit addresses are each shifted down to occupy entries (3) - (m') of the FIFO memory 240. Also, a newly received hit address is shifted into entry (1) of FIFO memory 240. Thus, the present invention, during a fetch interval, allows the
FIFO memory 240 to advantageously accept new hit addresses. At t=2, the fetch interval that commenced at t=l is still pending. Simultaneous with this pending fetch interval, the cache controller circuit 250 also accesses the cache memory 251 with the hit address (address2) located at the bottom entry (m') of the FLFO memory 240 (either physical bottom or logical bottom). The cache controller 250 then supplies the texture map data corresponding to this hit address2 to the texture filter 260.
Figure 7E illustrates the FIFO configuration at clock cycle t=3 where the second miss address is stored in entry (3) and the previously received hit addresses are each shifted down to occupy entries (4) - (m') of the FIFO memory 240. The hit address received at t=2 is shifted into entry (2) and a newly received hit address is shifted into entry (1) of FIFO memory 240. At t=3, the fetch interval that commenced at t=l is still pending. Simultaneous with this pending fetch interval, the cache controller circuit 250 also accesses the cache memory 251 with the hit address (address3) located at the bottom entry (m') of the FIFO memory 240 (either physical bottom or logical bottom). The cache controller 250 then supplies the texture map data corresponding to this hit address3 to the texture filter 260.
Figure 7F illustrates the FLFO configuration at clock cycle t=4 where the second miss address is stored in entry (4) and the only previously received hit address that is left is shifted down into entry (m1) of the FLFO memory 240. The hit addresses received at t=3 are shifted to occupy entries (2)-(3) and a newly received hit address is shifted into entry (1) of FLFO memory 240. At t=4, the fetch interval that commenced at t=l is still pending. Simultaneous with this pending fetch interval, the cache controller circuit 250 also accesses the cache memory 251 with the hit address (address4) located at the bottom entry (m') of the FLFO memory 240. The cache controller 250 then supplies the texture map data corresponding to this hit address4 to the texture filter 260. This process can continue for each next clock cycle provided: (1) there are previously received hit addresses, relative to the pending miss address; and (2) the fetch interval is still pending. It is appreciated that with reference to Figure7C through Figure 7F, after each hit address is processed, it is popped of off the FIFO memory 240.
Figure 7G illustrates the FIFO configuration at the completion of the pending fetch interval (t > 4) when the requested texture data for the miss address becomes stored in the cache memory 251. Next, the cache controller circuit 250 accesses the cache memory 251 with the miss address located at the bottom entry (m') of the FIFO memory 240. The cache controller 250 then supplies the texture map data corresponding to this miss address to the texture filter 260. Next, the hit addresses located in entries (1) - (4) are processed during the next four clock cycles as new texture map addresses are received at the top of FLFO memory 240.
During the cycles t= 1 to t=4, the TDA circuit 200 of the present invention advantageously processes hit addresses to supply texture map data during a fetch interval. Also during this interval, new hit addresses are accepted by the FIFO memory unit 240. In order to provide optimum performance gain in this environment, the entry size of the FIFO memory 240 should be roughly equal to the number of clock cycles required to perform a data fetch by the cache controller 250. In one embodiment of the present invention this value is 20 cycles, e.g., (m) should be roughly 20 entries. In one embodiment, each texture map data fetch interval involves the fetching of a block of texture map data, e.g., 64 bytes each corresponding to either four lines of 16 bytes each or eight lines of 8 bytes each. It is appreciated that in operation, TDA circuit 200 is very effective because texture mapping and texture filtering procedures often operate on texture data that are closely situated within a given address space neighborhood. Therefore, texture map address misses are not usually received back- to-back in time, but are rather spaced out between hit addresses. In this expected operational environment, FLFO memory 240 does not stall often as a result of encountering two pending miss addresses thus providing good data throughput of the cache memory 251.
OPERATION OF TDA CIRCUIT 200 OF THE PRESENT INVENTION
Figure 8 illustrates steps of process 400 performed by the TDA circuit 200 of the present invention for placing new texture map addresses (hit/miss) into the entries of FLFO memory 240. At start, it is assumed that FLFO 240 is empty. Process 400 commences at step 410 where it is checked if a new texture map address is received at address generator unit 220, e.g., new (u,v) coordinates and texture map base address. If so, the address generator circuit 220 computes a new texture map address at step 410. At step 420, the cache lookup circuit 230 checks if the new address is a hit address or a miss address. If the address is a hit address, then step 430 is entered where the hit address is pushed into the FIFO 240 and step 410 is re-entered. At step 420, if the address is a miss address then step 440 is entered.
At step 440, TDA circuit 200 pushes the miss address into the FIFO 240. This commences a fetch interval for the pending miss address. At step 450, when another newly received texture map address is received by the TMA circuit 200, then step 455 is entered. At step 455, the cache lookup circuit 230 checks if the new address is a hit address or a miss address. If the new address is a hit address then at step 460 the TMA circuit 200 checks if there is vacancy in the FLFO 240. If there is vacancy, then step 440 is entered where the newly received address is placed into the FIFO
240. At step 460, if there is no vacancy in the FIFO 240, then step 465 is entered where the TMA circuit 200 stalls and waits until the pending texture data fetching interval is complete so that a new
FLFO entry can become available.
At step 455 of Figure 8, if the newly received address is a miss address, then at step 470 the TMA circuit 200 checks if another miss address is already pending within the TMA FLFO 240. If so, then at step 475 the TMA circuit 200 stalls and waits until the pending texture data fetching interval is complete so that the newly received miss address can be supplied to the cache controller
250. At the completion of step 475, step 440 is then entered to push the new miss address into the FLFO 240. At this time, a fetch interval is commenced causing the cache controller circuit 250 to access main memory 102 (or the local frame buffer 1 10) to retrieve the texture map data for the newly stored pending miss address. At step 470, if another miss address is not already pending in FLFO 240, then step 440 is entered to push the new miss address into the FIFO 240. At this time, a fetch interval is commenced causing the cache controller circuit 250 to access main memory 102 (or local frame buffer 110) to retrieve the texture map data for the newly stored pending miss address.
Figure 9 illustrates steps of process 500 performed by the TDA circuit 200 of the present invention for removing texture map addresses from the bottom entry of FIFO memory 240. Process 500 commences at step 510 where TDA circuit 200 checks if FIFO 240 is not empty or empty. If not empty, then step 520 is entered. At step 520, the TMA circuit 200 of the present invention checks if the texture data corresponding to the address stored in the last entry (e.g., logical bottom or physical bottom) of the FLFO is present within the cache memory 251. If not, then step 510 is entered. At step 520, if the texture data for the last entry is stored within the texture cache memory 251, then step 530 is entered. At step 530, the TMA circuit 200 pops out the last FLFO entry from FIFO 240. At step
540, the cache controller 250 accesses texture data from the cache memory 251 corresponding to the last entry address and supplies the texture data to the filter unit 260.
CONCLUSION
The preferred embodiment of the present invention, an efficient texture data retrieval method used in conjunction with a texture data cache memory, is thus described. While the present invention has been described in particular embodiments, it should be appreciated that the present invention should not be construed as limited by such embodiments, but rather construed according to the below claims.

Claims

CLAIMSWhat is claimed is:
1. In a computer controlled graphics display system having a memory unit for containing texture map data used for texture mapping operations of graphics primitives, a circuit for retrieving texture map data, said circuit comprising: a cache memory for containing recently accessed texture map data of said texture map data of said memory unit; a cache lookup circuit coupled to a cache controller circuit for determining if a received texture map address is a hit address or a miss address with respect to contents of said cache memory; a FLFO memory coupled to said cache lookup circuit for storing a plurality of received texture map addresses; and a cache controller circuit coupled to said memory unit, said FLFO memory, and said cache memory, said cache controller circuit ( 1 ) for fetching first texture map data from said memory unit for storage into said cache memory, said first texture map data corresponding to a texture map miss address stored in a first location of said FIFO memory, and said cache controller circuit (2) for simultaneously providing second texture map data stored within said cache memory, said second texture map data corresponding to a texture map hit address stored in a second location of said FIFO memory.
2. A circuit as described in Claim 1 wherein said first location of said FLFO memory contains data received later in time with respect to data contained in said second location.
3. A circuit as described in Claim 1 wherein said cache memory is a 16-set fully- associated cache memory.
4. A circuit as described in Claim 1 wherein said cache memory is Ik bytes in size.
5. A circuit as described in Claim 1 further comprising an address generation circuit coupled to said cache lookup circuit, said address generation circuit for generating said received texture map addresses which respectively correspond to graphics primitives for rendering.
6. A circuit as described in Claim 1 wherein said FLFO memory is also for receiving and storing texture map hit addresses during the interval defined wherein said cache controller circuit is fetching said first texture map data from said memory unit for storage into said cache memory.
7. A circuit as described in Claim 1 further comprising a texture filter circuit coupled to said cache controller circuit, said texture filter circuit for processing texture map data supplied from said cache controller circuit.
8. A computer controlled graphics display system having a host processor for executing instructions and processing data, a bus coupled to said host processor, a display screen for displaying graphics images, and a circuit for retrieving texture map data, said circuit comprising: a memory unit for containing texture map data used for texture mapping operations of graphics primitives; a texture map cache memory for containing recently accessed texture map data of said texture map data of said memory unit; a cache lookup circuit coupled to a cache controller circuit for determining if a received texture map address is a hit address or a miss address with respect to contents of said cache memory; a FLFO memory coupled to said cache lookup circuit for storing a plurality of received texture map addresses; and a cache controller circuit coupled to said memory unit, said FLFO memory, and said texture map cache memory, said cache controller circuit (1) for fetching, during a fetch interval, first texture map data from said memory unit for storage into said texture map cache memory, said first texture map data corresponding to a texture map miss address stored in a first location of said FLFO memory, and said cache controller circuit,(2) for also providing, within said fetch interval, second texture map data stored within said texture map cache memory, said second texture map data corresponding to a first texture map hit address stored in a second location of said FIFO memory, said first texture map hit address received by said FIFO memory before said texture map miss address.
9. A system as described in Claim 8 wherein said cache controller is also for providing, during said fetch interval, third texture map data stored within said texture map cache memory, said third texture map data corresponding to a second texture map hit address stored in a third location of said FIFO memory, said second texture map hit address received by said FLFO memory before said texture map miss address.
10. A system as described in Claim 8 further comprising an address generation circuit coupled to said cache lookup circuit, said address generation circuit for generating said received texture map addresses which respectively correspond to graphics primitives.
1 1. A system as described in Claim 8 wherein said FIFO memory is also for receiving and storing texture map hit addresses during said fetch interval.
12. A system as described in Claim 8 wherein said texture map cache memory is a 16- set fully-associated cache memory.
13. A system as described in Claim 8 wherein said texture map cache memory is Ik bytes in size.
14. A system as described in Claim 8 further comprising a texture filter circuit coupled to said cache controller circuit, said texture filter circuit for processing texture map data supplied by said cache controller circuit.
15. In a computer controlled graphics display system, a method of retrieving texture map data, said method comprising the steps of: a) storing texture map data in a memory unit, said exture map data used for texture mapping operations of graphics primitives; b) storing in a cache memory recently accessed texture map data of said texture map data of said memory unit; c) determining if a particular texture map address is a hit address or a miss address with respect to contents of said cache memory; d) storing a plurality of received texture map addresses in a FIFO memory and removing individual stored texture map addresses from the bottom of said FIFO memory when texture data associated therewith is supplied from said cache memory; and e) during a fetch interval, using a cache controller circuit to fetch first texture map data from said memory unit for storage into said cache memory, said first texture map data corresponding to a texture map miss address stored in a first location of said FLFO memory; and f) within said fetch interval, using said cache controller to provide second texture map data stored within said cache memory, said second texture map data corresponding to a first texture map hit address stored in a second location of said FLFO memory, said first texture map hit address received by said FLFO memory before said texture map miss address.
16. A method as described in Claim 15 further comprising the step of using said cache controller to provide, within said fetch interval, third texture map data stored within said cache memory, said third texture map data corresponding to a second texture map hit address stored in a third location of said FIFO memory, said second texture map hit address received by said FLFO memory before said texture map miss address.
17. A method as described in Claim 15 further comprising the step of receiving and storing texture map hit addresses into said FLFO memory during said fetch interval.
18. A method as described in Claim 15 further comprising the step processing texture data supplied by said cache controller circuit using a texture filter circuit.
18. A method as described in Claim 15 further comprising the step processing texture data supplied by said cache controller circuit using a texture filter circuit.
19. A method as described in Claim 15 further comprising the step of stalling said above step d), provided a texture map miss address is received by said FIFO memory and another texture map miss address is already stored within said FIFO memory.
20. A method as described in Claim 15 further comprising the step of stalling said above step d), provided a texture map miss address reaches the bottom of said FIFO memory and its corresponding texture map data is not yet supplied from said memory unit.
PCT/US1997/023982 1996-12-20 1997-12-19 Enhanced texture map data fetching circuit and method WO1998028714A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP52908498A JP4110239B2 (en) 1996-12-20 1997-12-19 Enhanced texture map data fetching circuit and method
EP97953482A EP0946929A1 (en) 1996-12-20 1997-12-19 Enhanced texture map data fetching circuit and method
CA002275727A CA2275727A1 (en) 1996-12-20 1997-12-19 Enhanced texture map data fetching circuit and method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/770,453 US5831640A (en) 1996-12-20 1996-12-20 Enhanced texture map data fetching circuit and method
US08/770,453 1996-12-20

Publications (1)

Publication Number Publication Date
WO1998028714A1 true WO1998028714A1 (en) 1998-07-02

Family

ID=25088592

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1997/023982 WO1998028714A1 (en) 1996-12-20 1997-12-19 Enhanced texture map data fetching circuit and method

Country Status (7)

Country Link
US (1) US5831640A (en)
EP (1) EP0946929A1 (en)
JP (1) JP4110239B2 (en)
KR (1) KR20000062275A (en)
CA (1) CA2275727A1 (en)
TW (1) TW424219B (en)
WO (1) WO1998028714A1 (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU736560B2 (en) * 1996-08-30 2001-08-02 Sony Corporation Method and apparatus for texture data
WO2013089837A1 (en) * 2011-12-12 2013-06-20 Google Inc. Pre-fetching map tile data along a route
US8849942B1 (en) 2012-07-31 2014-09-30 Google Inc. Application programming interface for prefetching map data
US8886715B1 (en) 2011-11-16 2014-11-11 Google Inc. Dynamically determining a tile budget when pre-fetching data in a client device
US8972529B1 (en) 2011-08-04 2015-03-03 Google Inc. Management of pre-fetched mapping data incorporating user-specified locations
US9063951B1 (en) 2011-11-16 2015-06-23 Google Inc. Pre-fetching map data based on a tile budget
US9197713B2 (en) 2011-12-09 2015-11-24 Google Inc. Method and apparatus for pre-fetching remote resources for subsequent display on a mobile computing device
US9245046B2 (en) 2011-09-26 2016-01-26 Google Inc. Map tile data pre-fetching based on mobile device generated event analysis
US9275374B1 (en) 2011-11-15 2016-03-01 Google Inc. Method and apparatus for pre-fetching place page data based upon analysis of user activities
US9305107B2 (en) 2011-12-08 2016-04-05 Google Inc. Method and apparatus for pre-fetching place page data for subsequent display on a mobile computing device
US9332387B2 (en) 2012-05-02 2016-05-03 Google Inc. Prefetching and caching map data based on mobile network coverage
US9389088B2 (en) 2011-12-12 2016-07-12 Google Inc. Method of pre-fetching map data for rendering and offline routing
US9569463B1 (en) 2011-11-16 2017-02-14 Google Inc. Pre-fetching map data using variable map tile radius

Families Citing this family (83)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6044440A (en) * 1997-07-10 2000-03-28 Intel Corporation System and method to provide high graphics throughput by pipelining segments of a data stream through multiple caches
US6141020A (en) * 1997-11-12 2000-10-31 S3 Incorporated Opposing directional fill calculators in a graphics processor
US6088701A (en) * 1997-11-14 2000-07-11 3Dfx Interactive, Incorporated Command data transport to a graphics processing device from a CPU performing write reordering operations
US6130680A (en) * 1997-12-01 2000-10-10 Intel Corporation Method and apparatus for multi-level demand caching of textures in a graphics display device
US6259460B1 (en) * 1998-03-26 2001-07-10 Silicon Graphics, Inc. Method for efficient handling of texture cache misses by recirculation
US7038692B1 (en) * 1998-04-07 2006-05-02 Nvidia Corporation Method and apparatus for providing a vertex cache
US7136068B1 (en) * 1998-04-07 2006-11-14 Nvidia Corporation Texture cache for a computer graphics accelerator
US6011565A (en) * 1998-04-09 2000-01-04 S3 Incorporated Non-stalled requesting texture cache
US6236413B1 (en) * 1998-08-14 2001-05-22 Silicon Graphics, Inc. Method and system for a RISC graphics pipeline optimized for high clock speeds by using recirculation
US6201547B1 (en) * 1998-10-05 2001-03-13 Ati International Srl Method and apparatus for sequencing texture updates in a video graphics system
US6977649B1 (en) * 1998-11-23 2005-12-20 3Dlabs, Inc. Ltd 3D graphics rendering with selective read suspend
KR100283413B1 (en) * 1998-12-15 2001-04-02 김영환 Texture Mapping System
US6353438B1 (en) * 1999-02-03 2002-03-05 Artx Cache organization—direct mapped cache
US7050063B1 (en) * 1999-02-11 2006-05-23 Intel Corporation 3-D rendering texture caching scheme
US6919895B1 (en) * 1999-03-22 2005-07-19 Nvidia Corporation Texture caching arrangement for a computer graphics accelerator
US6683615B1 (en) 1999-06-09 2004-01-27 3Dlabs Inc., Ltd. Doubly-virtualized texture memory
US6677952B1 (en) 1999-06-09 2004-01-13 3Dlabs Inc., Ltd. Texture download DMA controller synching multiple independently-running rasterizers
US6650333B1 (en) 1999-06-09 2003-11-18 3Dlabs Inc., Ltd. Multi-pool texture memory management
US6587113B1 (en) 1999-06-09 2003-07-01 3Dlabs Inc., Ltd. Texture caching with change of update rules at line end
US7061500B1 (en) 1999-06-09 2006-06-13 3Dlabs Inc., Ltd. Direct-mapped texture caching with concise tags
US7050061B1 (en) 1999-06-09 2006-05-23 3Dlabs Inc., Ltd. Autonomous address translation in graphic subsystem
US6744438B1 (en) * 1999-06-09 2004-06-01 3Dlabs Inc., Ltd. Texture caching with background preloading
US6825848B1 (en) * 1999-09-17 2004-11-30 S3 Graphics Co., Ltd. Synchronized two-level graphics processing cache
US6750872B1 (en) 1999-09-17 2004-06-15 S3 Graphics, Co., Ltd. Dynamic allocation of texture cache memory
US6717577B1 (en) 1999-10-28 2004-04-06 Nintendo Co., Ltd. Vertex cache for 3D computer graphics
US6618048B1 (en) 1999-10-28 2003-09-09 Nintendo Co., Ltd. 3D graphics rendering system for performing Z value clamping in near-Z range to maximize scene resolution of visually important Z components
US6593931B1 (en) * 1999-12-01 2003-07-15 Intel Corporation Method and apparatus for improving system memory bandwidth utilization during graphics translational lookaside buffer cache miss fetch cycles
US6473834B1 (en) 1999-12-22 2002-10-29 Unisys Method and apparatus for prevent stalling of cache reads during return of multiple data words
US6415357B1 (en) 1999-12-23 2002-07-02 Unisys Corporation Caching method and apparatus
US6433789B1 (en) 2000-02-18 2002-08-13 Neomagic Corp. Steaming prefetching texture cache for level of detail maps in a 3D-graphics engine
US6891533B1 (en) * 2000-04-11 2005-05-10 Hewlett-Packard Development Company, L.P. Compositing separately-generated three-dimensional images
US7710425B1 (en) 2000-06-09 2010-05-04 3Dlabs Inc. Ltd. Graphic memory management with invisible hardware-managed page faulting
US6636214B1 (en) 2000-08-23 2003-10-21 Nintendo Co., Ltd. Method and apparatus for dynamically reconfiguring the order of hidden surface processing based on rendering mode
US6700586B1 (en) 2000-08-23 2004-03-02 Nintendo Co., Ltd. Low cost graphics with stitching processing hardware support for skeletal animation
US7538772B1 (en) 2000-08-23 2009-05-26 Nintendo Co., Ltd. Graphics processing system with enhanced memory controller
US6707458B1 (en) 2000-08-23 2004-03-16 Nintendo Co., Ltd. Method and apparatus for texture tiling in a graphics system
US6811489B1 (en) 2000-08-23 2004-11-02 Nintendo Co., Ltd. Controller interface for a graphics system
US7576748B2 (en) 2000-11-28 2009-08-18 Nintendo Co. Ltd. Graphics system with embedded frame butter having reconfigurable pixel formats
US7196710B1 (en) 2000-08-23 2007-03-27 Nintendo Co., Ltd. Method and apparatus for buffering graphics data in a graphics system
US6999080B1 (en) * 2000-11-30 2006-02-14 Microsoft Corporation System, method, and computer program product for general environment mapping
US6848025B2 (en) * 2001-10-26 2005-01-25 Elan Research, Inc. Method and system for programmable replacement mechanism for caching devices
US7825935B1 (en) * 2001-11-30 2010-11-02 Nvidia Corporation System, method and computer program product for using textures as instructions for graphics processing
US8463947B2 (en) * 2002-08-28 2013-06-11 Tellabs Operations, Inc. Method of finding rings for optimal routing of digital information
US6933945B2 (en) * 2003-03-31 2005-08-23 Sun Microsystems, Inc. Design for a non-blocking cache for texture mapping
US9081681B1 (en) 2003-12-19 2015-07-14 Nvidia Corporation Method and system for implementing compressed normal maps
KR20070008707A (en) * 2004-05-03 2007-01-17 코닌클리즈케 필립스 일렉트로닉스 엔.브이. Graphics pipeline for rendering graphics
US8736628B1 (en) 2004-05-14 2014-05-27 Nvidia Corporation Single thread graphics processing system and method
US8743142B1 (en) * 2004-05-14 2014-06-03 Nvidia Corporation Unified data fetch graphics processing system and method
US20060007234A1 (en) * 2004-05-14 2006-01-12 Hutchins Edward A Coincident graphics pixel scoreboard tracking system and method
US8860722B2 (en) * 2004-05-14 2014-10-14 Nvidia Corporation Early Z scoreboard tracking system and method
US8736620B2 (en) * 2004-05-14 2014-05-27 Nvidia Corporation Kill bit graphics processing system and method
US8711155B2 (en) * 2004-05-14 2014-04-29 Nvidia Corporation Early kill removal graphics processing system and method
US8687010B1 (en) 2004-05-14 2014-04-01 Nvidia Corporation Arbitrary size texture palettes for use in graphics systems
US6972769B1 (en) * 2004-09-02 2005-12-06 Nvidia Corporation Vertex texture cache returning hits out of order
US8078656B1 (en) 2004-11-16 2011-12-13 Nvidia Corporation Data decompression with extra precision
US7961195B1 (en) 2004-11-16 2011-06-14 Nvidia Corporation Two component texture map compression
US7928988B1 (en) 2004-11-19 2011-04-19 Nvidia Corporation Method and system for texture block swapping memory management
US7916149B1 (en) 2005-01-04 2011-03-29 Nvidia Corporation Block linear memory ordering of texture data
US8766995B2 (en) * 2006-04-26 2014-07-01 Qualcomm Incorporated Graphics system with configurable caches
US20070268289A1 (en) * 2006-05-16 2007-11-22 Chun Yu Graphics system with dynamic reposition of depth engine
US8884972B2 (en) 2006-05-25 2014-11-11 Qualcomm Incorporated Graphics processor with arithmetic and elementary function units
US8869147B2 (en) * 2006-05-31 2014-10-21 Qualcomm Incorporated Multi-threaded processor with deferred thread output control
US8644643B2 (en) 2006-06-14 2014-02-04 Qualcomm Incorporated Convolution filtering in a graphics processor
US7626588B1 (en) 2006-06-16 2009-12-01 Nvidia Corporation Prescient cache management
US7616209B1 (en) * 2006-06-16 2009-11-10 Nvidia Corporation Prescient cache management
US8766996B2 (en) * 2006-06-21 2014-07-01 Qualcomm Incorporated Unified virtual addressed register file
US8594441B1 (en) 2006-09-12 2013-11-26 Nvidia Corporation Compressing image-based data using luminance
US7928990B2 (en) * 2006-09-27 2011-04-19 Qualcomm Incorporated Graphics processing unit with unified vertex cache and shader register file
US8537168B1 (en) 2006-11-02 2013-09-17 Nvidia Corporation Method and system for deferred coverage mask generation in a raster stage
KR100882842B1 (en) * 2007-02-26 2009-02-17 삼성전자주식회사 Apparatus to use a fifo as a post-vertex cache and method thereof
US8724895B2 (en) * 2007-07-23 2014-05-13 Nvidia Corporation Techniques for reducing color artifacts in digital images
US9183607B1 (en) 2007-08-15 2015-11-10 Nvidia Corporation Scoreboard cache coherence in a graphics pipeline
US8373718B2 (en) * 2008-12-10 2013-02-12 Nvidia Corporation Method and system for color enhancement with color volume adjustment and variable shift along luminance axis
US8610732B2 (en) * 2008-12-11 2013-12-17 Nvidia Corporation System and method for video memory usage for general system application
US9411595B2 (en) 2012-05-31 2016-08-09 Nvidia Corporation Multi-threaded transactional memory coherence
KR101983833B1 (en) 2012-06-26 2019-09-04 삼성전자주식회사 Method and apparatus for providing shared caches
US9824009B2 (en) 2012-12-21 2017-11-21 Nvidia Corporation Information coherency maintenance systems and methods
US10102142B2 (en) 2012-12-26 2018-10-16 Nvidia Corporation Virtual address based memory reordering
US9230305B2 (en) 2012-12-31 2016-01-05 Nvidia Corporation Summed area computation using ripmap of partial sums
US9569385B2 (en) 2013-09-09 2017-02-14 Nvidia Corporation Memory transaction ordering
KR102147356B1 (en) 2013-09-30 2020-08-24 삼성전자 주식회사 Cache memory system and operating method for the same
CN104981797B (en) * 2014-02-07 2018-05-15 Sk科技有限公司 Cloud stream service system and the method and apparatus that cloud stream service is provided
US20150228106A1 (en) * 2014-02-13 2015-08-13 Vixs Systems Inc. Low latency video texture mapping via tight integration of codec engine with 3d graphics engine

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1995024682A1 (en) * 1994-03-07 1995-09-14 Silicon Graphics, Inc. Integrating texture memory and interpolation logic
EP0747826A2 (en) * 1995-06-06 1996-12-11 Hewlett-Packard Company Cache system with simultaneous tag comparison

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5469535A (en) * 1992-05-04 1995-11-21 Midway Manufacturing Company Three-dimensional, texture mapping display system
US5606650A (en) * 1993-04-22 1997-02-25 Apple Computer, Inc. Method and apparatus for storage and retrieval of a texture map in a graphics display system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1995024682A1 (en) * 1994-03-07 1995-09-14 Silicon Graphics, Inc. Integrating texture memory and interpolation logic
EP0747826A2 (en) * 1995-06-06 1996-12-11 Hewlett-Packard Company Cache system with simultaneous tag comparison

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
IKEDO AND MA: "PIXEL CACHE ARCHITECTURE WITH FIFO IMPLEMENTED WITHIN AN ASIC", PROCEEDINGS NINTH ANNUAL IEEE INTERNATIONAL ASIC CONFERENCE AND EXHIBIT, 23 September 1996 (1996-09-23) - 27 September 1996 (1996-09-27), NEW-YORK, NY, US, pages 19 - 22, XP002061330 *

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU736560B2 (en) * 1996-08-30 2001-08-02 Sony Corporation Method and apparatus for texture data
US8972529B1 (en) 2011-08-04 2015-03-03 Google Inc. Management of pre-fetched mapping data incorporating user-specified locations
US9245046B2 (en) 2011-09-26 2016-01-26 Google Inc. Map tile data pre-fetching based on mobile device generated event analysis
US9275374B1 (en) 2011-11-15 2016-03-01 Google Inc. Method and apparatus for pre-fetching place page data based upon analysis of user activities
US9063951B1 (en) 2011-11-16 2015-06-23 Google Inc. Pre-fetching map data based on a tile budget
US8886715B1 (en) 2011-11-16 2014-11-11 Google Inc. Dynamically determining a tile budget when pre-fetching data in a client device
US9569463B1 (en) 2011-11-16 2017-02-14 Google Inc. Pre-fetching map data using variable map tile radius
US9307045B2 (en) 2011-11-16 2016-04-05 Google Inc. Dynamically determining a tile budget when pre-fetching data in a client device
US9813521B2 (en) 2011-12-08 2017-11-07 Google Inc. Method and apparatus for pre-fetching place page data for subsequent display on a mobile computing device
US9305107B2 (en) 2011-12-08 2016-04-05 Google Inc. Method and apparatus for pre-fetching place page data for subsequent display on a mobile computing device
US9491255B2 (en) 2011-12-09 2016-11-08 Google Inc. Method and apparatus for pre-fetching remote resources for subsequent display on a mobile computing device
US9197713B2 (en) 2011-12-09 2015-11-24 Google Inc. Method and apparatus for pre-fetching remote resources for subsequent display on a mobile computing device
US9111397B2 (en) 2011-12-12 2015-08-18 Google Inc. Pre-fetching map tile data along a route
US9389088B2 (en) 2011-12-12 2016-07-12 Google Inc. Method of pre-fetching map data for rendering and offline routing
US9563976B2 (en) 2011-12-12 2017-02-07 Google Inc. Pre-fetching map tile data along a route
WO2013089837A1 (en) * 2011-12-12 2013-06-20 Google Inc. Pre-fetching map tile data along a route
US9332387B2 (en) 2012-05-02 2016-05-03 Google Inc. Prefetching and caching map data based on mobile network coverage
US8849942B1 (en) 2012-07-31 2014-09-30 Google Inc. Application programming interface for prefetching map data

Also Published As

Publication number Publication date
JP2001507152A (en) 2001-05-29
EP0946929A1 (en) 1999-10-06
US5831640A (en) 1998-11-03
CA2275727A1 (en) 1998-07-02
TW424219B (en) 2001-03-01
KR20000062275A (en) 2000-10-25
JP4110239B2 (en) 2008-07-02

Similar Documents

Publication Publication Date Title
US5831640A (en) Enhanced texture map data fetching circuit and method
KR100300972B1 (en) Texture mapping system and texture cache access method
US7164426B1 (en) Method and apparatus for generating texture
US6734867B1 (en) Cache invalidation method and apparatus for a graphics processing system
EP0998709B1 (en) Non-stalled requesting texture cache system and method
US6903737B2 (en) Method and apparatus for implementing spread memory layout
EP1994506B1 (en) Texture unit for multi processor environment
US6629188B1 (en) Circuit and method for prefetching data for a texture cache
US6058438A (en) Method and apparatus for performing high speed data transfers between a host memory and a geometry accelerator of a graphics machine
EP0883065B1 (en) Non-blocking pipelined cache
US20070052713A1 (en) Systems and methods for storing and fetching texture data using bank interleaving
US6587113B1 (en) Texture caching with change of update rules at line end
EP1721298A2 (en) Embedded system with 3d graphics core and local pixel buffer
US6683615B1 (en) Doubly-virtualized texture memory
US6778179B2 (en) External dirty tag bits for 3D-RAM SRAM
US6604175B2 (en) Data cache and method of storing data by assigning each independently cached area in the cache to store data associated with one item type
US7791612B2 (en) Fully associative texture cache having content addressable memory and method for use thereof
US7710425B1 (en) Graphic memory management with invisible hardware-managed page faulting
US6590579B1 (en) System for low miss rate replacement of texture cache lines
WO1998028713A9 (en) Enhanced methods and systems for caching and pipelining of graphics texture data
WO1998028713A1 (en) Enhanced methods and systems for caching and pipelining of graphics texture data
JPH0877367A (en) Image processor and image data processor using it

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): CA CN IL JP KR SG

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
ENP Entry into the national phase

Ref document number: 2275727

Country of ref document: CA

Ref country code: JP

Ref document number: 1998 529084

Kind code of ref document: A

Format of ref document f/p: F

Ref country code: CA

Ref document number: 2275727

Kind code of ref document: A

Format of ref document f/p: F

WWE Wipo information: entry into national phase

Ref document number: 1019997005655

Country of ref document: KR

WWE Wipo information: entry into national phase

Ref document number: 1997953482

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1997953482

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1019997005655

Country of ref document: KR

WWW Wipo information: withdrawn in national office

Ref document number: 1997953482

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1019997005655

Country of ref document: KR