WO1998031004A1 - Pixel reordering for improved texture mapping - Google Patents

Pixel reordering for improved texture mapping Download PDF

Info

Publication number
WO1998031004A1
WO1998031004A1 PCT/US1997/024264 US9724264W WO9831004A1 WO 1998031004 A1 WO1998031004 A1 WO 1998031004A1 US 9724264 W US9724264 W US 9724264W WO 9831004 A1 WO9831004 A1 WO 9831004A1
Authority
WO
WIPO (PCT)
Prior art keywords
pixel
memory
heap
module
references
Prior art date
Application number
PCT/US1997/024264
Other languages
French (fr)
Inventor
Michael B. Cox
Dinyar B. Lahewala
Dong-Ying Kuo
Original Assignee
S3 Incorporated
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by S3 Incorporated filed Critical S3 Incorporated
Priority to DE69736611T priority Critical patent/DE69736611T2/en
Priority to JP53095398A priority patent/JP2001507840A/en
Priority to CA002275237A priority patent/CA2275237C/en
Priority to EP97954326A priority patent/EP1016068B1/en
Priority to AU58127/98A priority patent/AU5812798A/en
Publication of WO1998031004A1 publication Critical patent/WO1998031004A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects

Definitions

  • the present invention relates generally to texture mapping in graphics systems, and more particularly to a system and method of reordering pixels to improve bandwidth and performance.
  • w Texture mapping is the process of mapping an image onto a surface in a three- dimensional graphics system. This technique is well-known in the art, and is described, for example, in J. Foley et al., Computer Graphics: Principles and Practice, 2d. ed., Addison-Wesley, 1990, at 741-44.
  • Fig. 1 there is shown an example of texture mapping
  • Texture map 101 is typically described in a rectangular coordinate scheme designated (u, v), and is ordinarily stored in some area of conventional memory, such as, for example, a conventional page-mode dynamic random-access memory (DRAM) or other paged memory.
  • DRAM page-mode dynamic random-access memory
  • Surface 104 in three-dimensional space has its own coordinate system (s, t).
  • surface 104 may be a primitive such as a polygon; many such polygons may be defined in three-space to form a three-
  • Each such polygon would then have its own coordinate system (s, t) similar to the surface 104 indicated in Fig. 1.
  • surface 104 is in turn mapped onto a two-dimensional display grid 103 stored in the frame buffer for display by the computer system. The mapping of surface 104 onto display
  • Display grid 103 is accomplished by matrix transforms that are well-known in the art.
  • Display grid 103 has coordinate system (x, y) and is typically implemented in an area of memory reserved for video display, such as video random-access memory (video RAM) e.g. VRAM or synchronous graphics random-access memory (SGRAM).
  • Video RAM video random-access memory
  • SGRAM synchronous graphics random-access memory
  • Display grid 103 contains individual elements known as pixels, represented by distinct memory locations in video RAM.
  • Each pixel in some region of display grid 103 maps onto a point on surface 104 and in turn to a point in texture map 101.
  • point A X y of display grid 103 maps onto point A s t in the coordinates of surface 102 and to point A uv in texture map 101, or a group of points forming a region in texture map 101.
  • Each of the mappings among display grid 103, surface 102, and texture map 101 may be point-to-point, point-to-region, region-to-point, or region-to-region.
  • the mapping from display grid 103 to surface 104 and in turn to texture map 101 also generates a value d representing the level of detail for the particular texel.
  • d is a measure of the perceived distance of the point in the texture map, as determined by a z-value for the point in the frame buffer. Points that are perceived to be farther away have a lower level of detail representing decreased resolution.
  • d is used to implement a multum in parvo map (MIP map) scheme wherein several texels may be averaged and mapped onto one pixel of the frame buffer. The higher the value of d, the lower the level of detail, and the more pixel-averaging is performed. In the limit, the entire texture map may be reduced to one pixel in the frame buffer.
  • MIP map multum in parvo map
  • rasterization engines draw the image into the frame buffer by the known technique of scan conversion of primitives such as polygons and lines (see, for example, Foley et al.).
  • Scan conversion takes as its input primitives defined in terms of vertices and orientations, and provides as its output a series of pixels to be drawn on the screen.
  • a rasterization engine performs the necessary mapping calculations to determine which texel of texture map 101 corresponds to the pixel.
  • the rasterization engine then issues whatever memory references are required, such as texture fetch, z-fetch, z-writeback, color fetch, color write-back, and the like) to retrieve texel information for writing to the pixel begin processed.
  • memory references are issued in the order generated by the scan conversion.
  • such memory references are stored and managed according to a first-in first-out (FIFO) scheme using a FIFO queue.
  • FIFO first-in first-out
  • a system and method of reordering memory references required to process independent pixels in a texture mapping system in order to achieve improved memory locality and hence improved graphics performance using conventional page-mode memory components Since the present invention does not require any changes to be made to the underlying graphics algorithm, memory locality for many such algorithms can be improved at once.
  • Various types of reads and writes may be performed in connection with rendering by a graphics engine. These include, for example, reading and writing for texture retrieval, retrieval of data from the frame buffer (such as color values, z- values, and the like), color write-back (final color write to frame buffer after resolving all applicable textures), and the like.
  • the present invention reorders the reads and writes associated with pixels when performing such processing so that references that require the currently open page are, in general, processed before references that require opening another page, in order to improve memory locality and minimize page- switching.
  • the technique described herein allows the graphics rasterization engine to achieve better bandwidth using conventional page-mode memory components and without require extreme memory interleaving.
  • the present invention employs a Pixel Priority Heap (PPH) to reorder pixels.
  • PPH Pixel Priority Heap
  • the PPH reorders addresses into texture memory to perform more reads from the same page before that page is closed and another opened.
  • the PPH is employed to reorder reads and writes from/ to other clients as well.
  • PPH may be used to reorder references to destination addresses in the frame buffer to improve locality; this is advantageous when PPH reordering of texture memory references results in changes to the order of frame buffer writes, so that a PPH is beneficial to frame buffer addresses as well.
  • Each pixel generated by scan conversion is written, one per cycle, to the PPH. If the PPH is full, the pixel is not written, and the system waits until an entry becomes available in the PPH before writing to it.
  • the system attempts to select one pixel from the PPH and write it to a memory controller first-in first-out (FIFO) queue to be textured. If there are no pixels in the PPH, or if the memory controller FIFO queue is full, no pixel is written to the memory controller FIFO queue.
  • the system attempts to find a pixel whose PPH address contains a bank and row address (RAS) matching the open page on the last pixel read or written. If such a match is found, the access can be performed relatively quickly (e.g. in one cycle) on some memory systems.
  • RAS bank and row address
  • the system attempts to find a pixel whose PPH address matches any open page, thus permitting intermediate-speed (e.g. three-cycle) access on some memory systems. If no such match is found, the system attempts to find any valid pixel in the PPH, regardless of bank or page address. On some memory systems, such an access is relatively slow (e.g. nine cycles for reads or six cycles for writes). In one embodiment, all such attempts to find matches are performed concurrently for improved performance.
  • the selected pixel is written to the memory controller FIFO queue and deleted from the PPH.
  • a memory controller issues the read or write corresponding to the pixel's PPH address. If the operation is a read, the data may then be returned from memory for subsequent processing depending on the application.
  • the PPH can implement conventional first-in first-out (FIFO) functionality. Such implementation may be useful, for example, when using the PPH in connection with z-buff ering.
  • z-buff er is a technique whereby each pixel written to the frame buffer has an associated stored z-value in a z-buffer.
  • Each incoming pixel write has a z-value that is compared with the stored z-value for the appropriate pixel, and is only written if the incoming pixel's z-value is less than the stored z-value.
  • the stored z-value is updated with the incoming pixel's z-value. See J. Foley et al.
  • the PPH also allows reads or writes to continue on an open page on a given bank while precharge is done on a page on an alternate bank.
  • the present invention operates in an essentially stateless manner. No decisions are explicitly based on previous decisions. Such essentially stateless operation is advantageous because it facilitates the use of simpler algorithms and simpler hardware.
  • the present invention emphasizes correct rather than optimal operation by making relatively simple decisions that may not be optimal but that are correct in the sense that they result in performance that is preferably no worse than a conventional FIFO technique.
  • Fig. 1 is a diagram showing texture mapping according to the prior art.
  • Fig. 2 is a block diagram showing overall architecture of an embodiment of the present invention.
  • Fig. 3 is a block diagram showing detailed architecture of an embodiment of the present invention.
  • Fig. 4 is a flowchart of the method of selecting pixels from the pixel priority heap according to the present invention.
  • Fig. 5 is a state diagram showing page update operations.
  • Fig. 6 is a flowchart of a method of selecting a page to be precharged according to one embodiment of the present invention.
  • Fig. 7 is a flowchart of a method of reading from the PPH according to one embodiment of the present invention.
  • Fig. 8 is a block diagram of an implementation of a heap module according to one embodiment of the present invention.
  • N bit-length of index i.e. address of chip and bank
  • System 200 contains the following elements.
  • Primitives 208 are stored in conventional manner and include, for example, definitions of polygons and lines.
  • Scan converter 201 converts primitives 208 into pixels in a conventional manner. Each pixel is defined in terms of values for u, v, and d, which represent u and v coordinates in texture space, along with a distance value (also known as a compression value) for determining the level of detail for the pixel.
  • the u, v, and d values for pixels are written to Pixel Priority Heap (PPH) 202, at a rate of one pixel per cycle.
  • PPH Pixel Priority Heap
  • u, v and d values are mapped onto memory references represented as bank, row and column, and these memory references are written to PPH 202. In yet another embodiment, some combination of these two representations is used, depending on the operation being performed, as described below.
  • PPH 202 is defined on one channel. If two or more channels are employed (e.g. left and right channels), it is preferable to have a PPH for each channel. This serves to queue pixels independently for each channel and can result in higher aggregate hit-rate because the locality improvement process is being effected for each channel separately.
  • a selector is employed to check address bits of pixels to determine the proper PPH for entry of the pixels. For example, in one embodiment the selector determines the proper PPH based on the last bit of the pixel address.
  • a 1MB chip (256K x 32) is provided for each channel. Banks are numbered from 1 to 2N-1, where there are 2N-1 banks (0 is reserved as a null value).
  • PPH 202 contains 2M-1 entries, each corresponding to one pixel to be textured. In one embodiment, two banks are provided for each chip. Three types of pixels are defined: 1) a pixel to be textured (a tpixe ⁇ ); 2) a pixel to be z-buffered (a z ⁇ ixel); and 3) a pixel to be written to (or read from) the frame buffer, or a texel to be written to texture memory (a cpixel).
  • Each pixel type is associated with a number of fields, such as for example: • uv-address: a (17 + N)-bit address into the texture map, where N bits uniquely identify the chip and bank, eight bits identify the column address (CAS), and nine bits identify the row address (RAS);
  • Cpixels and tpixels can be packed, so that the 24-bit color or 32-bit z-value can include four, two, or one colors or z-values. Packing and/ or unpacking is accomplished by conventional means.
  • the three types of pixels use the above-defined fields as follows:
  • tpixels use uv-addresses as PPH addresses into frame buffer 207.
  • Other types of pixels use fb-addresses.
  • Pixels are selected from PPH 202 to be written into memory controller FIFO queue 203 according to the method shown in Fig. 4, which favors pixels on open pages to pixels in other memory locations.
  • the steps shown in Fig. 4 are performed for each cycle.
  • System 200 first checks 401 whether PPH 202 is empty; if so, the method ends 402 and no pixel is written to queue 203. If PPH 202 is not empty, system 200 checks whether queue 203 is full; if so, the method ends 402 and no pixel is written to queue 203. If queue 203 is not full, system 200 searches 404 for a pixel in PPH 202 whose PPH address contains a bank and row address (RAS) matching the open page and bank of the last pixel read or written.
  • RAS bank and row address
  • system 200 writes 405 the pixel to queue 203. As stated previously, such access can be performed relatively quickly (e.g. in one cycle) on some memory systems. If no such pixel is found, system 200 searches 406 for a pixel in PPH 202 whose PPH address matches any open page. If such a pixel is found, system 200 writes 405 the pixel to queue 203. As stated previously, such access can be performed at intermediate speed (e.g. in three cycles) on some memory systems. If no such match is found, system 200 searches 407 for any valid pixel in PPH
  • system 200 writes 405 the pixel to queue 203.
  • searches 404, 406 and 407 are performed concurrently for improved performance.
  • Memory controller 204 retrieves pixels from queue 203 and performs accesses to texture map RAM 205 accordingly, based on the pixel representations from queue
  • Frame buffer 207 is output on display 209 which is implemented as a conventional cathode-ray tube or other output device.
  • system 200 uses the following data structures which define the fields of the entries in heap module 305 and pixels module 306.
  • each heap module 305 entry is a nine-bit comparator, an N-bit comparator, and an M-bit comparator.
  • Each entry provides three-bit global output (i.e., each entry provides three bits of output which must be provided as input to MPE 303), including the value of valid, and each one-bit result from the nine- and N- bit comparators.
  • N-bit bank 9-bit RAS; 8-bit CAS; ⁇ ;
  • Fig. 3 contains components called Available module 301, Heap module 305, modified priority encoder (MPE) 303, Control module 304 and Pixels module 306, as well as address registers 302. Each of these components is described below. In the descriptions, each module first generates outputs based on current register/ memory units, and then sets the register/ memory values for the next cycle.
  • MPE modified priority encoder
  • Available module 301 is a FIFO queue containing up to 2M-1 M-bit indices rep- resenting available slots in PPH 202. The following signals are used by available module 301: Input :
  • a client requests that a new pixel be entered into PPH 202 by asserting renable on Available module 301.
  • the new pixel is placed in address registers 302, including bank, RAS, CAS, and other.
  • Available module 301 returns a valid windex which write-enables Heap 202, Control module 204, and Pixels module 203 (any nonzero value write-enables these modules).
  • bank and RAS are written to Heap 202
  • the bank alone is written to Control module 204
  • the full pixel is written to Pixels module 204.
  • the rendering engine also reads the value of windex to check for a non-zero value.
  • any windex asserted by Available module 301 is always a valid index into an empty slot in PPH 202.
  • PPH 202 is implemented primarily by Heap module 305, MPE 303, and Control module 304. Each of these modules will now be described in turn.
  • Heap Module 305 Heap Module 305
  • Heap module 305 uses the following signals:
  • N-bit wbank N-bit wbank; /* pixel write into the PPH */ 9-bit RAS;
  • Heap module 305 maintains a copy of all valid pixel bank and RAS addresses.
  • tags providing information identifying page breaks and open pages to allow PPH 202 to determine in one cycle the best address to be written to the memory controller.
  • each location in heap module 305 compares its pixel's bank address with the i_last_bank address of the pixel last issued to memory controller 204, and its RAS address with the RAS address of the open page on the bank its pixel will be written to (i.e., it compares to i_open_pages[bank]).
  • these two comparators produce a three-bit result indicating the following:
  • Each pixel has such a three-bit result, so that up to 2M-1 such results are generated and provided as input (i_pixel_status) to MPE 303.
  • additional tags are provided to identify the primitive to which each pixel belongs. This is advantageous in order-dependent applications, i.e., in applications where it is preferable to preserve relative order of generation of primitives.
  • a primitive ID field of, for example, two bits may be provided for each pixel.
  • Heap module 305 thus implements the following code:
  • FIG. 8 there is shown an implementation of heap module 305 according to one embodiment of the present invention.
  • N 3
  • ix register 801 represents my ndex.
  • v 802 is the valid bit
  • b 803 is the bank
  • b 803 chooses one of the banks for comparison.
  • Comparator 805 compares RAS 804 with open pages list 809.
  • Comparator 806 compares the last bank accessed with b.
  • Comparator 808 compares the last index and the current index ix 801, in order to make sure that a real entry is available.
  • Multiplexer 807 takes the outputs from 805, 806 and 808 to generate a 3-bit signal to be sent to MPE 303.
  • MPE 303 uses the following signals:
  • MPE 303 chooses a "best" pixel to be output to memory controller 204 based on the following criteria, which is an implementation of the method described above in connection with Fig. 4:
  • MPE 303 implements the following code: if ( i_stall) ⁇
  • o_precharge 0; ⁇ else if (Ipbreak && I in_precharge && there is a j such that
  • Control module 304 keeps track of which pages in which bank are open. The following signals are used by control module 304:
  • N-bit lastjoank 0; N-bit precharge_bank;
  • N-bit bank[2M-l] ⁇ 0, 0, ... 0 ⁇ ;
  • Control module 304 maintains the lastjbank written and the last_index of the pixel written on the previous cycle, and also maintains a table of open_pages[]. It provides these as input to heap module 305.
  • control module 304 and heap module 305 together together (in a simple two-cycle protocol) maintain a correct version of the open_pages[] table.
  • a page break does occur, it is signaled by MPE 303 and no pixels are read from PPH 202 for two cycles (pixels may still be input into PPH 202 during these two cycles so long as PPH 202 is not full).
  • Control module 304 implements the following code:
  • Pixels module 306 is a standard RAM or a register file with control, containing pixel data and color information. The following signals are used by pixels module 306:
  • a pixel is read every cycle from pixels module 306 and written to memory controller 204.
  • the read from pixels module 306 is pipelined with the logic in control module 304, heap module 305, and MPE 303 so that the pixel selected in cycle t is read from pixels module 306 and written to the memory controller in cycle t+1.
  • no pixel is read from pixels module 306 if PPH 202 is empty.
  • FIG. 7 there is shown a flowchart of a PPH 202 read operation based on a two-cycle pipeline.
  • the timing in this pipeline is based on two-phase clocking.
  • PPH 202 processes signals from storage through combinational logic.
  • storage is updated.
  • steps 700 through 704 occur; on cycle t+1, steps 705 through 707 occur.
  • Cycle t First, the signals from control module 304 to heap module 305 must be valid. Then heap module 305 generates 701 all o_pixel_status[] bits. MPE 303 selects 702 an index and provides it to control module 304 and pixels module 306. Control module 304 updates 703 last_bank, lastjndex, and bank[i_index] . Pixels module 306 saves 704 the index obtained in 702 for use in the next cycle.
  • Cycle t+1 The index of the entry in heap module 305 that was selected in 702 last cycle (if any) is written 705 to available module 301 (o_done_index). Pixels module 306 writes 706 the pixel at rindex (last cycle's i ndex) to memory controller 204.
  • PPH 202 is capable of operation as a simple FIFO queue if iJFIFO is asserted to MPE 303.
  • MPE 303 keeps the lastjndex read from PPH 202 (this is initialized to 0) and on each cycle attempts to return ((lastjndex + 1) mod 2 ⁇ ).
  • mod 2M the lastjndex
  • available module 301 is a FIFO queue, and entries are returned to available module 301 in the order read from PPH 202.
  • PPH 202 Given MPE's 303 operation in FIFO mode to return consecutive entries as long as the pixel is valid, PPH 202 thus operates in this mode as a FIFO queue.
  • the client of PPH 202 must assert i_FIFO throughout the time that FIFO mode is desired, and must also ensure that PPH 202 is empty before it is transitioned between FIFO mode and PPH mode. Any other operation is undefined. Page Update Protocol
  • a page break when a page break occurs, two cycles are required to update the open pages table.
  • MPE 303 detects that a page break will occur as the result of a given pixel's bank and RAS address, it signals such to control module 304 via o pbreak. Since the pixel index ojndex is provided to control module 304 and is also provided to pixels module 306, the pixel is also written to memory controller 204 on the next cycle. Hence, whenever a pixel address that will cause a page break is issued to memory, there will soon be some length of dead time, e.g. six or eight cycles on some memory systems. PPH 202 can use this dead time to update tables on a page break.
  • MPE 303 During the two additional cycles MPE 303 requires for page update after a page break, MPE 303 writes no candidate pixels to control module 304 or pixels module 306.
  • the page update protocol operates as follows. When control module 304 is informed by MPE 303 that a page break has occurred, it asks heap module 305 for the RAS address that caused the page break. When heap module 305 returns the RAS address (o_RAS and o_valid_RAS), control module 304 updates open_pages[] to the new address for that bank. In terms of cycles that begin at c, the protocol is as follows:
  • MPE 303, control module 304, and heap module 305 each have two states with respect to the page update protocol.
  • MPE 303 is normally in state 505 and transitions to update state 506 at cycle c to initiate page update protocol.
  • Control module 304 is normally in state 503 and transitions to update state 504 at cycle c upon receiving i pbreak from MPE 303.
  • Heap module 501 is normally in state 501 and transitions to update state 502 at cycle c + 1 upon receiving i_pbreak from control module 304; at this time, heap module 501 also saves the RAS of the page break.
  • MPE 303 asserts i_valid_RAS
  • heap module 305 sends saved RAS to control module 304 and asserts o_valid_RAS
  • control module 304 updates the new RAS. All three modules 303, 304, and 305 then return to their normal states 505, 503, and 501 respectively.
  • PPH 202 When not in FIFO mode, PPH 202 in one embodiment precharges one page on some "alternate" bank when it is possible to do so. Precharging is a technique that is well-known in the art. By precharging a page on an alternate bank, the present invention is able to hide the precharge penalty by interleaving precharge with another read or write.
  • Fig. 6 there is shown a flowchart of a method of selecting a page to be precharged.
  • PPH 202 determines 601 whether there are any valid pixels left. A bank will be precharged only if there is a valid pixel that can use a page on that bank. In one embodiment, PPH 202 performs this determination by checking the valid bit of the pixels. If no valid pixels are left, no pages will be precharged. Otherwise, PPH 202 selects 602 a valid pixel.
  • PPH 202 determines 603 whether the selected pixel is on the currently-used bank. If so, the bank is not precharged, and PPH 202 returns to 601 to select another pixel. In one embodiment, PPH 202 performs the determination in 603 by checking whether the pixel status is 0x7 or 0x5, which indicates that the pixel is on the currently- used bank.
  • PPH 202 determines 604 whether the pixel is on a bank for which there is an open page that can be used. If so, using that open page incurs no penalty; therefore, the bank is not precharged and PPH 202 returns to 601 to select another pixel. In one embodiment, PPH 202 performs the determination in 604 by checking whether the bank containing the pixel has another pixel with matching bank; in other words, PPH 202 the pixel's bank against the banks of all other pixels with status 0x3. Alternatively, for simplicity, PPH 202 can simply check whether there is any pixel that can still use another bank; i.e. determine whether there any pixel on the bank with status 0x3.
  • PPH 202 proceeds to precharge 605 the page containing the pixel, as described more fully below. PPH 202 then ensures 606 that pixels subsequently chosen to use the precharged bank use the page that was precharged. In one embodiment, PPH 202 implements this by updating the table of open pages after the precharge 605 is performed. Thus, after PPH 202 selects pixels for the current bank and open page, it will automatically select pixels for the open page just precharged, since the pixel selection method prefers open pages to page breaks. In practice, PPH 202 uses essentially the same page update protocol used for page breaks, and therefore typically requires an additional two cycles before the page tables are actually updated. In the interim, it explicitly passes through only those pixels whose addresses are for the current bank and open page.
  • Alternate bank precharge is implemented by MPE 303, control module 304, and pixels module 306.
  • MPE 303 detects that an alternate bank can be precharged (by the method of Fig. 6), it writes the ojndex to control module 304 and pixels module 306, and asserts o_precharge.
  • MPE 303 also sets a single status bit, in precharge, which guides subsequent actions. Whenever in_precharge is set, MPE 303 issues transactions only on the current bank and open page. This is in contrast to its behavior when a page break occurs; in the latter case MPE 303 issues no further transactions until the page update protocol has completed.
  • control module 304 When control module 304 receives a precharge request, it saves the bank to be precharged in a register precharge jbank, and at the next cycle initiates essentially the same page update protocol (via o precharge) that is used on a page break, as described above. There are three differences: 1) names of the signals include precharge instead of pbreak; 2) heap module 305 does not delete the pixel that causes the precharge; and 3) MPE 303 allows pixels with addresses for the current page and bank to pass through during precharge.
  • heap module 305 on receiving i precharge are essentially the same as its actions on receiving ijpbreak; in particular, it saves the RAS to be precharged in save RAS.
  • control module 304 is returned the RAS via i_RAS, and it updates open_pages[precharge_bank] to i_RAS.
  • heap module 305 continues to generate candidate pixels for memory accesses. If any of these are for the current bank and open page, they are passed on to control module 304 and pixels module 306 by MPE 303, and are issued as memory transactions to the FIFO queue. No other transactions are allowed by MPE 303 until the in precharge bit is cleared. This occurs on the final cycle of the page update protocol, when heap module 305 asserts i_valid_RAS to control module 304 and MPE 303.
  • a precharge request to pixels module 306 is identical to a read or write except that ijprecharge is asserted; since pixels module 306 concatenates the precharge bit with the bank, RAS, and CAS address when it writes an entry to memory controller FIFO queue 203, pixels module 306 is effectively unaware that the transaction is a precharge rather than a read or a write. A page break and a precharge cannot occur simultaneously; MPE 303 keeps minimal state to prevent this.
  • the invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof.
  • the pixel priority heap described above may have application in any architecture involving paged memory systems, including both graphics systems and other applications. Accordingly, the disclosure of the present invention is intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the following claims.

Abstract

A system and method for reordering memory references for pixels to improved bandwidth and performance in texture mapping systems and other graphics systems by improving memory locality in conventional page-mode memory systems. Pixel memory references are received from a client graphics engine and placed in a pixel priority heap (202). The pixel priority heap (202) reorders the pixel memory references so that references requiring a currently open page are, in general, processed before references that require page breaks. Reordered pixel memory references are transmitted to a memory controller (204) for accessing memory (205, 207).

Description

PIXEL REORDERING FOR IMPROVED TEXTURE MAPPING
BACKGROUND OF THE INVENTION
5 1. Field of the Invention
The present invention relates generally to texture mapping in graphics systems, and more particularly to a system and method of reordering pixels to improve bandwidth and performance.
2. Description of Background Art w Texture mapping is the process of mapping an image onto a surface in a three- dimensional graphics system. This technique is well-known in the art, and is described, for example, in J. Foley et al., Computer Graphics: Principles and Practice, 2d. ed., Addison-Wesley, 1990, at 741-44.
Referring now to Fig. 1, there is shown an example of texture mapping
15 according to the prior art. The image to be mapped is referred to as a texture map 101, and its individual elements are referred to as texels. Texture map 101 is typically described in a rectangular coordinate scheme designated (u, v), and is ordinarily stored in some area of conventional memory, such as, for example, a conventional page-mode dynamic random-access memory (DRAM) or other paged memory. In the example of
20 Fig. 1, four pages 110, 111, 112, 113 are shown, each corresponding to a portion of the image area containing a corresponding portion of texture map 101.
Surface 104 in three-dimensional space has its own coordinate system (s, t). In a typical three-dimensional graphics system, surface 104 may be a primitive such as a polygon; many such polygons may be defined in three-space to form a three-
25 dimensional object or scene. Each such polygon would then have its own coordinate system (s, t) similar to the surface 104 indicated in Fig. 1. Based on the orientation of surface 104 in three-space, and on the position and orientation of the "camera", surface 104 is in turn mapped onto a two-dimensional display grid 103 stored in the frame buffer for display by the computer system. The mapping of surface 104 onto display
30 grid 103 is accomplished by matrix transforms that are well-known in the art. Display grid 103 has coordinate system (x, y) and is typically implemented in an area of memory reserved for video display, such as video random-access memory (video RAM) e.g. VRAM or synchronous graphics random-access memory (SGRAM). Display grid 103 contains individual elements known as pixels, represented by distinct memory locations in video RAM.
Each pixel in some region of display grid 103 maps onto a point on surface 104 and in turn to a point in texture map 101. Thus, in the example of Fig. 1, point AXy of display grid 103 maps onto point Ast in the coordinates of surface 102 and to point Auv in texture map 101, or a group of points forming a region in texture map 101. Each of the mappings among display grid 103, surface 102, and texture map 101 may be point-to-point, point-to-region, region-to-point, or region-to-region. In conventional implementations of texture mapping systems, the mapping from display grid 103 to surface 104 and in turn to texture map 101 also generates a value d representing the level of detail for the particular texel. Typically, d is a measure of the perceived distance of the point in the texture map, as determined by a z-value for the point in the frame buffer. Points that are perceived to be farther away have a lower level of detail representing decreased resolution. In retrieving texels from texture map 101, d is used to implement a multum in parvo map (MIP map) scheme wherein several texels may be averaged and mapped onto one pixel of the frame buffer. The higher the value of d, the lower the level of detail, and the more pixel-averaging is performed. In the limit, the entire texture map may be reduced to one pixel in the frame buffer.
Conventional rasterization engines draw the image into the frame buffer by the known technique of scan conversion of primitives such as polygons and lines (see, for example, Foley et al.). Scan conversion takes as its input primitives defined in terms of vertices and orientations, and provides as its output a series of pixels to be drawn on the screen. As each pixel is generated by scan conversion, a rasterization engine performs the necessary mapping calculations to determine which texel of texture map 101 corresponds to the pixel. The rasterization engine then issues whatever memory references are required, such as texture fetch, z-fetch, z-writeback, color fetch, color write-back, and the like) to retrieve texel information for writing to the pixel begin processed. Thus, memory references are issued in the order generated by the scan conversion. Conventionally, such memory references are stored and managed according to a first-in first-out (FIFO) scheme using a FIFO queue.
It is known that conventional page-mode DRAM components incur access-time penalties when accessing memory locations from different memory pages. For example, in some memory architectures such as SGRAM, an access to an open page requires one cycle, a read from a page not open requires nine cycles, a write to a page not open requires six cycles, and an access to an open page on a different bank requires three cycles. Thus, the above-described scheme of issuing memory references in the order generated by scan conversion may incur such penalties, as the referenced areas of texture map 101 may lie in different pages. In fact, depending on the distortion of the texture boundaries resulting from the particular mapping transformation being employed, references generated in scan conversion order may require repeated page- switching back and forth. Since memory bandwidth is generally the bottleneck in fast generation of three-dimensional images, such repeated page-swapping results in diminished performance.
There have been attempts in the prior art to alleviate the above-described problem. One example is the use of specialized memory in place of conventional page- mode memory components. See, for example, H. Fuchs and J. Poulton, "Pixel-Planes: A VLSI-Oriented Design for a Raster Graphics Engine," in VLSI Design, vol. 2., no. 3, 1981; M. Deering et al, "FBRAM: A New Form of Memory Optimized for 3D
Graphics," in Computer Graphics, Proceedings of SIGGRAPH, 1995; A. Schilling et al., "Texram: A Smart Memory for Texturing," in IEEE Computer Graphics and Applications, 1996. Such systems generally improve memory bandwidth by, for example, associating memory directly with processing on an application-specific integrated circuit (ASIC), or by associating logic with memory directly on a DRAM chip. See A. Schilling et al. However, such techniques require highly specialized components that are generally more expensive than conventional page-mode DRAM.
Another attempt to reduce memory bandwidth is described in K. Akeley, "RealityEngine Graphics," in Computer Graphics, Proceedings of SIGGRAPH, 1993. Akeley describes a system of extreme memory interleaving. This technique requires significantly more memory chips than do conventional memory systems, and consequently is more expensive. What is needed is a system of reducing memory bandwidth by minimizing page-switching in conventional page-mode memory, so as to improve performance of graphic engines for a wide range of client algorithms without substantially increasing the number of components or the overall cost.
SUMMARY OF THE INVENTION
In accordance with the present invention, there is provided a system and method of reordering memory references required to process independent pixels in a texture mapping system in order to achieve improved memory locality and hence improved graphics performance using conventional page-mode memory components. Since the present invention does not require any changes to be made to the underlying graphics algorithm, memory locality for many such algorithms can be improved at once.
Various types of reads and writes may be performed in connection with rendering by a graphics engine. These include, for example, reading and writing for texture retrieval, retrieval of data from the frame buffer (such as color values, z- values, and the like), color write-back (final color write to frame buffer after resolving all applicable textures), and the like. The present invention reorders the reads and writes associated with pixels when performing such processing so that references that require the currently open page are, in general, processed before references that require opening another page, in order to improve memory locality and minimize page- switching. The technique described herein allows the graphics rasterization engine to achieve better bandwidth using conventional page-mode memory components and without require extreme memory interleaving.
The present invention employs a Pixel Priority Heap (PPH) to reorder pixels. In one embodiment, the PPH reorders addresses into texture memory to perform more reads from the same page before that page is closed and another opened. In alternative embodiments, the PPH is employed to reorder reads and writes from/ to other clients as well. For example, PPH may be used to reorder references to destination addresses in the frame buffer to improve locality; this is advantageous when PPH reordering of texture memory references results in changes to the order of frame buffer writes, so that a PPH is beneficial to frame buffer addresses as well. Each pixel generated by scan conversion is written, one per cycle, to the PPH. If the PPH is full, the pixel is not written, and the system waits until an entry becomes available in the PPH before writing to it.
Concurrently, the system attempts to select one pixel from the PPH and write it to a memory controller first-in first-out (FIFO) queue to be textured. If there are no pixels in the PPH, or if the memory controller FIFO queue is full, no pixel is written to the memory controller FIFO queue. In selecting a pixel from the PPH to be written into the memory controller FIFO queue, the system attempts to find a pixel whose PPH address contains a bank and row address (RAS) matching the open page on the last pixel read or written. If such a match is found, the access can be performed relatively quickly (e.g. in one cycle) on some memory systems. If no such match is found, the system attempts to find a pixel whose PPH address matches any open page, thus permitting intermediate-speed (e.g. three-cycle) access on some memory systems. If no such match is found, the system attempts to find any valid pixel in the PPH, regardless of bank or page address. On some memory systems, such an access is relatively slow (e.g. nine cycles for reads or six cycles for writes). In one embodiment, all such attempts to find matches are performed concurrently for improved performance.
The selected pixel is written to the memory controller FIFO queue and deleted from the PPH. On subsequent cycles, a memory controller issues the read or write corresponding to the pixel's PPH address. If the operation is a read, the data may then be returned from memory for subsequent processing depending on the application.
In addition, the PPH can implement conventional first-in first-out (FIFO) functionality. Such implementation may be useful, for example, when using the PPH in connection with z-buff ering. As is known in the art, z-buff er is a technique whereby each pixel written to the frame buffer has an associated stored z-value in a z-buffer. Each incoming pixel write has a z-value that is compared with the stored z-value for the appropriate pixel, and is only written if the incoming pixel's z-value is less than the stored z-value. When the incoming pixel write is performed, the stored z-value is updated with the incoming pixel's z-value. See J. Foley et al. The PPH also allows reads or writes to continue on an open page on a given bank while precharge is done on a page on an alternate bank. In a preferred embodiment, the present invention operates in an essentially stateless manner. No decisions are explicitly based on previous decisions. Such essentially stateless operation is advantageous because it facilitates the use of simpler algorithms and simpler hardware. The present invention emphasizes correct rather than optimal operation by making relatively simple decisions that may not be optimal but that are correct in the sense that they result in performance that is preferably no worse than a conventional FIFO technique.
BRIEF DESCRIPTION OF THE DRAWINGS
Fig. 1 is a diagram showing texture mapping according to the prior art. Fig. 2 is a block diagram showing overall architecture of an embodiment of the present invention.
Fig. 3 is a block diagram showing detailed architecture of an embodiment of the present invention.
Fig. 4 is a flowchart of the method of selecting pixels from the pixel priority heap according to the present invention.
Fig. 5 is a state diagram showing page update operations. Fig. 6 is a flowchart of a method of selecting a page to be precharged according to one embodiment of the present invention.
Fig. 7 is a flowchart of a method of reading from the PPH according to one embodiment of the present invention.
Fig. 8 is a block diagram of an implementation of a heap module according to one embodiment of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
Overall Operation The following detailed description makes reference to several variables and terms, defined here for convenience:
• N bit-length of index (i.e. address of chip and bank); preferred embodiment uses N=32;
• 2N-1 number of banks of memory on this channel (0 represents null value); • 2N-2 number of chips on this channel (two banks each);
• M bit-length of index of a pixel entry in Pixel Priority Heap; preferred embodiment uses M=5;
• 2M-1 maximum number of entries in Pixel Priority Heap (0 represents null value).
Referring now to Fig. 2, there is shown a block diagram of the overall high-level architecture of an embodiment of the present invention. System 200 contains the following elements. Primitives 208 are stored in conventional manner and include, for example, definitions of polygons and lines. Scan converter 201 converts primitives 208 into pixels in a conventional manner. Each pixel is defined in terms of values for u, v, and d, which represent u and v coordinates in texture space, along with a distance value (also known as a compression value) for determining the level of detail for the pixel. In one embodiment, the u, v, and d values for pixels are written to Pixel Priority Heap (PPH) 202, at a rate of one pixel per cycle. In another embodiment, u, v and d values are mapped onto memory references represented as bank, row and column, and these memory references are written to PPH 202. In yet another embodiment, some combination of these two representations is used, depending on the operation being performed, as described below.
In one embodiment, PPH 202 is defined on one channel. If two or more channels are employed (e.g. left and right channels), it is preferable to have a PPH for each channel. This serves to queue pixels independently for each channel and can result in higher aggregate hit-rate because the locality improvement process is being effected for each channel separately. With multiple PPH's, a selector is employed to check address bits of pixels to determine the proper PPH for entry of the pixels. For example, in one embodiment the selector determines the proper PPH based on the last bit of the pixel address.
In one embodiment, a 1MB chip (256K x 32) is provided for each channel. Banks are numbered from 1 to 2N-1, where there are 2N-1 banks (0 is reserved as a null value). In one embodiment, PPH 202 contains 2M-1 entries, each corresponding to one pixel to be textured. In one embodiment, two banks are provided for each chip. Three types of pixels are defined: 1) a pixel to be textured (a tpixeϊ); 2) a pixel to be z-buffered (a zγixel); and 3) a pixel to be written to (or read from) the frame buffer, or a texel to be written to texture memory (a cpixel). Each pixel type is associated with a number of fields, such as for example: • uv-address: a (17 + N)-bit address into the texture map, where N bits uniquely identify the chip and bank, eight bits identify the column address (CAS), and nine bits identify the row address (RAS);
• fb-address: a (17 + N)-bit address into the frame buffer;
• color: 24 bits (eight for red, eight for green, eight for blue); • z-value: 32 bits.
Cpixels and tpixels can be packed, so that the 24-bit color or 32-bit z-value can include four, two, or one colors or z-values. Packing and/ or unpacking is accomplished by conventional means.
In one embodiment, the three types of pixels use the above-defined fields as follows:
• tpixel: uv-address, fb-address, color;
• zpixel: fb-address, z-value;
• cpixel: fb-address, color.
According to one embodiment of the present invention, tpixels use uv-addresses as PPH addresses into frame buffer 207. Other types of pixels use fb-addresses.
Pixels are selected from PPH 202 to be written into memory controller FIFO queue 203 according to the method shown in Fig. 4, which favors pixels on open pages to pixels in other memory locations. The steps shown in Fig. 4 are performed for each cycle. System 200 first checks 401 whether PPH 202 is empty; if so, the method ends 402 and no pixel is written to queue 203. If PPH 202 is not empty, system 200 checks whether queue 203 is full; if so, the method ends 402 and no pixel is written to queue 203. If queue 203 is not full, system 200 searches 404 for a pixel in PPH 202 whose PPH address contains a bank and row address (RAS) matching the open page and bank of the last pixel read or written. If such a pixel is found, system 200 writes 405 the pixel to queue 203. As stated previously, such access can be performed relatively quickly (e.g. in one cycle) on some memory systems. If no such pixel is found, system 200 searches 406 for a pixel in PPH 202 whose PPH address matches any open page. If such a pixel is found, system 200 writes 405 the pixel to queue 203. As stated previously, such access can be performed at intermediate speed (e.g. in three cycles) on some memory systems. If no such match is found, system 200 searches 407 for any valid pixel in PPH
202, regardless of bank or page address. If a valid pixel is found, system 200 writes 405 the pixel to queue 203. As stated previously, on conventional memory systems, such an access is relatively slow (e.g. nine cycles for reads or six cycles for writes). In one embodiment, searches 404, 406 and 407 are performed concurrently for improved performance.
Memory controller 204 retrieves pixels from queue 203 and performs accesses to texture map RAM 205 accordingly, based on the pixel representations from queue
203. Operations 206 such as blending, z-comparing, and the like are performed on the retrieved pixels as required by the application, for output into frame buffer 207 which is conventionally implemented in video RAM. Frame buffer 207 is output on display 209 which is implemented as a conventional cathode-ray tube or other output device. Data Structures
According to one embodiment, system 200 uses the following data structures which define the fields of the entries in heap module 305 and pixels module 306. As noted above, associated with each heap module 305 entry is a nine-bit comparator, an N-bit comparator, and an M-bit comparator. Each entry provides three-bit global output (i.e., each entry provides three bits of output which must be provided as input to MPE 303), including the value of valid, and each one-bit result from the nine- and N- bit comparators.
Data Structure pp jieap: struct pph_heap {
1-bit valid;
N-bit bank;
9-bit RAS;
M-bit my index;
1-bit = index (
M-bit input index,
M-bit this->my index
1-bit = open page (
9-bit input RAS,
9-bit this->RAS
1-bit = bank(
N-bit input bank,
N-bit this->bank
Data Structure γγh pixel: struct pph_pixel {
N-bit bank; 9-bit RAS; 8-bit CAS; }; Implementation
Referring now to Fig. 3, there is shown a detailed block diagram of an embodiment of the present invention. Fig. 3 contains components called Available module 301, Heap module 305, modified priority encoder (MPE) 303, Control module 304 and Pixels module 306, as well as address registers 302. Each of these components is described below. In the descriptions, each module first generates outputs based on current register/ memory units, and then sets the register/ memory values for the next cycle.
Available Module 301
Available module 301 is a FIFO queue containing up to 2M-1 M-bit indices rep- resenting available slots in PPH 202. The following signals are used by available module 301: Input :
M-bit i done index ;
1-bit renable ;
Internal :
M-bit available [ 2M-1 ] ;
Output :
M-bit windex ;
A client (rendering engine) requests that a new pixel be entered into PPH 202 by asserting renable on Available module 301. The new pixel is placed in address registers 302, including bank, RAS, CAS, and other. Available module 301 returns a valid windex which write-enables Heap 202, Control module 204, and Pixels module 203 (any nonzero value write-enables these modules). When a non-zero windex is asserted, bank and RAS are written to Heap 202, the bank alone is written to Control module 204, and the full pixel is written to Pixels module 204. To determine if PPH 202 is full, the rendering engine (not shown) also reads the value of windex to check for a non-zero value. If PPH 202 is full, windex will be 0. When Heap module 305 removes a pixel that has been successfully output to memory controller 204, it returns to Available module 301 the o_done_index of the slot that contained the pixel. Whenever o_done_index is non- zero, write to available module 301 is enabled, and the value of o_done_index is itself entered in the FIFO queue. Finally, Available module 301 is initialized to contain all indices into PPH 202. Since an index is removed from Available module 301 whenever a new pixel is written to PPH 202, and since the same index is returned to the FIFO queue whenever the pixel is removed from PPH 202, any windex asserted by Available module 301 is always a valid index into an empty slot in PPH 202.
Thus, because the heap module 305 writes to the available module 301 only the indices of those pixels that have been removed from PPH 202, it is guaranteed that these writes always succeed. Heap module 305 cannot write a value of 0 to available module 301 (i_done_index = 0 does nothing), and so neither can a value of 0 be returned when renable is asserted. Each cycle, an index can be read from and another index written to available module 301.
PPH 202 is implemented primarily by Heap module 305, MPE 303, and Control module 304. Each of these modules will now be described in turn. Heap Module 305
Heap module 305 uses the following signals:
Input :
N-bit wbank; /* pixel write into the PPH */ 9-bit RAS;
M-bit windex;
1-bit i_pbreak; /* page-break indication */
1-bit i_precharge; /* precharge */
9-bit i_open_pages [ 2N-1 ] ; M-bit i_last_index;
N-bit i_last_bank; Internal :
1-bit pbreak = 0;
1-bit in_precharge = 0; 9-bit save_RAS = 0; struct pph_heap heap[2M_l] =
{ {0,0,0,1}, {0,0,0,2}, ... {0,0,0,2M}}; Output :
3-bit o_pixel_status [ 2M-1 ] ; /* valid+comparator results */
9-bit o_RAS; /* page-break protocol */
1-bit o_valid_RAS;
M-bit o_done_index; /* pixel removal */
Heap module 305 maintains a copy of all valid pixel bank and RAS addresses.
Associated with each entry are tags providing information identifying page breaks and open pages to allow PPH 202 to determine in one cycle the best address to be written to the memory controller. This is implemented as two comparators (=_open_page() and =_bank()) . An additional comparator (=_index()) is provided to allow each location in Heap module 305 to delete a pixel that was dispatched to memory controller 204 in the previous cycle. On each cycle, each location in heap module 305 compares its pixel's bank address with the i_last_bank address of the pixel last issued to memory controller 204, and its RAS address with the RAS address of the open page on the bank its pixel will be written to (i.e., it compares to i_open_pages[bank]). Combined with a valid bit for each location, these two comparators produce a three-bit result indicating the following:
• 111: the pixel is destined for the open page on the last bank written to;
• 011: the pixel is destined for an open page on another bank;
• 001 or 101: the pixel is valid but is not destined to any open page; • 000: the pixel is not valid.
Each pixel has such a three-bit result, so that up to 2M-1 such results are generated and provided as input (i_pixel_status) to MPE 303.
In one embodiment, additional tags are provided to identify the primitive to which each pixel belongs. This is advantageous in order-dependent applications, i.e., in applications where it is preferable to preserve relative order of generation of primitives. A primitive ID field of, for example, two bits may be provided for each pixel.
Heap module 305 thus implements the following code:
Concurrently on each cycle:
; if ( i_last_index == 0 ) o_done_index = 0 ; o_valid_RAS = (pbreak | in_precharge) ; o_RAS = save_RAS ; }
Concurrently on each cycle, each entry "this" in the heap [ ] performs :
{ if (this->=_index (this->my_index, i_last_index) ) { save_RAS = this->RAS; if ( ! i_precharge) {
/* Return slot to Available */ o_done_index = this->my_index; this->valid = 0;
} o_pixel_status [this->my_index] = 0; } else {
/* Each entry provides three bits on output * (from MSB to LSB) : banks match, pages
* match, pixel valid. Each entry tests its
* bank against the i_last_bank, and each uses
* its own open bank as an index into the
* i_open_pages [] to test its own RAS on that * bank against the page that is open on that
* bank. */ o_pixel_status [this->my_index]
(this->=_bank(i_last_bank, this- >bank)«2 I (this->=_open_page (open_pages [this-
>bank] , this->RAS) «1) | this->valid; Concurrently on each cycle:
{ if (windex>0) { heap [windex] .valid = 1; heap [windex] .bank = wbank; heap [windex] .RAS = RAS; } pbreak = i_pbreak; in_precharge = i_precharge;
Referring now to Fig. 8, there is shown an implementation of heap module 305 according to one embodiment of the present invention. In the example shown, N=3, for eight banks of memory on each channel and 8 MB of memory. ix register 801 represents my ndex. v 802 is the valid bit, b 803 is the bank, b 803 chooses one of the banks for comparison. Comparator 805 compares RAS 804 with open pages list 809. Comparator 806 compares the last bank accessed with b. Comparator 808 compares the last index and the current index ix 801, in order to make sure that a real entry is available. Multiplexer 807 takes the outputs from 805, 806 and 808 to generate a 3-bit signal to be sent to MPE 303.
MPE 303
MPE 303 uses the following signals:
Input : 1-bit i_FIFO; /* PPH is in FIFO mode */
1-bit i_stall;
3-bit i_pixel_status [2M-1] ;
1-bit i_valid_RAS; Internal: M-bit last_index = 0;
/* initialized so after +, we will */ /* have first available */
M-bit tmp;
1-bit in_precharge = 0; 1-bit pbreak = 0;
Output :
M-bit o_index;
1-bit o_pbreak;
1-bit o_precharge;
MPE 303 chooses a "best" pixel to be output to memory controller 204 based on the following criteria, which is an implementation of the method described above in connection with Fig. 4:
• Choose the index of a pixel with i_pixel_status = 111 if there is any such pixel (same bank, same page) (step 404);
• Choose the index of a pixel with i_pixel_status = 011 if there is any such pixel (open page on another bank) (step 406);
• Choose the index of a pixel with i_pixel_status = 002 or 101 if there is any such pixel (any valid pixel) (step 407); • Return 0 (no valid pixels) (step 401 or 407).
MPE 303 implements the following code: if ( i_stall) {
/* Memory controller FIFO must assert stall one
* cycle before the FIFO is actually full ... * this is because Pixels is one cycle behind
* MPE, and that pixel must be written before
* the PPH stalls output. */ o_index = 0; o_pbreak = 0; o_precharge = 0; } else { if (i_FIFO) { tmp = ( last_index + 1 ) > 2M ? 1 : (last_index + 1) ; if (i_pixel_status [tmp] & 0x01) { o_index = tmp; o_pbreak = 0; o_precharge = 0; last__index = tmp; } lse {
Concurrently on each cycle over all 2M_1 inputs { /* bank precharge
* TBD — could add "there is an m such * that 0x07". */ if (Ipbreak && ! in_precharge && there is no j such that
(i_pixel_status [j ]==0x03) && there is a k such that (i_pixel_status[k]==0x01) ){ o_index = k; o_pbreak = 0; o_precharge = 1; in_precharge = 1; } else if (Ipbreak && there is a j such that (i_pixel_status[j]==0x07) ){
/* pixel reordering */ o_index = j; o_pbreak = 0;
o_precharge = 0; } else if (Ipbreak && I in_precharge && there is a j such that
(i_pixel_status[j]==0x03) ) { o_index = j; o_pbreak = 0; o_precharge = 0; } else if (Ipbreak && I in_precharge && there is a j such that
(i_ρixel_status[j] & 0x01) ) { /* note the & implies don't care
* for bank */ o_index = j; o__pbreak = 1; o_precharge = 0; pbreak = 1; } else { o_index = 0; o_pbreak =0; o_precharge=0 ;
} if ( i_valid_RAS) { in_precharge = 0 ; pbreak = 0 ;
Control Module 304
Control module 304 keeps track of which pages in which bank are open. The following signals are used by control module 304:
Input :
N-bit wbank; /* pixel write to PPH */
M-bit windex;
M-bit i_index; /* pixel read from PPH */
1-bit i_pbreak; /* page-break and precharge
1-bit i_precharge; /* protocol */
1-bit i_valid_RAS;
9-bit i_RAS; Internal:
1-bit pbreakO - 0, pbreakl = 0;
1-bit in_precharge0 = 0; in_prechargel = 0;
N-bit lastjoank = 0; N-bit precharge_bank;
M-bit last_index = 0;
N-bit bank[2M-l] = {0, 0, ... 0};
9-bit open_pages[N-l] = {0, 0, ... 0}; Output : N-bit o_last_bank; /* to heap's comparators */
M-bit o_last_index;
9-bit o_open_pages [2N-1] ;
1-bit o_pbreak; /* page update protocols */
1-bit o_precharge;
Control module 304 maintains the lastjbank written and the last_index of the pixel written on the previous cycle, and also maintains a table of open_pages[]. It provides these as input to heap module 305. In addition, control module 304 and heap module 305 together (in a simple two-cycle protocol) maintain a correct version of the open_pages[] table. When a page break does occur, it is signaled by MPE 303 and no pixels are read from PPH 202 for two cycles (pixels may still be input into PPH 202 during these two cycles so long as PPH 202 is not full).
Control module 304 implements the following code:
Concurrently on each cycle:
{ o_pbreak = pbreakO ; o_precharge = in_prechargeO ; o_last_bank = last_bank; o_last_index = last_index ; o_open_pages [ ] = open_pages [ ] ; if ( i_valid_RAS ) { if (pbreakl ) openjpages [ last_bank] = i_RAS ; if ( in_prechargel ) open_pages [precharge_bank] = i_RAS ; } if ( i_index > 0 ) { if ( i_precharge) precharge_bank = bank [i_index] ; else last_bank = bank [i_index] ; } last_index = i_index; pbreakl = pbreakO; pbreakO = i_pbreak; in_prechargel = in_precharge0; in_precharge0 = i_precharge; if (windex > 0) bank [windex] = wbank; }
Pixels Module 306
Pixels module 306 is a standard RAM or a register file with control, containing pixel data and color information. The following signals are used by pixels module 306:
Input:
1-bit i precharge;
M-bit i index;
M-bit windex;
N-bit wbank;
9-bit wRAS;
8-bit wCAS;
Internal:
M-bit rindex;
(17+n)-bit pixel [2M-1] ;
Output :
1-bit o precharge;
N-bit o bank;
9-bit o_RAS;
8-bit o_CAS ;
1-bit FIFO wr; A pixel is read every cycle from pixels module 306 and written to memory controller 204. The read from pixels module 306 is pipelined with the logic in control module 304, heap module 305, and MPE 303 so that the pixel selected in cycle t is read from pixels module 306 and written to the memory controller in cycle t+1. Of course, no pixel is read from pixels module 306 if PPH 202 is empty.
Pixels module 306 implements the following code: if ( rindex > 0 ) { o_precharge = i_precharge; o_bank, o_RAS , o_CAS , o_other = pixel [ rindex] ; FIFO_wr = 1 ; rindex = i_index; } else FIFO_wr = 0;
Read Operation
Referring now to Fig. 7, there is shown a flowchart of a PPH 202 read operation based on a two-cycle pipeline. The timing in this pipeline is based on two-phase clocking. In the first phase, PPH 202 processes signals from storage through combinational logic. In the second phase, storage is updated. On cycle t, steps 700 through 704 occur; on cycle t+1, steps 705 through 707 occur.
Cycle t: First, the signals from control module 304 to heap module 305 must be valid. Then heap module 305 generates 701 all o_pixel_status[] bits. MPE 303 selects 702 an index and provides it to control module 304 and pixels module 306. Control module 304 updates 703 last_bank, lastjndex, and bank[i_index] . Pixels module 306 saves 704 the index obtained in 702 for use in the next cycle.
Cycle t+1: The index of the entry in heap module 305 that was selected in 702 last cycle (if any) is written 705 to available module 301 (o_done_index). Pixels module 306 writes 706 the pixel at rindex (last cycle's i ndex) to memory controller 204.
FIFO Mode
PPH 202 is capable of operation as a simple FIFO queue if iJFIFO is asserted to MPE 303. Under such operation, MPE 303 keeps the lastjndex read from PPH 202 (this is initialized to 0) and on each cycle attempts to return ((lastjndex + 1) mod 2^). In particular, if the pixel at this consecutive index (mod 2M) is valid (which MPE 303 can determine from i_pixel_status[]), MPE 303 returns that next index (ojndex); MPE 303 in this case also saves the index as lastjndex. Ii the pixel at the consecutive index is invalid, MPE 303 sets ojndex = 0, and does not modify lastjndex.
As described above, available module 301 is a FIFO queue, and entries are returned to available module 301 in the order read from PPH 202. Given MPE's 303 operation in FIFO mode to return consecutive entries as long as the pixel is valid, PPH 202 thus operates in this mode as a FIFO queue. In one embodiment, the client of PPH 202 must assert i_FIFO throughout the time that FIFO mode is desired, and must also ensure that PPH 202 is empty before it is transitioned between FIFO mode and PPH mode. Any other operation is undefined. Page Update Protocol
In one embodiment, when a page break occurs, two cycles are required to update the open pages table. When MPE 303 detects that a page break will occur as the result of a given pixel's bank and RAS address, it signals such to control module 304 via o pbreak. Since the pixel index ojndex is provided to control module 304 and is also provided to pixels module 306, the pixel is also written to memory controller 204 on the next cycle. Hence, whenever a pixel address that will cause a page break is issued to memory, there will soon be some length of dead time, e.g. six or eight cycles on some memory systems. PPH 202 can use this dead time to update tables on a page break. During the two additional cycles MPE 303 requires for page update after a page break, MPE 303 writes no candidate pixels to control module 304 or pixels module 306. The page update protocol operates as follows. When control module 304 is informed by MPE 303 that a page break has occurred, it asks heap module 305 for the RAS address that caused the page break. When heap module 305 returns the RAS address (o_RAS and o_valid_RAS), control module 304 updates open_pages[] to the new address for that bank. In terms of cycles that begin at c, the protocol is as follows:
Figure imgf000023_0001
Figure imgf000024_0002
Referring now to Fig. 5, there is shown a state diagram of page update operations. MPE 303, control module 304, and heap module 305 each have two states with respect to the page update protocol. MPE 303 is normally in state 505 and transitions to update state 506 at cycle c to initiate page update protocol. Control module 304 is normally in state 503 and transitions to update state 504 at cycle c upon receiving i pbreak from MPE 303. Heap module 501 is normally in state 501 and transitions to update state 502 at cycle c + 1 upon receiving i_pbreak from control module 304; at this time, heap module 501 also saves the RAS of the page break. At cycle c + 2, MPE 303 asserts i_valid_RAS, heap module 305 sends saved RAS to control module 304 and asserts o_valid_RAS, and control module 304 updates
Figure imgf000024_0001
the new RAS. All three modules 303, 304, and 305 then return to their normal states 505, 503, and 501 respectively.
Alternate Bank Precharge
When not in FIFO mode, PPH 202 in one embodiment precharges one page on some "alternate" bank when it is possible to do so. Precharging is a technique that is well-known in the art. By precharging a page on an alternate bank, the present invention is able to hide the precharge penalty by interleaving precharge with another read or write. Referring now to Fig. 6, there is shown a flowchart of a method of selecting a page to be precharged. First, PPH 202 determines 601 whether there are any valid pixels left. A bank will be precharged only if there is a valid pixel that can use a page on that bank. In one embodiment, PPH 202 performs this determination by checking the valid bit of the pixels. If no valid pixels are left, no pages will be precharged. Otherwise, PPH 202 selects 602 a valid pixel.
PPH 202 then determines 603 whether the selected pixel is on the currently-used bank. If so, the bank is not precharged, and PPH 202 returns to 601 to select another pixel. In one embodiment, PPH 202 performs the determination in 603 by checking whether the pixel status is 0x7 or 0x5, which indicates that the pixel is on the currently- used bank.
If the selected pixel is not on the currently-used bank, PPH 202 then determines 604 whether the pixel is on a bank for which there is an open page that can be used. If so, using that open page incurs no penalty; therefore, the bank is not precharged and PPH 202 returns to 601 to select another pixel. In one embodiment, PPH 202 performs the determination in 604 by checking whether the bank containing the pixel has another pixel with matching bank; in other words, PPH 202 the pixel's bank against the banks of all other pixels with status 0x3. Alternatively, for simplicity, PPH 202 can simply check whether there is any pixel that can still use another bank; i.e. determine whether there any pixel on the bank with status 0x3.
If the selected pixel is not on a bank that has an open page that can be used, PPH 202 proceeds to precharge 605 the page containing the pixel, as described more fully below. PPH 202 then ensures 606 that pixels subsequently chosen to use the precharged bank use the page that was precharged. In one embodiment, PPH 202 implements this by updating the table of open pages after the precharge 605 is performed. Thus, after PPH 202 selects pixels for the current bank and open page, it will automatically select pixels for the open page just precharged, since the pixel selection method prefers open pages to page breaks. In practice, PPH 202 uses essentially the same page update protocol used for page breaks, and therefore typically requires an additional two cycles before the page tables are actually updated. In the interim, it explicitly passes through only those pixels whose addresses are for the current bank and open page.
The method of alternate bank precharge according to one embodiment will now be described. Alternate bank precharge is implemented by MPE 303, control module 304, and pixels module 306. When MPE 303 detects that an alternate bank can be precharged (by the method of Fig. 6), it writes the ojndex to control module 304 and pixels module 306, and asserts o_precharge. MPE 303 also sets a single status bit, in precharge, which guides subsequent actions. Whenever in_precharge is set, MPE 303 issues transactions only on the current bank and open page. This is in contrast to its behavior when a page break occurs; in the latter case MPE 303 issues no further transactions until the page update protocol has completed. When control module 304 receives a precharge request, it saves the bank to be precharged in a register precharge jbank, and at the next cycle initiates essentially the same page update protocol (via o precharge) that is used on a page break, as described above. There are three differences: 1) names of the signals include precharge instead of pbreak; 2) heap module 305 does not delete the pixel that causes the precharge; and 3) MPE 303 allows pixels with addresses for the current page and bank to pass through during precharge.
The operation of heap module 305 on receiving i precharge are essentially the same as its actions on receiving ijpbreak; in particular, it saves the RAS to be precharged in save RAS. On the cycle after this, control module 304 is returned the RAS via i_RAS, and it updates open_pages[precharge_bank] to i_RAS. During the two cycles that heap module 305 and control module 304 exchange information, heap module 305 continues to generate candidate pixels for memory accesses. If any of these are for the current bank and open page, they are passed on to control module 304 and pixels module 306 by MPE 303, and are issued as memory transactions to the FIFO queue. No other transactions are allowed by MPE 303 until the in precharge bit is cleared. This occurs on the final cycle of the page update protocol, when heap module 305 asserts i_valid_RAS to control module 304 and MPE 303.
A precharge request to pixels module 306 is identical to a read or write except that ijprecharge is asserted; since pixels module 306 concatenates the precharge bit with the bank, RAS, and CAS address when it writes an entry to memory controller FIFO queue 203, pixels module 306 is effectively unaware that the transaction is a precharge rather than a read or a write. A page break and a precharge cannot occur simultaneously; MPE 303 keeps minimal state to prevent this. From the above description, it will be apparent that the invention disclosed herein provides a novel and advantageous system and method of reordering pixels in graphics systems to improve bandwidth and performance. The foregoing discussion discloses and describes merely exemplary methods and embodiments of the present invention. As will be understood by those familiar with the art, the invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. In particular, the pixel priority heap described above may have application in any architecture involving paged memory systems, including both graphics systems and other applications. Accordingly, the disclosure of the present invention is intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the following claims.

Claims

What is Claimed is:
1. A system for reordering memory references for pixels to improve memory locality in a page-mode memory architecture, comprising: an available module for receiving pixel memory references from a client graphics engine; and a pixel priority heap coupled to the available module for receiving pixel memory references from the available module, for reordering pixel memory references to improve memory locality, and for transmitting reordered pixel memory references to a memory controller.
2. The system of claim 1, wherein the pixel memory references comprise addresses into a texture memory space.
3. The system of claim 1, wherein the pixel memory references comprise addresses into a frame buffer.
4. The system of claim 1, wherein the pixel priority heap reorders pixel memory references responsive to searching for memory references on open pages and banks.
5. The system of claim 1, wherein, responsive to a mode selection signal, the pixel priority heap selectively transmits memory references in the order received.
6. The system of claim 1, wherein: the pixel priority heap comprises a plurality of slots for storage of pixel memory references, each slot having a status selectable from an available status indicating that no reference is stored in the slot and an unavailable status indicating that a reference is stored in the slot; and the available module further stores indices representing available slots in the pixel priority heap.
7. The system of claim 1, further comprising a memory controller coupled to the pixel priority heap for receiving transmitted reordered pixel memory references and for accessing memory in response to the reordered pixel memory references of the pixel priority heap.
8. The system of claim 1, wherein the pixel priority heap comprises: a heap module for storing pixel memory references and tags describing storage locations of the pixel memory references; a control module coupled to the heap module for tracking open memory pages; and a modified priority encoder coupled to the heap module and to the control module, for selecting a pixel memory reference to be transmitted to the memory controller.
9. The system of claim 8, further comprising a pixels module coupled to the modified priority encoder and to the memory controller for pipelining pixel memory references to the memory controller.
10. The system of claim 1, wherein the pixel priority heap selectively precharges a memory page on an alternate bank.
11. A system for reordering memory references for pixels to improve memory locality in a page-mode memory architecture, comprising: a heap module comprising a plurality of slots, each slot for storing a pixel memory reference and tags describing a storage location of the pixel memory reference, each slot further having a status selectable from an available status indicating that no reference is stored in the slot and an unavailable status indicating that a reference is stored in the slot; an available module coupled to the heap module for receiving pixel memory references from a client graphics engine, for storing indices representing available slots in the pixel priority heap and for transmitting pixel memory references to the heap module; a control module coupled to the heap module for tracking open memory pages; a modified priority encoder coupled to the heap module and to the control module, for selecting a pixel memory reference to be transmitted responsive to searching for memory references on open pages and banks to improve memory locality; a pixels module coupled to the modified priority encoder for pipelining pixel memory references; and a memory controller coupled to the pixels module for receiving pipelined reordered pixel memory references from the pixels module and for accessing memory in response to the pixel memory references.
12. A method of reordering memory references for pixels to improve memory locality in a page-mode memory architecture, comprising: a) writing a pixel memory reference into a pixel priority heap; and b) concurrently with a): b.l) selecting a pixel from the pixel priority heap; and b.2) transmitting the selected pixel to a memory controller.
13. The method of claim 12, wherein b.l) comprises: b.1.1) determining whether any pixel memory references in the pixel priority heap have an address matching an open page and bank of an immediately previously transmitted pixel memory reference; b.l.2) responsive to b.1.1) indicating that a matching pixel memory reference exists, transmitting the pixel memory reference; and b.1.3) responsive to b.1.1) indicating that a matching pixel memory reference does not exist: b.l.4) determining whether any pixel memory references in the pixel priority heap have an address matching any open page; b.l.5) responsive to b.l.4) indicating that a matching pixel memory reference exists, transmitting the pixel memory reference; and b.l.6) responsive to b.l.4) indicating that a matching pixel memory reference does not exist and responsive to the existence of any valid pixel memory reference: b.l.7) issuing a page break; and b.l.8) selecting and transmitting a valid pixel memory reference.
14. The method of claim 12, further comprising: c) selectively precharging a memory page on an alternate bank.
15. The method of claim 14, wherein c) comprises: cl) determining if any valid pixel memory references exist; and c.2) responsive to cl) indicating that at least one valid pixel memory reference exists: c.3) selecting a valid pixel memory reference having an associated memory page; c.4) responsive to the selected reference being associated with a currently-used bank of memory, returning to cl); c.5) responsive to the selected reference being associated with a bank having a usable open page, returning to cl); and c.6) responsive to the selected reference not being associated with a currently-used bank of memory and not being associated with a bank having a usable open page, precharging the memory page associated with the selected reference.
AMENDED CLAIMS
[received by the International Bureau on 22 June 1998 (22.06.98) ; original cl aims 1 - 15 replaced by amended claims 1 - 14 (4 pages) ]
1. A system for reordering memory references for pixels to improve memory locality in a page-mode memory architecture, comprising: an available module for receiving pixel memory references from a client graphics engine; and a pixel priority heap coupled to the available module for receiving pixel memory references from the available module, for reordering pixel memory references to improve memory locality, and for transmitting reordered pixel memory references to a memory controller;
wherein: the pixel priority heap comprises a plurality of slots for storage of pixel memory references, each slot having a status selectable from an available status indicating that no reference is stored in the slot and an unavailable status indicating that a reference is stored in the slot; and the available module further stores indices representing available slots in the pixel priority heap.
2. The system of claim 1, wherein the pixel memory references comprise addresses into a texture memory space.
3. The system of claim 1, wherein the pixel memory references comprise addresses into a frame buffer.
4. The system of claim 1, wherein the pixel priority heap reorders pbcel memory references responsive to searching for memory references on open pages and banks.
5. The system of claim 1, wherein, responsive to a mode selection signal, the pixel priority heap selectively transmits memory references in the order received.
6. The system of claim 1, further comprising a memory controller coupled to the pixel priority heap for receiving transmitted reordered pixel memory references and for accessing memory in response to the reordered pixel memory references of the pixel priority heap.
7. The system of claim 1, wherein the pixel priority heap further comprises: a heap module containing the plurality of slots of the pixel priority heap, the heap module for storing pixel memory references and tags describing storage locations of the pixel memory references; a control module coupled to the heap module for tracking open memory pages; and a modified priority encoder coupled to the heap module and to the control module, for selecting a pixel memory reference to be transmitted to the memory controller.
8. The system of claim 7, further comprising a pixels module coupled to the modified priority encoder and to the memory controller for pipelining pixel memory references to the memory controller.
9. The system of claim 1, wherein the pixel priority heap selectively precharges a memory page on an alternate bank.
10. A system for reordering memory references for pixels to improve memory locality in a page-mode memory architecture, comprising: a heap module comprising a plurality of slots, each slot for storing a pixel memory reference and tags describing a storage location of the pixel memory reference, each slot further having a status selectable from an available status indicating that no reference is stored in the slot and an unavailable status indicating that a reference is stored in the slot; an available module coupled to the heap module for receiving pixel memory references from a client graphics engine, for storing indices representing available slots in the pixel priority heap and for transmitting pixel memory references to the heap module; a control module coupled to the heap module for tracking open memory pages; a modified priority encoder coupled to the heap module and to the control module, for selecting a pixel memory reference to be transmitted responsive to searching for memory references on open pages and banks to improve memory locality; a pixels module coupled to the modified priority encoder for pipelining pixel memory references; and a memory controller coupled to the pixels module for receiving pipelined reordered pixel memory references from the pixels module and for accessing memory in response to the pixel memory references.
11. A method of reordering memory references for pixels to improve memory locality in a page-mode memory architecture, comprising: a) writing a pixel memory reference into an available slot of a pixel priority heap, the pixel priority heap comprising a plurality of slots for storage of pixel memory references, each slot having a status selectable from an available status indicating that no reference is stored in the slot and an unavailable status indicating that a reference is stored in the slot; and b) concurrently with a): b.l) selecting an unavailable slot of the pixel priority heap containing a pixel memory reference; and b.2) transmitting the pixel memory reference contained in the selected unavailable slot to a memory controller.
12. The method of claim 11, wherein b) comprises: b.1.1) determining whether any unavailable slots of the pixel priority heap contain pixel memory references matching an open page and bank of an immediately previously transmitted pixel memory reference; b.1.2) responsive to b.1.1) indicating that a matching pixel memory reference exists, transmitting the pixel memory reference; and b.1.3) responsive to b.1.1) indicating that a matching pixel memory reference does not exist: b.l .4) determining whether any unavailable slots of the pixel priority heap contain pixel memory references matching any open page; b.l.5) responsive to b.l .4) indicating that a matching pixel memory reference exists, transmitting the pixel memory reference; and b.l.6) responsive to b.l. ) indicating that a matching pixel memory reference does not exist and responsive to the existence of any valid pixel memory reference: b.l.7) issuing a page break; and b.l .8) selecting and transmitting a valid pixel memory reference.
13. The method of claim 11, further comprising: c) selectively precharging a memory page on an alternate bank.
14. The method of claim 13, wherein c) comprises: cl) determining if any valid pixel memory references exist; and c.2) responsive to cl) indicating that at least one valid pixel memory reference exists: c.3) selecting a valid pixel memory reference having an associated memory page; c.4) responsive to the selected reference being associated with a currently-used bank of memory, returning to cl); c.5) responsive to the selected reference being associated with a bank having a usable open page, returning to cl); and c.6) responsive to the selected reference not being associated with a currently-used bank of memory and not being as so ciated with a bank having a usable open page, precharging the memory page associated with the selected reference.
PCT/US1997/024264 1997-01-09 1997-12-19 Pixel reordering for improved texture mapping WO1998031004A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
DE69736611T DE69736611T2 (en) 1997-01-09 1997-12-19 Reorganization of memory references for pixels in a page-oriented memory architecture
JP53095398A JP2001507840A (en) 1997-01-09 1997-12-19 Pixel relocation for improved texture mapping
CA002275237A CA2275237C (en) 1997-01-09 1997-12-19 Pixel reordering for improved texture mapping
EP97954326A EP1016068B1 (en) 1997-01-09 1997-12-19 Reordering of memory references for pixels in a page-mode memory architecture
AU58127/98A AU5812798A (en) 1997-01-09 1997-12-19 Pixel reordering for improved texture mapping

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/780,787 US5852451A (en) 1997-01-09 1997-01-09 Pixel reordering for improved texture mapping
US08/780,787 1997-01-09

Publications (1)

Publication Number Publication Date
WO1998031004A1 true WO1998031004A1 (en) 1998-07-16

Family

ID=25120690

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1997/024264 WO1998031004A1 (en) 1997-01-09 1997-12-19 Pixel reordering for improved texture mapping

Country Status (7)

Country Link
US (1) US5852451A (en)
EP (1) EP1016068B1 (en)
JP (1) JP2001507840A (en)
AU (1) AU5812798A (en)
CA (1) CA2275237C (en)
DE (1) DE69736611T2 (en)
WO (1) WO1998031004A1 (en)

Families Citing this family (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3586991B2 (en) * 1996-08-30 2004-11-10 ソニー株式会社 Texture data reading device and rendering device
AU9808298A (en) * 1997-10-15 1999-05-03 Electric Planet, Inc. A system and method for generating an animatable character
AU5688199A (en) 1998-08-20 2000-03-14 Raycer, Inc. System, apparatus and method for spatially sorting image data in a three-dimensional graphics pipeline
US6771264B1 (en) 1998-08-20 2004-08-03 Apple Computer, Inc. Method and apparatus for performing tangent space lighting and bump mapping in a deferred shading graphics processor
US6405267B1 (en) 1999-01-22 2002-06-11 S3 Graphics Co., Ltd. Command reordering for out of order bus transfer
US6577320B1 (en) 1999-03-22 2003-06-10 Nvidia Corporation Method and apparatus for processing multiple types of pixel component representations including processes of premultiplication, postmultiplication, and colorkeying/chromakeying
US6542159B1 (en) * 1999-05-19 2003-04-01 Ati International S.R.L. Apparatus to control memory accesses in a video system and method thereof
US6469703B1 (en) 1999-07-02 2002-10-22 Ati International Srl System of accessing data in a graphics system and method thereof
US6756986B1 (en) * 1999-10-18 2004-06-29 S3 Graphics Co., Ltd. Non-flushing atomic operation in a burst mode transfer data storage access environment
US6811489B1 (en) 2000-08-23 2004-11-02 Nintendo Co., Ltd. Controller interface for a graphics system
US6707458B1 (en) 2000-08-23 2004-03-16 Nintendo Co., Ltd. Method and apparatus for texture tiling in a graphics system
US6700586B1 (en) 2000-08-23 2004-03-02 Nintendo Co., Ltd. Low cost graphics with stitching processing hardware support for skeletal animation
US7538772B1 (en) 2000-08-23 2009-05-26 Nintendo Co., Ltd. Graphics processing system with enhanced memory controller
US7576748B2 (en) 2000-11-28 2009-08-18 Nintendo Co. Ltd. Graphics system with embedded frame butter having reconfigurable pixel formats
US6636214B1 (en) 2000-08-23 2003-10-21 Nintendo Co., Ltd. Method and apparatus for dynamically reconfiguring the order of hidden surface processing based on rendering mode
US7196710B1 (en) 2000-08-23 2007-03-27 Nintendo Co., Ltd. Method and apparatus for buffering graphics data in a graphics system
US6564304B1 (en) * 2000-09-01 2003-05-13 Ati Technologies Inc. Memory processing system and method for accessing memory including reordering memory requests to reduce mode switching
US6891545B2 (en) * 2001-11-20 2005-05-10 Koninklijke Philips Electronics N.V. Color burst queue for a shared memory controller in a color sequential display system
US6778175B2 (en) * 2002-02-05 2004-08-17 Xgi Technology Inc. Method of arbitration of memory request for computer graphics system
US6919896B2 (en) * 2002-03-11 2005-07-19 Sony Computer Entertainment Inc. System and method of optimizing graphics processing
US6989837B2 (en) * 2002-12-16 2006-01-24 S3 Graphics Co., Ltd. System and method for processing memory with YCbCr 4:2:0 planar video data format
US7599044B2 (en) 2005-06-23 2009-10-06 Apple Inc. Method and apparatus for remotely detecting presence
US7577930B2 (en) 2005-06-23 2009-08-18 Apple Inc. Method and apparatus for analyzing integrated circuit operations
US9298311B2 (en) 2005-06-23 2016-03-29 Apple Inc. Trackpad sensitivity compensation
US7433191B2 (en) 2005-09-30 2008-10-07 Apple Inc. Thermal contact arrangement
US7598711B2 (en) 2005-11-23 2009-10-06 Apple Inc. Power source switchover apparatus and method
US20090091564A1 (en) * 2007-10-03 2009-04-09 Raju Thevan System and method for rendering electronic documents having overlapping primitives
US9019292B2 (en) * 2011-09-12 2015-04-28 Microsoft Technology Licensing Llc Reordering graph execution for processing optimization
US10061537B2 (en) 2015-08-13 2018-08-28 Microsoft Technology Licensing, Llc Data reordering using buffers and memory

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5613083A (en) * 1994-09-30 1997-03-18 Intel Corporation Translation lookaside buffer that is non-blocking in response to a miss for use within a microprocessor capable of processing speculative instructions
US5630096A (en) * 1995-05-10 1997-05-13 Microunity Systems Engineering, Inc. Controller for a synchronous DRAM that maximizes throughput by allowing memory requests and commands to be issued out of order

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5269001A (en) * 1989-07-28 1993-12-07 Texas Instruments Incorporated Video graphics display memory swizzle logic circuit and method
US5287470A (en) * 1989-12-28 1994-02-15 Texas Instruments Incorporated Apparatus and method for coupling a multi-lead output bus to interleaved memories, which are addressable in normal and block-write modes
US5715407A (en) * 1992-03-06 1998-02-03 Rambus, Inc. Process and apparatus for collision detection on a parallel bus by monitoring a first line of the bus during even bus cycles for indications of overlapping packets
US5440713A (en) * 1992-05-29 1995-08-08 Industrial Technology Research Institute M-way N-port paged-interleaved memory system
FR2705804B1 (en) * 1993-05-27 1995-08-11 Sgs Thomson Microelectronics Multi-tasking processor architecture.
FR2705805B1 (en) * 1993-05-27 1996-06-28 Sgs Thomson Microelectronics Image processing system.
US5551002A (en) * 1993-07-01 1996-08-27 Digital Equipment Corporation System for controlling a write cache and merging adjacent data blocks for write operations
US5655122A (en) * 1995-04-05 1997-08-05 Sequent Computer Systems, Inc. Optimizing compiler with static prediction of branch probability, branch frequency and function frequency
US5790130A (en) * 1995-06-08 1998-08-04 Hewlett-Packard Company Texel cache interrupt daemon for virtual memory management of texture maps

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5613083A (en) * 1994-09-30 1997-03-18 Intel Corporation Translation lookaside buffer that is non-blocking in response to a miss for use within a microprocessor capable of processing speculative instructions
US5630096A (en) * 1995-05-10 1997-05-13 Microunity Systems Engineering, Inc. Controller for a synchronous DRAM that maximizes throughput by allowing memory requests and commands to be issued out of order

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
ADVANCES IN THE 68040 MEMORY SUBSYSTEM ELECTRONIC ENGINEERING, February 1990, Vol. 62, No. 758, EDENFIELD R., pages 35-41. *
GRAPHICS DRIVE VIDEO DRAMS ELECTRONICS ENGINEERING TIMES, November 1995, No. 876, WILSON R., pages 74-79. *
See also references of EP1016068A4 *

Also Published As

Publication number Publication date
JP2001507840A (en) 2001-06-12
DE69736611D1 (en) 2006-10-12
EP1016068A1 (en) 2000-07-05
EP1016068B1 (en) 2006-08-30
US5852451A (en) 1998-12-22
EP1016068A4 (en) 2004-10-13
CA2275237A1 (en) 1998-07-16
CA2275237C (en) 2002-12-03
AU5812798A (en) 1998-08-03
DE69736611T2 (en) 2007-09-20

Similar Documents

Publication Publication Date Title
US5852451A (en) Pixel reordering for improved texture mapping
US6011565A (en) Non-stalled requesting texture cache
US6288730B1 (en) Method and apparatus for generating texture
US6426753B1 (en) Cache memory for high latency and out-of-order return of texture data
KR100325902B1 (en) Floating point processor for high performance 3D graphic accelerator and its processing function
KR100478767B1 (en) Graphic processing with deferred shading
JP4203128B2 (en) Block and band oriented scanning in 3D triangle drawing
US5987567A (en) System and method for caching texture map information
US7050063B1 (en) 3-D rendering texture caching scheme
JP4410940B2 (en) Texturing system for use in 3D imaging systems
US8200917B2 (en) Multi-media processor cache with cache line locking and unlocking
US7439983B2 (en) Method and apparatus for de-indexing geometry
US7880745B2 (en) Systems and methods for border color handling in a graphics processing unit
US6085292A (en) Apparatus and method for providing non-blocking pipelined cache
US20080150951A1 (en) 3-d rendering engine with embedded memory
US20070052713A1 (en) Systems and methods for storing and fetching texture data using bank interleaving
JPH09500462A (en) Computer graphics system with high performance multi-layer z-buffer
US6300953B1 (en) Apparatus and method for grouping texture cache requests
JPH06175646A (en) Frame buffer and raster processor for graphic system and method for buffering pixel variable
KR100550240B1 (en) Tiled graphics architecture
US20050195200A1 (en) Embedded system with 3D graphics core and local pixel buffer
JP3083724B2 (en) Graphic system signal processing method and apparatus
US6173372B1 (en) Parallel processing of data matrices
KR0170665B1 (en) Ram for 3-dimensional graphics
US6088033A (en) Method and apparatus for processing picture elements

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GE GH HU ID IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZW AM AZ BY KG KZ MD RU TJ TM

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SZ UG ZW AT BE CH DE DK ES FI FR GB GR IE IT LU MC

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
ENP Entry into the national phase

Ref document number: 2275237

Country of ref document: CA

Ref country code: CA

Ref document number: 2275237

Kind code of ref document: A

Format of ref document f/p: F

WWE Wipo information: entry into national phase

Ref document number: 1997954326

Country of ref document: EP

ENP Entry into the national phase

Ref country code: JP

Ref document number: 1998 530953

Kind code of ref document: A

Format of ref document f/p: F

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

WWP Wipo information: published in national office

Ref document number: 1997954326

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 1997954326

Country of ref document: EP