WO2000011613A2 - Performing hidden surface removal in a graphics processor with deferred shading - Google Patents

Performing hidden surface removal in a graphics processor with deferred shading Download PDF

Info

Publication number
WO2000011613A2
WO2000011613A2 PCT/US1999/019190 US9919190W WO0011613A2 WO 2000011613 A2 WO2000011613 A2 WO 2000011613A2 US 9919190 W US9919190 W US 9919190W WO 0011613 A2 WO0011613 A2 WO 0011613A2
Authority
WO
WIPO (PCT)
Prior art keywords
stamp
sample
value
primitive
samples
Prior art date
Application number
PCT/US1999/019190
Other languages
French (fr)
Other versions
WO2000011613A3 (en
WO2000011613A9 (en
Inventor
Jerome F. Duluk, Jr.
Steven L. Dodgen
Emerson S. Fang
Hengwei Hsu
Jason R. Redgrave
Sushma Trivedi
Original Assignee
Apple Computer, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Apple Computer, Inc. filed Critical Apple Computer, Inc.
Priority to AU57825/99A priority Critical patent/AU5782599A/en
Publication of WO2000011613A2 publication Critical patent/WO2000011613A2/en
Publication of WO2000011613A3 publication Critical patent/WO2000011613A3/en
Publication of WO2000011613A9 publication Critical patent/WO2000011613A9/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/50Lighting effects
    • G06T15/80Shading
    • G06T15/87Gouraud shading
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • G06T15/30Clipping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T1/00General purpose image data processing
    • G06T1/20Processor architectures; Processor configuration, e.g. pipelining
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/40Filling a planar surface by adding surface attributes, e.g. colour or texture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/04Texture mapping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • G06T15/20Perspective computation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • G06T15/40Hidden part removal
    • G06T15/405Hidden part removal using Z-buffer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/50Lighting effects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/50Lighting effects
    • G06T15/80Shading
    • G06T15/83Phong shading

Definitions

  • PIPELINE PROCESSOR HAVING ADVANCED FEATURES (Atty. Doc. No. A-66364) Serial No. , filed 20 August 1999, entitled SYSTEM, APPARATUS AND METHOD FOR BALANCING RENDERING RESOURCES IN A THREE-DIMENSIONAL GRAPHICS PIPELINE (Atty. Doc.
  • PROCESSOR (Atty. Doc. No. A-66360).
  • This invention relates to three-dimensional computer graphics computer systems, and more particularly to a system for performing conservative hidden surface removal in a graphics processor with deferred shading.
  • Computer graphics is the art and science of generating pictures with a computer. This picture or image generation process is commonly called rendering.
  • 3D three-dimensional
  • 3D three-dimensional
  • Real-time display devices such as CRTs used as computer monitors, refresh the display by continuously displaying the image over and over. This refresh usually occurs row-by-row, where each row is called a raster line or scan line.
  • raster lines are numbered from bottom to top, but are displayed in order from top to bottom.
  • a sequence of images is displayed, giving the illusion of motion in three- dimensional space.
  • Interactive 3D computer graphics allows a user to change his viewpoint or change the geometry in real-time, thereby requiring the rendering system to create new images on-the-fly in real-time.
  • each renderable object generally has its own local object coordinate system, and therefore needs to be translated (or transformed) from object coordinates to pixel display coordinates.
  • this is a 4-step process: 1) translation (including scaling for size enlargement or shrink) from object coordinates to world coordinates, which is the coordinate system for the entire scene;
  • pixel coordinates also called screen coordinates.
  • Screen coordinates are points in three- dimensional space, and can be in either screen-precision (i.e., pixels) or object-precision (high precision numbers, usually floating-point), as described later.
  • These translation steps can be compressed into one or two steps by pre-computing appropriate translation matrices before any translation occurs.
  • Many techniques are used for generating pixel color values, including Gouraud shading, Phong shading, and texture mapping.
  • Figure 1 shows a three-dimensional object, a tetrahedron, with its own coordinate axes (x obJ ,y obJ ,z obJ ).
  • the three-dimensional object is translated, scaled, and placed in the viewing point's coordinate system based on (x eye .y eye . z eye )- Tne object is projected onto the viewing plane, thereby correcting for perspective.
  • the object appears to have become two-dimensional; however, the object's z coordinates are preserved so they can be used later by hidden surface removal techniques.
  • the object is finally translated to screen coordinates, based on (x screen .y screen . z screen ). where z. creen is going perpendicularly into the page.
  • Points on the object now have their x and y coordinates described by pixel location (and fractions thereof) within the display screen and their z coordinates in a scaled version of distance from the viewing point.
  • the geometry representing the surfaces closest to the scene viewing point must be determined.
  • the visible surfaces within the volume subtended by the pixel's area determine the pixel color value, while hidden surfaces are prevented from affecting the pixel.
  • Non-opaque surfaces closer to the viewing point than the closest opaque surface (or surfaces, if an edge of geometry crosses the pixel area) affect the pixel color value, while all other non-opaque surfaces are discarded.
  • the term "occluded" is used to describe geometry which is hidden by other non-opaque geometry.
  • image-precision and “object-precision” are defined: "Image-precision algorithms are typically performed at the resolution of the display device, and determine the visibility at each pixel. Object-precision algorithms are performed at the precision with which each object is defined, and determine the visibility of each object.”
  • the depth complexity of a scene is a measure of the wasted processing. For example, for a scene with a depth complexity of ten, 90% of the computation is wasted on hidden pixels.
  • This wasted computation is typical of hardware renderers that use the simple Z buffer technique (discussed later herein), generally chosen because it is easily built in hardware. Methods more complicated than the Z Buffer technique have heretofore generally been too complex to build in a cost-effective manner.
  • An important feature of the method and apparatus invention presented here is the avoidance of this wasted computation by eliminating hidden portions of geometry before they are rasterized, while still being simple enough to build in cost- effective hardware.
  • the point When a point on a surface (frequently a polygon vertex) is translated to screen coordinates, the point has three coordinates: 1) the x coordinate in pixel units (generally including a fraction); 2) the y coordinate in pixel units (generally including a fraction); and 3) the z coordinate of the point in either eye coordinates, distance from the virtual screen, or some other coordinate system which preserves the relative distance of surfaces from the viewing point.
  • positive z coordinate values are used for the "look direction" from the viewing point, and smaller values indicate a position closer to the viewing point.
  • a surface is approximated by a set of planar polygons
  • the vertices of each polygon are translated to screen coordinates.
  • the screen coordinates are interpolated from the coordinates of vertices, typically by the processes of edge walking and span interpolation.
  • a z coordinate value is generally included in each pixel value (along with the color value) as geometry is rendered.
  • the Deering Reference includes a diagram of a generic 3D graphics pipeline (i.e., a renderer, or a rendering system) that it describes as "truly generic, as at the top level nearly every commercial 3D graphics accelerator fits this abstraction", and this pipeline diagram is reproduced here as Figure 2.
  • Such pipeline diagrams convey the process of rendering, but do not describe any particular hardware.
  • This document presents a new graphics pipeline that shares some of the steps of the generic 3D graphics pipeline. Each of the steps in the generic 3D graphics pipeline will be briefly explained here. Processing of polygons is assumed throughout this document, but other methods for describing 3D geometry could be substituted. For simplicity of explanation, triangles are used as the type of polygon in the described methods.
  • the first step within the floating-point intensive functions of the generic 3D graphics pipeline after the data input is the transformation step (step 214), which was described above.
  • the second step the clip test, checks the polygon to see if it is at least partially contained in the view volume (sometimes shaped as a frustum) (step 216). If the polygon is not in the view volume, it is discarded; otherwise processing continues.
  • the third step is face determination, where polygons facing away from the viewing point are discarded (step 218). Generally, face determination is applied only to objects that are closed volumes.
  • the fourth step, lighting computation generally includes the set up for Gouraud shading and/or texture mapping with multiple light sources of various types, but could also be set up for Phong shading or one of many other choices (step 222).
  • the fifth step, clipping deletes any portion of the polygon that is outside of the view volume because that portion would not project within the rectangular area of the viewing plane (step 224).
  • polygon clipping is done by splitting the polygon into two smaller polygons that both project within the area of the viewing plane. Polygon clipping is computationally expensive.
  • the sixth step, perspective divide does perspective correction for the projection of objects onto the viewing plane (step 226).
  • step seven the screen space conversion step (step 228).
  • step 230 set up for incremental render, computes the various begin, end, and increment values needed for edge walking and span interpolation (e.g.: x, y, and z coordinates; RGB color; texture map space u and v coordinates; and the like).
  • edge walking (step 232) incrementally generates horizontal spans for each raster line of the display device by incrementing values from the previously generated span
  • pixel values also include color values, which can be generated by simple Gouraud shading (i.e., interpolation of vertex color values) or by more computationally expensive techniques such as texture mapping (possibly using multiple texture maps blended together), Phong shading (i.e., per-fragment lighting), and/or bump mapping (perturbing the interpolated surface normal).
  • a double-buffered MUX output look-up table operation is performed (step 238).
  • the blocks with rounded corners typically represent functions or process operations, while sharp cornered rectangles typically represent stored data or memory.
  • the z buffered blend either keeps the new pixel values (if it is closer to the viewing point than previously stored value for that pixel location) by writing it into the frame buffer, or discards the new pixel values (if it is farther).
  • antialiasing methods can blend the new pixel color with the old pixel color.
  • the z buffered blend generally includes most of the per-fragment operations, described below.
  • the generic 3D graphics pipeline includes a double buffered frame buffer, so a double buffered MUX is also included. An output lookup table is included for translating color map values. Finally, digital to analog conversion makes an analog signal for input to the display device.
  • a major drawback to the generic 3D graphics pipeline is its drawing intensive functions are not deterministic at the pixel level given a fixed number of polygons. That is, given a fixed number of polygons, more pixel-level computation is required as the average polygon size increases.
  • the floating-point intensive functions are proportional to the number of polygons, and independent of the average polygon size. Therefore, it is difficult to balance the amount of computational power between the floating-point intensive functions and the drawing intensive functions because this balance depends on the average polygon size.
  • Prior art Z Buffers are based on conventional Random Access Memory (RAM or DRAM), Video RAM (VRAM), or special purpose DRAMs.
  • RAM or DRAM Random Access Memory
  • VRAM Video RAM
  • special purpose DRAM One example of a special purpose D
  • OpenGL ® is a software interface to graphics hardware which consists of several hundred functions and procedures that allow a programmer to specify objects and operations to produce graphical images.
  • the objects and operations include appropriate characteristics to produce color images of three-dimensional objects.
  • Most of OpenGL ® (Version 1.2) assumes or requires a that the graphics hardware include a frame buffer even though the object may be a point, line, polygon, or bitmap, and the operation may be an operation on that object.
  • the general features of OpenGL ® (just one example of a graphical interface) are described in the reference "The OpenGL ® Graphics System: A Specification
  • the invention is not limited to structures, procedures, or methods which are compatible or consistent with OpenGL ® , or with any other standard or non-standard graphical interface.
  • the inventive structure and method may be implemented in a manner that is consistent with the OpenGL ® , or other standard graphical interface, so that a data set prepared for one of the standard interfaces may be processed by the inventive structure and method without modification.
  • inventive structure and method provides some features not provided by OpenGL ® , and even when such generic input/output is provided, the implementation is provided in a different manner.
  • pipeline state does not have a single definition in the prior-art.
  • the OpenGL ® specification sets forth the type and amount of the graphics rendering machine or pipeline state in terms of items of state and the number of bits and bytes required to store that state information.
  • pipeline state tends to include object vertex pertinent information including for example, the vertices themselves the vertex normals, and color as well as "non-vertex" information.
  • object vertex pertinent information including for example, the vertices themselves the vertex normals, and color as well as "non-vertex" information.
  • object geometry information is provided to describe the scene.
  • the object or objects are specified in terms of vertex information, where an object is modeled, defined, or otherwise specified by points, lines, or polygons (object primitives) made up of one or more vertices.
  • a vertex is a location in space and may be specified for example by a three-space (x,y,z) coordinate relative to some reference origin.
  • other information such as a surface normal, color, texture, transparency, and the like information pertaining to the characteristics of the vertex. This information is essentially "per-vertex" information.
  • a color value may be specified in the data stream for a particular vertex and then not respecified in the data stream until the color changes for a subsequent vertex.
  • the color value may still be characterized as per-vertex data even though a color value is not explicitly included in the incoming data stream for each vertex.
  • Texture mapping presents an interesting example of information or data which could be considered as either per-vertex information or pipeline state information.
  • one or more texture maps may be specified, each texture map being identified in some manner, such as with a texture coordinate or coordinates.
  • Other information, not related on a one-to-one basis to the geometry object primitives, used by the renderer such as lighting location and intensity, material settings, reflective properties, and other overall rules on which the renderer is operating may more accurately be referred to as pipeline state.
  • These particular depth tests which change from time to time, may be considered to be pipeline state at that time.
  • Parameters considered to be renderer (pipeline) state in OpenGL ® are identified in Section 6.2 of the aforementioned OpenGL ® Specification (Version 1.2, at pages 193-217).
  • APIs Application Program Interfaces
  • OpenGL ® Open Graphics Library
  • D3D Data Stored Graphics Library
  • OpenGL ® Open Graphics Library
  • Some exemplary OpenGL ® per-fragment operations are briefly reviewed so that any generic similarities and differences between the inventive structure and method and conventional structures and procedures can be more readily appreciated.
  • a frame buffer stores a set of pixels as a two-dimensional array. Each picture- element or pixel stored in the frame buffer is simply a set of some number of bits. The number of bits per pixel may vary depending on the particular GL implementation or context. Corresponding bits from each pixel in the frame buffer are grouped together into a bitplane; each bitplane containing a single bit from each pixel. The bitplanes are grouped into several logical buffers referred to as the color, depth, stencil, and accumulation buffers. The color buffer in turn includes what is referred to under OpenGL ® as the front left buffer, the front right buffer, the back left buffer, the back right buffer, and some additional auxiliary buffers.
  • the values stored in the front buffers are the values typically displayed on a display monitor while the contents of the back buffers and auxiliary buffers are invisible and not displayed.
  • Stereoscopic contexts display both the front left and the front right buffers, while monoscopic contexts display only the front left buffer.
  • the color buffers must have the same number of bitplanes, but particular implementations of context may not provide right buffers, back buffers, or auxiliary buffers at all, and an implementation or context may additionally provide or not provide stencil, depth, or accumulation buffers.
  • the color buffers consist of either unsigned integer color indices or R, G, B, and, optionally, a number "A" of unsigned integer values; and the number of bitplanes in each of the color buffers, the depth buffer (if provided), the stencil buffer (if provided), and the accumulation buffer (if provided), is fixed and window dependent. If an accumulation buffer is provided, it should have at least as many bit planes per R, G, and B color component as do the color buffers.
  • a fragment produced by rasterization with window coordinates of (x w , y w ) modifies the pixel in the frame buffer at that location based on a number of tests, parameters, and conditions.
  • OpenGL ® does not provide for an explicit "color test” between the alpha test and stencil test. Per-Fragment operations under OpenGL ® are applied after all the color computations. Each of these tests or operations is briefly described below.
  • the pixel ownership test determines if the pixel at location (x w , y w ) in the frame buffer is currently owned by the GL context. If it is not, the window system decides the fate of the incoming fragment. Possible results are that the fragment is discarded or that some subset of the subsequent per- fragment operations are applied to the fragment. This pixel ownership test allows the window system to properly control the GL's behavior.
  • the pixel ownership test involves determining which process and associated window owns each of the pixels. If a particular process does not "own" a pixel, it fails the pixel ownership test relative to the frame buffer and that pixel is thrown away. Note that under the typical paradigm, the pixel ownership test is run by each process, and that for a give pixel location in the frame buffer, that pixel may pass the pixel ownership test for one of the processes, and fail the pixel ownership test for the other process.
  • a particular pixel can pass the ownership test for only one process because only one process can own a particular frame buffer pixel at the same time.
  • the pixel ownership test may not be particularly relevant. For example, if the scene is being rendered to an off-screen buffer, and subsequently Block Transferred or "blitted" to the desktop, pixel ownership is not really even relevant. Each process automatically or necessarily passes the pixel ownership test (if it is executed) because each process effectively owns its own off-screen buffer and nothing is in front of that buffer.
  • the pixel is not owned by that process, then there is no need to write a pixel value to that location, and all subsequent processing for that pixel may be ignored.
  • all the data associated with a particular pixel on the screen is read during rasterization. All information for any polygon that feeds that pixel is read, including information as to the identity of the process that owns that frame buffer pixel, as well as the Z buffer, the color value, the old color value, the alpha value, stencil bits, and so forth.
  • a process owns the pixel, then the other downstream process are executed (for example, scissor test, alpha test, and the like) On the other hand, if the process does not own the pixel and fails the ownership test for that pixel, the process need not consider that pixel further and that pixel is skipped for subsequent tests.
  • the scissor test determines if (x ⁇ y lies within a scissor rectangle defined by four coordinate values corresponding to a left bottom (left, bottom) coordinate, a width of the rectangle, and a height of the rectangle. The values are set with the procedure "void Scissor( int left, int bottom, sizei width, sizei height)" under OpenGL ® . If left ⁇ x w ⁇ left+width and bottom ⁇ y w ⁇ bottom+height, then the scissor test passes; otherwise the scissor test fails and the particular fragment being tested is discarded. Various initial states are provided and error conditions monitored and reported. In simple terms, a rectangle defines a window which may be an on-screen or off-screen window.
  • the window is defined by an x-left, x-right, y-top, and y-bottom coordinate (even though it may be expressed in terms of a point and height and width dimensions from that point).
  • This scissor window is useful in that only pixels from a polygon fragment that fall in that screen aligned scissor window will change. In the event that a polygon straddles the scissor window, only those pixels that are inside the scissor window may change.
  • the pipeline calculates everything it needs to in order to determine the z value and color of that pixel. Once z value and color are determined, that information is used to determine what information should be placed in the frame buffer
  • the scissor test provides means for discarding pixels and/or fragments before they actually get to the frame buffer to cause the output to change.
  • Alpha Test Color is defined by four values, red (R), green (G), blue (B), and alpha (A).
  • the RGB values define the contribution from each of the primary colors, and alpha is related to the transparency. Typically, color is a 32-bit value, 8-bits for each component, though such representation is not limited to 32-bits.
  • Alpha test compares the alpha value of a given pixel to an alpha reference value. The type of comparison may also be specified, so that for example the comparison may be a greater-than operation, a less-than operation, and so forth. If the comparison is a greater-than operation, then the pixel's alpha value has to be greater than the reference to pass the alpha test. Any pixel not passing the alpha test is thrown away or discarded.
  • the OpenGL ® Specification describes the manner in which alpha test is implemented in OpenGL ® .
  • Alpha test is a per-fragment operation and after all of the fragment coloring calculations and lighting and shading operations are completed. Each of these per-fragment operations may be thought of as part of the conventional z buffer blending operations.
  • Color Test Color test is similar to the alpha test described hereinbefore, except that rather than performing the magnitude or logical comparisons between the pixel alpha (A) value and a reference value, the color test performs a magnitude or logical comparison between one or a combination of the R, G, or B color components and reference value(s).
  • the comparison test may be for example, greater-than, less-than, equal-to, greater-than-or-equal-to, "greater-than- and less- thai] c " where c ⁇ nd c are sore predetermined reference values, and so forth.
  • Color test might, for example, be useful to provide blue-screen functionality.
  • the comparison test may also be performed on a single color component or on a combination of color components.
  • one typically has one value for each component for the color test there are effectively two values per component, a maximum value and a minimum value.
  • stencil test conditionally discards a fragment based on the outcome of a comparison between a value stored in a stencil buffer at location (x w , y and a reference value.
  • stencil comparison functions are permitted such that whether the stencil test passes can depend upon whether the reference value is less than, less than or equal to, equal to, greater than or equal to, greater than, or not equal to the masked stored value in the stencil buffer.
  • the Under OpenGL ® if the stencil test fails, the incoming fragment is discarded.
  • the reference value and the comparison value can have multiple bits, typically 8 bits so that 256 different values may be represented.
  • stencil bits are part of the pipeline state.
  • the type of stencil test to perform can be specified at the time the geometry is rendered.
  • the stencil bits are used to implement various filtering, masking or stenciling operations. For example, if a particular fragment ends up affecting a particular pixel in the frame buffer, then the stencil bits can be written to the frame buffer along with the pixel information.
  • the depth buffer test discards the incoming fragment if a depth comparison fails.
  • the comparison is enabled or disabled with the generic Enable and Disable commands using the OpenGL ® symbolic constant DEPTH_TEST.
  • depth test is disabled, the depth comparison and subsequent possible updates to the depth buffer value are bypassed and a fragment is passed to the next operation.
  • the stencil bits are also involved and are modified even if the test is bypassed.
  • the stencil value is modified if the depth buffer test passed. If depth test is enabled, the depth comparison takes place and the depth buffer and stencil value may subsequently be modified. The manner in which the depth test is implemented in OpenGL ® is described in greater detail in the OpenGL ® specification at page 145.
  • Depth comparisons are implemented in which possible outcomes are as follows: the depth buffer test either never passes or always passes, if the incoming fragment's z food value is less than, less than or equal to, equal to, greater than, greater than or equal to, or not equal to the depth value stored at the location given by the incoming fragment's (x w , y w ) coordinates. If the depth buffer test fails, the incoming fragment is discarded. The stencil value at the fragment's (x ⁇ y w ) coordinate is updated according to the function currently in effect for depth buffer test failure. Otherwise, the fragment continues to the next operation and the value of the depth buffer at the fragment's (x*,, y w ) location is set to the fragment's z, value. In this case the stencil value is updated according to the function currently in effect for depth buffer test success.
  • the necessary OpenGL ® state is an eight-valued integer and a single bit indicating whether depth buffering is enabled or disabled.
  • alpha blending (also referred toas blending) combines the incoming fragment's R, G, B, and A values with the R, G, B, and A values stored in the frame buffer at the incoming fragment's
  • Blending is basically an operation that takes color in the frame buffer and the color in the fragment, and blends them together.
  • the manner in which blending is achieved, that is the particular blending function, may be selected from various alternatives for both the source and destination. Blending is described in the OpenGL ® specification at page 146-149 and is hereby incorporated by reference.
  • Various blend equations are available under OpenGL ® .
  • the blending equation is evaluated separately for each color component and its corresponding weighting coefficient.
  • Each of the four R, G, B, A components has its own weighting factor.
  • the blending test (or blending equation) is part of pipeline state and can potentially change for every polygon, but more typically would change only for the object made up or several polygons.
  • dithering selects between two color values or indices.
  • RGBA mode consider the value of any of the color components as a fixed-point value with m bits to the left of the binary point, where m is the number of bits allocated to that component in the frame buffer; call each such value c.
  • dithering selects a value d such that d e ⁇ max ⁇ 0, [c]-1 , [c] ⁇ . This selection may depend on the x w and y w coordinates of the pixel.
  • color index mode the same rule applies with c being a single color index.
  • the value of c must not be larger than the maximum value representable in the frame buffer for either the component or the index.
  • Various logical operations may be implemented between source (s) and destination (d), including for example: clear, set, and, noop, xor, or, nor, nand, invert, copy, inverted and, equivalence, reverse or, reverse and, inverted copy, and inverted or.
  • the logicop arguments and corresponding operations, as well as additional details of the OpenGL ® logicop implementation, are set forth in the OpenGL ® specification at pates 150-151.
  • Logical operations are performed independently for each color index buffer that is selected for writing, or for each red, green, blue, and alpha value of each color buffer that is selected for writing.
  • the required state is an integer indicating the logical operation, and two bits indicating whether the logical operation is enabled or disabled.
  • pixels are referred to as the smallest individually controllable element of the display device. But, because images are quantized into discrete pixels, spatial aliasing occurs.
  • a typical aliasing artifact is a "staircase" effect caused when a straight line or edge cuts diagonally across rows of pixels.
  • Some rendering systems reduce aliasing effects by dividing pixels into subpixels, where each sub- pixel can be colored independently. When the image is to be displayed, the colors for all sub-pixels within each pixel are blended together to form an average color for the pixel.
  • a renderer that uses up to 16 sub- pixels per pixel is described in "RealityEngine Graphics", by Akeley, pages 109 to 116 of SIGGRAPH93
  • A-Buffer used to perform blending
  • This technique is also included in the Akeley Reference
  • the A-buffer is an antialiasing technique that reduces aliasing by keeping track of the percent coverage of a pixel by a rendered polygon.
  • the main drawback to this technique is the need to sort polygons front-to-back (or back-to-front) at each pixel in order to get acceptable antialiased polygons.
  • CAM Content Addressable Memories
  • Most Content Addressable Memories perform a bit-for-bit equality test between an input vector and each of the data words stored in the CAM. This type of CAM frequently provides masking of bit positions in order to eliminate the corresponding bit in all words from affecting the equality test. It is inefficient to perform magnitude comparisons in a equality-testing CAM because a large number of clock cycles is required to do the task.
  • CAMs are presently used in translation look-aside buffers within a virtual memory systems in some computers. CAMs are also used to match addresses in high speed computer networks.
  • Magnitude comparison CAM is defined here as any content addressible memory where the stored data are treated as numbers, and arithmetic magnitude comparisons (i.e. less-than, greater-than, less-than-or-equal-to, and the like) are performed on the data in parallel. This is in contrast to ordinary CAM which treats stored data strictly as bit vectors, not as numbers.
  • a magnitude comparison content addressable memory is described in, and incorporated herein by reference, is U.S. Patent Number 4,996,666, by Jerome F. Duluk Jr., entitled "Content-Addressable Memory System Capable of Fully Parallel
  • Duluk Patent Magnitude Comparisons
  • Structures within the Duluk Patent specifically referenced shall include the prefix “Duluk Patent” (for example, “Duluk Patent MCCAM Bit Circuit”).
  • Duluk Patent MCCAM Bit Circuit Other types of magnitude comparison content addressable memories may also be used.
  • the "MCCAM” abbreviation is conveniently used in this description to refer to various types, structures, and methods for magnitude comparison content addressable memory and is not limited to the particular magnitude comparison content addressable memory described in U.S. Patent
  • the basic internal structure of an MCCAM is a set of memory bits organized into words, where each word can perform one or more arithmetic magnitude comparisons between the stored data and input data.
  • each word can perform one or more arithmetic magnitude comparisons between the stored data and input data.
  • a parallel search comparison operation is called a "query" of the stored data.
  • the invention described here augments the capability of the MCCAM by adding various features, including the ability to output all the query result bits every clock cycle and to logically "or” together these output query result bits to form additional outputs.
  • CHSR hidden surface removal
  • the inventive apparatus and method provide conservative hidden surface removal (CHSR) in a deferred shading graphics pipeline (DSGP).
  • the pipeline renders primitives, and the invention is described relative to a set of renderable primitives that include: 1 ) triangles, 2) lines, and 3) points. Polygons with more than three vertices are divided into triangles in the Geometry block (described hereinafter), but the DSGP pipeline could be easily modified to render quadrilaterals or polygons with more sides.
  • the inventive renderer effectively renders any polygon primitive.
  • the invention advantageously takes into account whether and in what part of the display screen a given primitive may appear or have an effect.
  • the pipeline divides the 3D window being drawn into a series of smaller regions, called tiles and stamps.
  • the pipeline performs deferred shading, in which pixel colors are not determined until after hidden-surface removal.
  • MCCAM Addressable Memory
  • inventive Conservative Hidden Surface Removal procedure advantageously maintains compatibility with other standard APIs, such as OpenGL ® , including their support of dynamic rule changes for the primitives (e.g. changing the depth test or stencil test during a scene).
  • inventive deferred shader the conventional rendering paradigm, wherein non-deferred shaders typically execute a sequence of rules for every geometry item and then check the final rendered result, is broken.
  • the inventive structure and method anticipate or predict what geometry will actually affect the final values in the frame buffer without having to make or generate all the colors for every pixel inside of every piece of geometry. In principle, the spatial position of the geometry is examined, and a determination is made for any particular sample, the one geometry item that affects the final color in the z buffer, and then generates only that color.
  • the CHSR processes each primitive in time order and, for each sample that a primitive touches, CHSR makes conservative decision based on the various Application Program Interface
  • the CHSR process can be considered a finite state machine (FSM) per sample.
  • FSM finite state machine
  • each per-sample FSM is called a sample finite state machine.
  • Each sample FSM maintains per-sample data including: (1) z coordinate information; (2) primitive information (any information needed to generate the primitive's color at that sample or pixel, or a pointer to such information); and (3) one or more sample state bits (for example, these bits could designate the z value or z values to be accurate or conservative). While multiple z values per sample can be easily used, multiple sets of primitive information per sample would be expensive.
  • the sample FSM maintains primitive information for one primitive.
  • Each sample FSM may also maintain transparency information, which is used for sorted transparencies.
  • the DSGP can operate in two distinct modes: 1 ) time order mode, and 2) sorted transparency mode.
  • Time order mode is designed to preserve, within any particular tile, the same temporal sequence of primitives.
  • time order mode time order of vertices and modes are preserved within each tile, where a tile is a portion of the display window bounded horizontally and vertically.
  • time order preserved we mean that for a given tile, vertices and modes are read in the same order as they are written.
  • sorted transparency mode the process of reading geometry from a tile is divided into multiple passes. In the first pass, the opaque geometry(i.e., geometry that can completely hide more distant geometry) is processed, and in subsequent passes, potentially transparent geometry is processed.
  • Sorted transparency mode can spatially sort (on a sample-by-sample basis) the geometry into either back- to-front or front-to-back order, thereby providing a mechanism for the visible transparent objects to be blended in spatial order (rather than time order), resulting in a more correct rendering.
  • the sorted transparency method is performed jointly by the Sort block and the Cull block.
  • an object primitive (current primitive) is selected from a group of primitives, each primitive comprising a plurality of stamps.
  • stamps in the current primitive are compared to stamps from previously evaluated primitives in the group of primitives, and a first stamp is selected from the current primitive by the stamp selection process as a current stamp (CS), and optionally by the SAM for performance reasons.
  • CS is compared to a second stamp or a CPVS selected from previously evaluated stamps that have not been discarded. The second stamp is discarded when no part of the second stamp would affect a final graphics display image based on the comparison with the CS.
  • the part of second stamp that would not affect the final image is deleted from the second stamp.
  • the CS is discarded when no part of the second stamp would affect a final graphics display image based on the comparison with the second stamp. If part, but not all, of the CS would not affect the final image based on the comparison with the second stamp, then the part of CS that would not affect the final image is deleted from the CS.
  • the stamps that have not been discarded have their pixels, or samples, colored by the part of the pipeline downstream from these first steps in performing hidden surface removal.
  • the set of non-discarded stamps can be limited to one stamp per sample.
  • the second stamp and the CS include the same sample and both can not be discarded, the second stamp is dispatched and the CS is kept in the list of non-discarded stamps.
  • the visibility of the second stamp and the CS depends on parameters evaluated later in the computer graphics pipeline, the second stamp and the CS are dispatched.
  • the selection of the first stamp by for example the SAM and the stamp selection process, as a current stamp (CS) is based on a relationship test of depth states of samples in the first stamp with depth states of samples of previously evaluated stamps; and an aspect of the inventive apparatus simultaneously performs the relationship test on a multiplicity of stamps.
  • a set of currently potentially visible stamps is maintained separately from the set of current depth values (CDVs), wherein the inventive hidden surface removal method includes the following steps or procedures.
  • an object primitive (current primitive) is selected from a group of primitives, each primitive comprising a plurality of stamps.
  • a first stamp from the current primitive is selected as a currently stamp (CS).
  • a currently potentially visible stamp (CPVS) is selected from the set of CPVSs such that the CPVS overlaps the CS.
  • the depth value of the CS is compared to the corresponding value in the set of CDVs, and this comparison operation takes into account the pipeline state and updates the CDVs.
  • Samples in the selected CPVS that are determined to be not visible are deleted for the selected CPVS. If all samples in the selected CPVS are deleted, the selected CPVS is deleted from the set of CPVS's. If any sample in the CS is determined to be visible, the CS is added to the set of the CPVS's with only its visible samples included. If for any sample both the CS and selected CPVS are visible, then at least those visible samples in the selected CPVS are sent down the pipeline for color computations.
  • the invention provides structure and method for processing in parallel all CPVS's that overlap the CS. Furthermore, the parallel processing is pipelined such that a CS can be processed at the rate of one CS per clock cycle. Also multiple CS's can be processed in parallel.
  • the invention provides structure and method for a hidden surface removal system for a deferred shader computer graphics pipeline, wherein the pipeline includes a Magnitude Comparison
  • MCCAM Content Addressable Memory
  • Stamp Selection unit coupled to the MCCAM cull unit, for identifying, based on the first group and a perimeter of the primitive, a second group of potentially visible samples associated with the primitive
  • Z-Cull unit coupled to the stamp selection unit and the MCCAM cull unit, for identifying visible stamp portions by evaluating a pipeline state, and comparing depth states of the second group with stored depth state values
  • Stamp Portion Memory unit coupled to the Z-Cull unit, for storing visible stamp portions based on control signals received from the Z-Cull unit, wherein the Stamp Portion Memory unit dispatches stamps having a visibility dependent on parameters evaluated later in the computer graphics pipeline.
  • the invention provides structure and method of rendering a graphics image including the steps of: receiving a plurality of primitives to be rendered; selecting a sample location; rendering a front most opaque sample at the selected sample location, and defining the z value of the front most opaque sample as Zfar; comparing z values of a first plurality of samples at the selected sample location; defining to be Znear a first sample, at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the first plurality of samples; rendering the first sample; setting Zfar to the value of Znear; comparing z values of a second plurality of samples at the selected sample location; defining as Znear the z value of a second sample at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the second plurality of samples; and rendering the second sample.
  • Figure 1 is a diagrammatic illustration showing a tetrahedron, with its own coordinate axes, a viewing point's coordinate system, and screen coordinates
  • Figure 2 is a diagrammatic illustration showing a conventional generic renderer for a 3D graphics pipeline.
  • Figure 3 is a diagrammatic illustration showing a first embodiment of the inventive 3-Dimensional
  • Figure 4 is a diagrammatic illustration showing a second embodiment of the inventive 3-Dimensional
  • Figure 5 is a diagrammatic illustration showing a third embodiment of the inventive 3-Dimensional
  • Figure 6 is a diagrammatic illustration showing a fourth embodiment of the inventive 3-Dimensional
  • Figure 7 is a diagrammatic illustration showing a fifth embodiment of the inventive 3-Dimensional
  • Figure 8 is a diagrammatic illustration showing a sixth embodiment of the inventive 3-Dimensional
  • Figure 9 is a diagramatic illustration showing an exemplary flow of data through blocks of an embodiment of the pipeline.
  • Figure 10 is a diagrammatic illustration showing an embodiment of the inventive 3-Dimensional graphics pipeline including information passed between the blocks.
  • Figure 11 is a diagramatic illustration showing the manner in which an embodiment of the Cull block produces fragments from a partially obscured triangle.
  • Figure 12 illustrates a block diagram of the Cull block according to one embodiment of the present invention.
  • Figure 13 illustrates the relationships between tiles, pixels, and stamp portions in an embodiment of the invention.
  • Figure 14 illustrates a detailed block diagram of the Cull block according to one embodiment of the present invention.
  • Figure 15 illustrates a Setup Output Primitive Packet according to one embodiment of the present invention.
  • Figure 16 illustrates a flow chart of a conservative hidden surface removal method according to one embodiment of the present invention.
  • Figure 17A illustrates a sample tile including a primitive and a bounding box.
  • Figure 17B shows the largest z values (ZMax) for each stamp in the tile.
  • Figure 17C shows the results of the z value comparisons between the ZMin for the primitive and the ZMaxes for every stamp.
  • Figure 18 illustrates an example of a stamp selection process of the conservative hidden surface removal method according to one embodiment of the present invention.
  • Figure 19 illustrates an example showing a set of the left most and right most positions of a primitive in each subraster line that contains at least one sample point,
  • Figure 20 illustrates a stamp containing four pixels.
  • Figures 21 A - 21 D illustrate an example of the operation of the Z Cull unit.
  • Figure 22 illustrates an example of how samples are processed by the Z Cull unit.
  • Figures 23A - 23D illustrate an example of early dispatch.
  • Figure 24 illustrates a sample level example of early dispatch processing.
  • Figure 25 illustrates an example of processing samples with alpha test with a CHSR method according to one embodiment of the present invention.
  • Figure 26 illustrates aspects of stencil testing relative to rendering operations for an embodiment of CHSR.
  • Figure 27 illustrates aspects of alpha blending relative to rendering operations for an embodiment of CHSR.
  • Figure 28A illustrates part of a Spatial Packet containing three control bits: DoAlphaTest, DoABIend and Transparent.
  • Figure 28B illustrates how the alpha values are evaluated to set the DoABIend control bit.
  • Figure 29 illustrates a flow chart of a sorted transparency mode CHSR method according to one embodiment of the present invention.
  • the invention is directed to a system for performing Conservative Hidden Surface Removal (CHSR) Deferred Shading Graphics Processor (DSGP) graphics processor.
  • CHSR Conservative Hidden Surface Removal
  • DSGP Deferred Shading Graphics Processor
  • This present invention includes numerous embodiments of the DSGP pipeline.
  • the DSGP pipeline can be connected to a computer via a variety of possible interfaces, including but not limited to for example, an Advanced Graphics Port (AGP) and/or a PCI bus interface, amongst the possible interface choices. VGA and video output are generally also included.
  • Embodiments of the invention supports both OpenGL ® and Direct3D APIs.
  • the OpenGL ® specification entitled “The OpenGL ® Graphics System: A Specification (Version 1.2)" by Mark Segal and Kurt Akeley, edited by Jon Leech, is included incorporated by reference.
  • the inventive structure and method anticipate or predict what geometry will actually affect the final values in the frame buffer without having to make or generate all the colors for every pixel inside of every piece of geometry.
  • the spatial position of the geometry is examined, and a determination is made for any particular sample, the one geometry item that affects the final color in the z buffer, and then generate only that color.
  • Samples are processed in parallel, and generally all the samples in all the pixels within a stamp are processed in parallel. Hence, if one stamp can be processed per clock cycle (and there are 4 pixels per stamp and 4 samples per pixel), then 16 samples are processed per clock cycle.
  • a "stamp" defines the number of pixels and samples processed at one time. This per-stamp processing is generally pipelined, with pipeline stalls injected if a stamp needs to be processed again before the same stamp (from a previous primitive) has completed (that is, unless out-of-order stamp processing can be handled).
  • each sample FSM can be replaced by a pointer into a memory where all the primitive information is stored.
  • the Color Pointer is used to point to a primitive's information in Polygon Memory.
  • Stamps are also used to reduce the number of data packets transmitted down the pipeline. That is, when one sample within a stamp is dispatched (either early dispatch or end-of-tile dispatch), other samples within the same stamp and the same primitive are also dispatched (such a joint dispatch is hereinafter called a Visible Stamp Portion, or VSP).
  • VSP Visible Stamp Portion
  • alpha test if alpha values for a primitive arise only from the alpha values at the vertices (not from other places such as texturing), then a simplified alpha test can be done for entire primitives. That is, the vertex processing block (which is the Geometry block in a preferred embodiment) can determine when any interpolation of the vertex alpha values would be guaranteed to pass the alpha test, and for that primitive, disable the alpha test. This cannot be done if the alpha values cannot be determined before CHSR is performed. If a frame does not start with depth clear, then the sample FSMs are set to their most conservative state (with near z values at the minimum and far z values at the maximum).
  • the DSGP can operate in two distinct modes: 1 ) time order mode, and 2) sorted transparency mode.
  • Time order mode is described above, and is designed to preserve, within any particular tile, the same temporal sequence of primitives.
  • the Sorted Transparency mode is described immediately below.
  • the control of the pipeline operating mode is done in the Sort block 6000.
  • Sort block 6000 is located in the pipeline between a Mode Extraction block (MEX) 4000 and Setup (STP) block 8000.
  • Sort block operates primarily to take geometry scattered around the display window and sort it into tiles.
  • Sort block also manages the Sort Memory, which stores all the geometry from the entire scene before it is rasterized, along with some mode information.
  • Sort memory comprises a double-buffered list of vertices and modes.
  • One page collects a scene's geometry (vertex by vertex and mode by mode), while the other page is sending its geometry (primitive by primitive and mode by mode) down the rest of the pipeline.
  • vertices and modes are written sequentially into the sort memory as they are received by the sort block.
  • the read is done on a tile-by-tile basis, and the read process operates in two modes: (1 ) time order mode, and (2) sorted transparency mode.
  • time order mode time order of vertices and modes are preserved within each tile, where a tile is a portion of the display window bounded horizontally and vertically. By time order preserved, we mean that for a given tile, vertices and modes are read in the same order as they are written. 4.0.5 Sorted Transparency Mode
  • sorted transparency mode In sorted transparency mode, the process of reading geometry from a tile is divided into multiple passes. In the first pass, the Sort block outputs guaranteed opaque geometry, and in subsequent passes the Sort block outputs potentially transparent geometry. Within each sorted transparency mode pass, the time ordering is preserved, and mode data is inserted in its correct time-order location. Sorted transparency mode can be performed in either back-to-front or front-to-back order. In a preferred embodiment, the sorted transparency method is performed jointly by the Sort block and the Cull block.
  • Each frame (also called a scene or user frame) of 3D graphics primitives is rendered into a 3D window on the display screen.
  • a window consists of a rectangular grid of pixels, and the window is divided into tiles (hereinafter tiles are assumed to be 16x16 pixels, but could be any size). If tiles are not used, then the window is considered to be one tile.
  • Each tile is further divided into stamps (hereinafter stamps are assumed to be 2x2 pixels, thereby resulting in 64 stamps per tile, but stamps could be any size within a tile).
  • Each pixel includes one or more of samples, where each sample has its own color values and z value (hereinafter, pixels are assumed to include four samples, but any number could be used).
  • a fragment is the collection of samples covered by a primitive within a particular pixel. The term "fragment" is also used to describe the collection of visible samples within a particular primitive and a particular pixel.
  • the renderer calculates the color value (RGB or RGBA) and z value for each pixel of each primitive, then compares the z value of the new pixel with the current z value in the Z-buffer. If the z value comparison indicates the new pixel is "in front of the existing pixel in the frame buffer, the new pixel overwrites the old one; otherwise, the new pixel is thrown away.
  • Z buffer rendering typically results in a great deal of wasted processing effort if the scene contains many hidden surfaces.
  • the renderer may calculate color values for ten or twenty times as many pixels as are visible in the final picture. This means the computational cost of any per-pixel operation —such as Phong shading or texture-mapping — is multiplied by ten or twenty. Therefore an improved method of rendering that reduces Z buffer rendering inefficiency is needed.
  • HSR hidden surface removal
  • the objective of a deferred shading pipeline is to generate pixel colors for only those primitives that appear in the final image (i.e., exact HSR).
  • Eliminating samples before the resource intensive coloring process is performed on the samples improves the efficiency of the rendering process.
  • the primitives should generally be accumulated before HSR begins.
  • the HSR process determines the single visible primitive at each sample within all the pixels. Once the visible primitive is determined for a sample, then the primitive's color at that sample location is determined. Additional efficiency can be achieved by determining a single per-pixel color for all the samples within the same pixel, rather than computing per-sample colors.
  • HSR process For a frame with at least some alpha blending of primitives (generally due to transparency), there are some samples that are colored by two or more primitives. This means the HSR process must determine a set of visible primitives per sample.
  • Alpha blending is described in greater detail in the aforementioned OpenGL ® specification. in some APIs, such as OpenGL ® , the HSR process can be complicated by other operations (that is by operations other than depth test) that can discard primitives. These other operations include: pixel ownership test, scissor test, alpha test, color test, and stencil test (as described elsewhere in this specification).
  • Some of these operations discard a primitive based on its color (such as alpha test), which is not determined in a deferred shading pipeline until after the HSR process (this is because alpha values are often generated by the texturing process, included in pixel fragment coloring). For example, a primitive that would normally obscure a more distant primitive (generally at a greater z value) can be discarded by alpha test, thereby causing it to not obscure the more distant primitive. A HSR process that does not take alpha test into account could mistakenly discard the more distant primitive. Hence, there may be an inconsistency between deferred shading and alpha test (similarly, with color test and stencil test); that is, pixel coloring is postponed until after hidden surface removal, but hidden surface removal can depend on pixel colors. Simple solutions to this problem include: 1) eliminating non-depth-dependent tests from the
  • CHSR conservative hidden surface removal
  • This complication is compounded by an API allowing the application program to change the depth test within a frame.
  • Different geometry may be subject to drastically different rules for visibility.
  • a deferred shading pipeline must preserve the time ordering of primitives, and the correct pipeline state (for example, the depth test) must be associated with each primitive.
  • FIG. 2 A conventional 3D graphics pipeline is illustrated in Figure 2.
  • the following describes a first embodiment of the inventive 3D Deferred Shading Graphics Pipeline Version 1 (hereinafter "DSGPvl"), as shown in Figure 3.
  • DSGPvl inventive 3D Deferred Shading Graphics Pipeline Version 1
  • Figure 3 the inventive pipeline ( Figure 3) has been obtained from the generic conventional pipeline ( Figure 2) by replacing the drawing intensive functions 231 with: (1 ) a scene memory 250 for storing the pipeline state and primitive data describing each primitive, called scene memory in the figure; (2) an exact hidden surface removal process 251 ; (3) a fragment coloring process 252; and
  • the scene memory 250 stores the primitive data for a frame, along with their attributes, and also stores the various settings of pipeline state throughout the frame.
  • Primitive data includes vertex coordinates, texture coordinates, vertex colors, vertex normals, and the like
  • primitive data also includes the data generated by the setup for incremental render, which includes spatial, color, and edge derivatives.
  • the scene memory 250 can be double buffered, thereby allowing the HSR process to perform computations on one frame while the floating-point intensive functions perform computations on the next frame.
  • the scene memory can also be triple buffered.
  • the scene memory could also be a scratchpad for the HSR process, storing intermediate results for the HSR process, allowing the HSR process to start before all primitive have been stored into the scene memory.
  • every primitive is associated with the pipeline state information that was valid when the primitive was input to the pipeline. The simplest way to associate the pipeline state with each primitive is to include the entire pipeline state within each primitive.
  • the preferred way to store information in the scene memory is to keep separate lists: one list for pipeline state settings and one list for primitives. Furthermore, the pipeline state information can be split into a multiplicity of sub-lists, and additions to each sub-list occurs only when part of the sub-list changes.
  • the preferred way to store primitives is done by storing a series of vertices, along with the connectivity information to re-create the primitives. This preferred way of storing primitives eliminates redundant vertices that would otherwise occur in polygon meshes and line strips.
  • the HSR process described relative to DSGPvl is required to be an exact hidden surface removal (EHSR) because it is the only place in the DSGPvl where hidden surface removal is done.
  • EHSR exact hidden surface removal
  • the exact hidden surface removal (EHSR) process 251 determines precisely which primitives affect the final color of the pixels in the frame buffer. This process accounts for changes in the pipeline state, which introduces various complexities into the process. Most of these complications stem from the per-fragment operations (ownership test, scissor test, alpha test, and the like), as described above. These complications are solved by the innovative conservative hidden surface removal (CHSR) process, described later, so that exact hidden surface removal is not required.
  • CHSR innovative conservative hidden surface removal
  • the fragment coloring process generates colors for each sample or group of samples within a pixel. This can include: Gouraud shading, texture mapping, Phong shading, and various other techniques for generating pixel colors.
  • This process is different from edge walk 232 and span interpolation 234 because this process must be able to efficiently generate colors for subsections of primitives. That is, a primitive may be partially visible, and therefore, colors need to be generated for only some of its pixels, and edge walk and span interpolation assume the entire primitive must be colored.
  • the HSR process may generate a multiplicity of visible subsections of a primitive, and these may be interspersed in time amongst visible subsections of other primitives.
  • the fragment coloring process 252 should be capable of generating color values at random locations within a primitive without needing to do incremental computations along primitive edges or along the x-axis or y-axis.
  • the blending process 253 of the inventive pipeline combines the fragment colors together to generate a single color per pixel.
  • this blending process 253 does not include Z buffer operations because the exact hidden surface removal process 251 as already determined which primitives are visible at each sample.
  • the blending process 253 may keep separate color values for each sample, or sample colors may be blended together to make a single color for the entire pixel. If separate color values are kept per sample and are stored separately into the Frame buffer 240 , then final pixel colors are generated from sample colors during the scan out process as data is sent to the digital to analog converter 242.
  • the scene memory 250 stores: (1 ) primitive data; and (2) pipeline state.
  • (1 ) primitive data As described above for DSGPvl , the scene memory 250 stores: (1 ) primitive data; and (2) pipeline state.
  • (2) pipeline state In a second embodiment of the Deferred Shading Graphics Pipeline 260 (Version 2)
  • this scene memory 250 is split into two parts: a spatial memory 261 part and polygon memory 262 part.
  • the split of the data is not simply into primitive data and pipeline state data.
  • pipeline state needed for HSR is stored into spatial memory 261 , while the rest is stored into polygon memory 262.
  • Examples of pipeline state needed for HSR include (as defined, for example, in the OpenGL ® Specification) are DepthFunc, DepthMask, StencilEnable, etc.
  • Primitive data is also split.
  • the part of the primitive data needed for HSR is stored into spatial memory 261 , and the rest of the primitive data is stored into polygon memory 262.
  • the part of primitive data needed for HSR includes vertex locations and spatial derivatives (i.e., dz/dx, dz/dy, dx/dy for edges, etc.).
  • the part of primitive data not needed for HSR includes vertex colors, texture coordinates, color derivatives, etc. If per-fragment lighting is performed in the pipeline, the entire lighting equation is applied to every fragment. But in a deferred shading pipeline, only visible fragments require lighting calculations.
  • the polygon memory may also include vertex normals, vertex eye coordinates, vertex surface tangents, vertex binormals, spatial derivatives of all these attributes, and other per-primitive lighting information.
  • vertex normals vertex eye coordinates
  • vertex surface tangents vertex surface tangents
  • vertex binormals spatial derivatives of all these attributes, and other per-primitive lighting information.
  • HSR process is done on a per-tile basis. Splitting the scene memory 250 into spatial memory 261 and polygon memory 262 has the advantage of reducing total memory bandwidth.
  • the output from setup for incremental render 230 is input to the spatial data separation process 263, which stores all the data needed for HSR into spatial memory 261 and the rest of the data into polygon memory 262.
  • the EHSR process 264 receives primitive spatial data (e.g., vertex screen coordinates, spatial derivatives, etc.) and the part of the pipeline state needed for HSR (including all control bits for the per-fragment testing operations).
  • primitive spatial data e.g., vertex screen coordinates, spatial derivatives, etc.
  • the data matching process 265 matches the vertex state and pipeline state with visible fragments, and tile information is stored in tile buffers 266.
  • the remainder of the pipeline is primarily concerned with the scan out process including sample to/from pixel conversion 267, reading and writing to the frame buffer, double buffered MUX output look-up, and digital to analog (D/A) conversion of the data stored in the frame buffer to the actual analog display device signal values.
  • sample to/from pixel conversion 267 reading and writing to the frame buffer
  • double buffered MUX output look-up double buffered MUX output look-up
  • D/A digital to analog
  • a fifth embodiment of the Deferred Shading Graphics Pipeline (Version 5) (DSGPv ⁇ ), illustrated in Figure 7, exact hidden surface removal is used as in the third embodiment, however, the tiling is added, and a tile sorting procedure is added after data separation, and the read is by tile prior to spatial setup.
  • the polygon memory of the first three embodiments is replaced with a state memory. 4.1.8 Deferred Shading Graphics Pipeline, Sixth Embodiment (Version 6)
  • the pipeline takes data from the host computer's I/O bus, processes it, and sends it to the computer's display.
  • the pipeline is divided into twelve blocks, plus three memory stores and the frame buffer.
  • Figure 9 shows the flow of data through the pipeline 1000. The blocks that make up the pipeline are discussed below.
  • Figure 10 shows a method flow diagram of pipeline 1000 which also illustrates the information that is passed between the blocks.
  • Command Fetch and Decode (CFD) 2000 handles communication with the host computer through the I/O bus. It converts its input into a series of packets, which it passes to the Geometry block. Most of the input stream consists of geometrical data — lines, points, and polygons. The descriptions of these geometrical objects can include colors, surface normals, texture coordinates, etc. The input stream also contains rendering information, such as lighting, blending modes, and buffer functions.
  • Geometry block 3000 handles four major tasks: transforms, decomposition of all polygons into triangles, clipping, and per-vertex lighting calculations needed for Gouraud shading.
  • transforms decomposition of all polygons into triangles
  • clipping clipping
  • per-vertex lighting calculations needed for Gouraud shading.
  • the Geometry block transforms incoming graphics primitives into a uniform coordinate space
  • the DSGP pipeline provides six user- definable clipping planes. After clipping, the Geometry block breaks polygons with more than three vertices into sets of triangles, to simplify processing.
  • the Geometry block calculates the vertex colors that the Fragment block uses to perform the shading.
  • MEX Mode Extraction
  • the "everything else” lights, colors, texture coordinates, etc. — is stored in a special buffer called the Polygon Memory, where it can be retrieved by the Mode Injection block.
  • the Polygon Memory is double buffered, so the Mode Injection block can read data for one frame, while the Mode Extraction block is storing data for the next frame.
  • the mode data stored in the Polygon Memory falls into three major categories: per-frame data (such as lighting), per-primitive data (such as material properties) and per-vertex data (such as color). The Mode Extraction and Mode injection blocks further divide these categories to optimize efficiency.
  • the Mode Extraction block sends the Sort block a packet containing the vertex data and a pointer into the Polygon Memory.
  • the pointer is called the Color Pointer, although it is used to retrieve all sorts of other information in addition to color.
  • the packet also contains fields indicating whether the vertex represents a point, the endpoint of a line, or the corner of a triangle.
  • the vertices are sent in a strict time sequential order, the same order in which they were fed into the pipeline.
  • the packet also specifies whether the current vertex forms the last one in a given primitive (i.e., "completes" the primitive). In the case of triangle strips or fans, and line strips or loops, the vertices are shared between adjacent primitives. In this case, the packets indicate how to identify the other vertices in each primitive.
  • the Sort block 6000 receives vertices from the Mode Extraction block and sorts the resulting points, lines, and triangles by tile. In the double-buffered Sort Memory 7000, it maintains a list of vertices representing the graphic primitives, and a set of Tile Pointer Lists-one list for each tile in the frame.
  • Sort block 6000 receives a vertex that completes a primitive (such as the third vertex in a triangle), it checks to see which tiles the primitive touches. For each tile a primitive touches, the Sort block adds a pointer to the vertex to that tile's Tile Pointer List. When the Sort block has finished sorting all the geometry in a frame, it sends the data to Setup block 8000. Each Sort block output packet represents a complete primitive. Sort sends its output in tile-by- tile order: all of the primitives that touch a given tile, then all of the primitives that touch the next tile, and so on. Note that this means that Sort may send the same primitive many times, once for each tile it touches.
  • the Setup block 8000 calculates spatial derivatives for lines and triangles. Setup block 8000 processes one tile's worth of data, one primitive at a time. When Setup block 8000 is done with a primitive, it sends the data on to the Cull block. The Setup block also breaks stippled lines into separate line segments
  • the Setup block outputs Setup Output Primitive Packets, each of which represents one primitive: a triangle, line segment or point.
  • Cull block 9000 performs the conservative hidden surface removal process.
  • the Cull block 9000 processing is divided into two steps: Magnitude Comparison Content Addressable Memory (MCCAM) Cull, and Subpixel Cull.
  • the Cull block accepts data one tile's worth at a time.
  • the MCCAM Cull discards primitives that are hidden completely by previously processed geometry.
  • the Subpixel Cull takes the remaining primitives (which are partly or entirely visible), and determines the visible fragments.
  • the Subpixel Cull outputs one stamp's worth of fragments at a time, called a Visible Stamp Portion (VSP).
  • VSP Visible Stamp Portion
  • Figure 11 shows an example of how the Cull block produces fragments from a partially obscured triangle. Each pixel in a VSP is divided up into a number of samples to determine how much of the pixel is covered by a given fragment. The Pixel block uses this information when it blends the fragments to produce the final color for the pixel.
  • the Mode Injection block 10000 retrieves mode information —such as colors, material properties, and so on— from the Polygon Memory 5000 and passes it downstream as required. To save bandwidth, the individual downstream blocks cache recently used mode information. The Mode Injection block keeps track of what information is cached downstream, and only sends information as necessary.
  • mode information such as colors, material properties, and so on
  • the Fragment block 11000 primarily performs interpolation. Fragment block 11000 interpolates color values for Gouraud shading, surface normals for Phong shading and texture coordinates for texture mapping. Fragment block 11000 also interpolates surface tangents for use in the bump mapping algorithm, if bump maps are in use. In addition, the Fragment block performs perspective corrected interpolation using barycentric coefficients.
  • Texture block 12000 applies texture maps to the pixel fragments. Texture maps are stored in the Texture Memory 13000. Unlike the other memory elements described previously, the Texture Memory is single-buffered. It is loaded from the host computer's memory using the AGP interface. A single polygon can use up to four textures. Textures are mip-mapped. That is, each texture comprises a series of texture maps at different levels of detail, each map representing the appearance of the texture at a given distance from the eye point. To produce a texture value for a given pixel fragment, the Texture block performs tri-linear interpolation from the texture maps, to approximate the correct level of detail. The Texture block also performs other interpolation methods, such as anisotropic interpolation.
  • the Texture block supplies interpolated texture values (generally as RGBA color values) to the Phong block on a per-fragment basis.
  • Bump maps represent a special kind of texture map. Instead of a color, each texel of a bump map contains a height field gradient. 5.10 Phong (PHG)
  • Phong block 14000 performs Phong shading for each pixel fragment. It uses the material and lighting information supplied by the Mode Injection block, the texture colors from the Texture block, and the surface normal generated by the Fragment block to determine the fragment's apparent color. If bump mapping is in use, the Phong block uses the interpolated height field gradient from the Texture block to perturb the fragment's surface normal before shading.
  • the Pixel block 15000 receives VSPs, where each fragment has an independent color value.
  • the Pixel block performs pixel ownership test, scissor test, alpha test, stencil operations, depth test, blending, dithering and logic operations on each sample in each pixel (see OpenGL ® Spec 1.1 , Section 4.1 , "Per- Fragment Operations," p. 109).
  • the Pixel block When the Pixel block has accumulated a tile's worth of finished pixels, it blends the samples within each pixel (thereby performing antialiasing of pixels) and sends them to the Backend, to be stored in the Frame Buffer 17000.
  • the Pixel block blends the fragments and the background color to produce the final pixels.
  • the Pixel block weights each fragment according to how much of the pixel it covers-or to be more precise, by the number of samples it covers.
  • the Pixel block performs stencil testing, alpha blending, and antialiasing of pixels. When the Pixel block accumulates a tile's worth of finished pixels, it sends them to the Backend, to be stored in the Frame Buffer.
  • the Backend 16000 receives a tile's worth of pixels at a time from the Pixel block, and stores them into the Frame Buffer 17000.
  • the Backend also sends a tile's worth of pixels back to the Pixel block, because specific Frame Buffer values can survive from frame to frame. For example, stencil bit values can be constant over many frames, but can be used in all those frames.
  • the Backend performs 2D drawing and sends the finished frame to the output devices.
  • the Backend provides the interface between the Frame Buffer and the computer monitor and video output.
  • Cull Block Overview Figure 12 illustrates a block diagram of Cull block 9000.
  • the Cull block is responsible for: 1 ) pre- shading hidden surface removal; and 2) breaking down primitive geometry entities (triangles, lines and points) to stamp based geometry entities called Visible Stamp Portions (VSPs).
  • VSPs Visible Stamp Portions
  • the Cull block does, in general, a conservative culling of hidden surfaces.
  • Cull block 9000 does not handle some "fragment operations" such as alpha test and stencil test.
  • Z Cull 9012 can store two depth values per sample, but Z Cull 9012 only stores the attributes of one primitive per sample. Thus, whenever a sample requires blending colors from two pieces of geometry, the
  • Cull block sends the first primitive (using time order) down the pipeline, even though there may be later geometry that hides both pieces of the blended geometry.
  • the Cull block receives input in the form of packets from the Setup block 8000.
  • One type of packet received by the Cull block is a mode packet.
  • Mode packets provide the Cull block control information including the start of a new tile, a new frame, and the end of a frame.
  • Cull block 9000 also receives Setup Output Primitive Packets.
  • the Setup Output Primitive Packets each describe, on a per tile basis, either a triangle, a line or a point.
  • the data field in Setup Output Primitive Packets contain bits to indicate the primitive type (triangle, line or point). The interpretation of the rest of the geometry data field depends upon the primitive type.
  • a non-geometry data field contains the Color Pointer and mode bits that control the culling mode that can be changed on a per primitive bases.
  • Mode packets include mode bits that indicate whether alpha test is on, whether Z buffer write is enabled, whether culling is conservative or accurate, whether depth test is on, whether blending is on, whether a primitive is anti-aliased and other control information.
  • Sort block 6000 bins the incoming geometry entities to tiles.
  • Setup block 8000 pre-processes the primitives to provide more detailed geometric information for the Cull block to do the hidden surface removal.
  • Setup block 8000 pre-calculates the slope value for all the edges, the bounding box of the primitive within the tile, minimum depth value (front most) of the primitive within the tile, and other relevant data.
  • Mode Extraction block 4000 Prior to Sort, Mode Extraction block 4000 has already extracted the color, light, texture and related mode data, the Cull block only gets the mode data that is relevant to the Cull block and a pointer, called Color Pointer, that points to color, light and texture data stored in Polygon Memory 5000.
  • a pointer called Color Pointer
  • the Cull block performs two main functions.
  • the primary function is to remove geometry that is guaranteed to not affect the final results in Frame Buffer 17000 (i.e., a conservative form of hidden surface removal).
  • the second function is to break primitives into units of Visible Stamp Portions (VSP).
  • VSP Visible Stamp Portions
  • a stamp portion is the intersection of a primitive with a given stamp.
  • a VSP is a visible portion of a geometry entity within a stamp.
  • each stamp is comprised of four pixels, and each pixel has four predetermined sample points. Thus each stamp has 16 predetermined sample points.
  • Mode Injection block 10000 is then given by the number and the set of sample points covered by a primitive in a given stamp.
  • the Cull block sends one VSP at a time to the Mode Injection block 10000.
  • the Cull block processes primitives one tile at a time. However, for the current frame, the pipeline is in one of two modes: 1 ) time order mode; or 2) sorted transparency mode.
  • time order mode the time order of vertices and modes are preserved within each tile, and the tile is processed in a single pass through the data. That is, for a given tile, vertices and modes are read in the same order as they are written, but are skipped if they do not affect the current tile.
  • sorted transparency mode the processing of each tile is divided into multiple passes, where, in the first pass, guaranteed opaque geometry is processed (the Sort block only sends non-transparent geometry for this pass). In subsequent passes, potentially transparent geometry is processed (the Sort block repeatedly sends all the transparent geometry for each pass).
  • the time ordering is preserved, and mode data is inserted in its correct time-order location.
  • time order mode when there is only "simple opaque geometry" (i.e. no scissor testing, alpha testing, color testing, stencil testing, blending, or logicop) in a tile, the Cull block will process all the primitives in the tile before dispatching any VSPs to Mode Injection. This is because the Cull block hidden surface removal method can unambiguously determine, for each sample, the single primitive that covers (i.e., colors) that sample.
  • the case of "simple opaque geometry” is a typically infrequent special case.
  • time order mode when the input geometry is not limited to "simple opaque geometry" within a tile, this may cause early dispatch of VSPs (an entire set of VSPs or selected VSPs).
  • each tile is processed in multiple passes (assuming there is at least some transparent geometry in the tile). In each pass, there is no early dispatch of VSPs.
  • a PrimType parameter indicates the primitive type (triangle, line or point).
  • the spatial location of the primitive is done using a "unified description". That is, the packet describes the primitive as a quadrilateral (not screen aligned), and triangles and points are degenerate cases. This "unified description" is described in more detail in the provisional patent application entitled “Graphics Processor with Deferred Shading,” filed August 20, 1998, which is hereby incorporated by reference.
  • the packet includes a color pointer, used by Mode Injection.
  • the packet also includes several mode bits, many of which can change primitive by primitive. The following are considered to be "mode bits", and are input to state machines in Z Cull 9012: CullFlushOverlap,
  • Cull block 9000 receives the following packet types:
  • Cull block 9000 receives a primitive, Cull attempts to eliminate it by querying the Magnitude Comparison Content Addressable Memory (MCCAM) Cull 9002, shown in Figure 12, with the primitive's bounding box. If MCCAM Cull 9002 indicates that a primitive is completely hidden within the tile, then the primitive is eliminated. If MCCAM Cull 9002 cannot reject the primitive completely, it will generate a stamp list, each stamp in the list may contain a portion of the primitive that may be visible. This list of potentially visible stamps is sent to the Stamp Selection Logic 9008 of Cull block 9000. Stamp Selection
  • MCCAM Magnitude Comparison Content Addressable Memory
  • Logic 9008 uses the geometry data of the primitive to determine the set of stamps within each stamp row of the tile that are actually touched by the primitive. Combined with the stamp list produced by MCCAM Cull
  • the Stamp Selection Logic unit dispatches one potentially visible stamp 9006 at a time to the Z Cull block 9012.
  • Each stamp is divided into a grid of 16 by 16 sub-pixels. Each horizontal grid line is called a subraster line.
  • Each of the 16 sample points per stamp has to fall (for antialiased primitives) at the center of one of the 256 possible sub-pixel locations.
  • Each pixel has four sample points within its boundary, as shown with stamp 9212 in Figure 13A.
  • Figure 13B and Figure 13C illustrate the manner in which the Stamp Portion is input into the Z-Cull process and as stored in SPM, respectively.
  • Sample locations within pixels can be made programmable. With programmable sample locations, multiple processing passes can be made with different sample locations thereby increasing the effective number of samples per pixel. For example, four passes could be performed with four different sets of sample locations, thereby increasing the effective number of samples per pixel to fourteen.
  • the display image is divided into tiles to more efficiently render the image.
  • the tile size as a fraction of the display size can be defined based upon the graphics pipeline hardware resources.
  • the process of determining the set of stamps within a stamp row that is touched by a primitive involves calculating the left most and right most positions of the primitive in each subraster line that contains at least one sample point. These left most and right most subraster line positions are referred to as XleftSubS, and XrightSubS, which stands for x left most subraster line for sample i and x right most subraster line for sample i respectively. Samples are numbered from 0 to 15. The determination of XleftSubS, and XrightSubS, is typically called the edge walking process. If a point on an edge (xO, yO) is known, then the value of x1 corresponding to the y position of y1 can easily be determined by:
  • the set of 16 pairs of XleftSubS, and XrightSubS is also sent by the Stamp Selection Logic unit to Z Cull 9012.
  • Z Cull unit 9012 receives one stamp number (or StamplD) at a time.
  • stamp number contains a portion of a primitive that may be visible as determined by MCCAM Cull 9002.
  • XleftSubS, and XrightSubS are used to determine which of the 16 sample points are covered by the primitive.
  • Sample i is covered if Xsample,, the x coordinate value of sample i satisfies:
  • the primitive's z value is computed at that sample point.
  • the current z values and z states for all 16 sample points are read from the Sample Z buffer
  • Each sample point can have a z state of "conservative" or "accurate”.
  • Alpha test, and other tests are performed by pipeline stages after Cull block 9000. Therefore, for example, a primitive that may appear to affect the final color in the frame buffer based on depth test, may in fact be eliminated by alpha test before the depth test is performed, and thus the primitive does not affect the final color in the frame buffer.
  • the Cull block 9000 uses conservative z values.
  • a conservative z value defines the outer limit of a z value for a sample based on the geometry that has been processed up to that point.
  • a conservative z value means that the actual z value is either at that point or at a smaller z value.
  • the conservative z is the maximum z value that the point can have. If the depth test is render if greater than, then the conservative z value is a minimum z value. Conversely, if the depth test is render if less than, then the conservative z value is a maximum z value. For a render if less than depth test, any sample for a given sample location, with a z value less than the conservative z is thus a conservative pass because it is not known at that point in the processes whether it will pass.
  • An accurate z value is a value such that the surface which that z represents is the actual z value of the surface. With an accurate z it is known that the z value represents a surface that is known to be visible and anything in front of it is visible and everything behind it is obscured, at that point in the process.
  • the status of a sample is maintained by a state machine, and as the process continues the status of a sample may switch between accurate and conservative.
  • a single conservative z value is used.
  • two z values are maintained for each sample location, a near z value (Znear) and a far z value (Zfar).
  • the far z value is a conservative z value
  • the near z value is an optimistic z value.
  • Using two z values allows samples to be determined to be accurate again after being labeled as conservative. This improves the efficiency of the pipeline because an accurate z value can be used to eliminate more geometry than a conservative z value. For example, if a sample is received that is subject to alpha test, in the Cull block it is not known whether the sample will be eliminated due to alpha test.
  • the z value may have to be made conservative if the position of the sample subject to alpha test would pass the depth test.
  • the sample that is subject to alpha test is then sent down the pipeline. Since, the sample subject to alpha test is not kept, the z value of the stored sample cannot later be converted back to accurate.
  • the sample subject to alpha test can, depending on its relative position, be stored as the Zfar/Znear sample. Subsequent samples can then be compared with the sample subject to alpha test as well as the second stored sample.
  • the Cull block determines, based on the depth test, that one of the subsequent samples, such as an opaque sample in front of the sample subject to alpha test, renders the sample subject to alpha test not visible, then that subsequent sample can be labeled as accurate.
  • OpenGL ® primitives are processed in groups. The beginning and ending of a group of pimitives are identified by the commands, begin and end respectively.
  • the depth test is defined independently for each group of primitives. The depth test is one component of the pipeline state.
  • Each sample point has a Finite State Machine (FSM) independent of other samples.
  • FSM Finite State Machine
  • the z state combined with the mode bits received by Cull drive the sample FSMs.
  • the sample FSMs control the comparison on a per sample basis between the primitive's z value and the Z Cull 9012 z value. The result of the comparison is used to determine whether the new primitive is visible or hidden at each sample point that the primitive covers.
  • the maximum of the 16 sample points' z value is used to update the MCCAM Cull 9002.
  • a sample's FSM also determines how the Sample Z Buffer in Z Cull 9012 should be updated for that sample, and whether the sample point of the new VSP should be dispatched early. In addition, the sample FSM determines if any old VSP that may contain the sample point should be destroyed or should be dispatched early. For each sample Z Cull 9012 generates four control bits that describe how the sample should be processed, and sends them to the Stamp Portion Mask unit 9014. These per sample control bits are: SendNew, KeepOld, SendOld, and NewVSPMask. If the primitive contains a sample point that is visible, then a NewVSPMask control bit is asserted which causes Stamp Portion Memory (SPM) 9018 to generate a new VSP coverage mask.
  • SPM Stamp Portion Memory
  • the remaining three control bits determine how SPM 9018 updates the VSP coverage mask for the primitive.
  • geometry is spatially sorted on a per-sample basis, and, within each sample, is rendered in either back-to-front or front-to-back order. In either case, only geometry that is determined to be in front of the front-most opaque geometry needs to be send down the pipeline, and this determination is done in Cull 9012.
  • back-to-front sorted transparency mode transparent primitives are rasterized in spatial order starting with the layer closest to the front most opaque layer instead of the regular mode of time order rasterization. Two z values are used for each sample location, Zfar and Znear.
  • the transparent primitives go through Z Cull unit 9012 several times.
  • Sort block 6000 illustrated in Figure 9, sends only the opaque primitives. The z values are updated as described above.
  • the z values for opaque primitives are referred to as being of type Zfar.
  • the opaque VSPs are dispatched.
  • the second time Sort block 6000 only sends the transparent primitives for the tile to Cull block 9000. Initially the Znear portion of the Sample Z Buffer are preset to the smallest z value possible. A sample point with a z value behind Zfar is hidden, but a z value in front of Zfar and behind Znear is closer to the opaque layer and therefore replaces the current Znear's z value. This pass determines the z value of the layer that is closest to the opaque layer. The VSPs representing the closest to opaque layer are dispatched.
  • Z Cull receives the second pass of transparent primitives. This process continues until Z Cull determines that it has processed all possible layers of transparent primitives.
  • Z Cull in sorted transparent mode is also controlled by the sample finite state machines.
  • the number of transparent passes is equal to the number of visible transparent surfaces.
  • the passes can be done as: a) The Opaque Pass (there is only one Opaque Pass) does the following: the front-most opaque geometry is identified (labeled Zfar) and sent down the pipeline. b) The first Transparent Pass does the following: 1) at the beginning of the pass, keep the Zfar value from the Opaque Pass, and set Znear to zero; 2) identifies the back-most transparent surface between Znear (initialized to zero at the start of the pass) and Zfar; 2) determine the new Znear value; and, 3) at the end of the pass, send this back-most transparent surface down the pipeline.
  • the subsequent passes do the following: 1) at the beginning of the pass, set the Zfar value to the Znear value from the last pass, and set Znear to zero; 2) identify the next farthest transparent surface between Znear and Zfar; 3) determine the new
  • the number of transparent passes can be limited to a preselected maximum, even if the number of visible transparent surfaces at a sample is greater.
  • the passes can be done as: a) In the First Opaque Pass (there are two opaque passes, the other one is the Last Opaque Pass), the front-most opaque geometry is identified (labeled Zfar), but this geometry is not sent down the pipeline, because, only the z-value is valuable in this pass. This Zfar value is the boundary between visible transparent layers and hidden transparent layers. This pass is done with the time order mode sample FSM. b) The next pass, the first Transparent Pass, renders the front-most transparent geometry and also counts the number of visible transparencies at each sample location.
  • This pass does the following: 1 ) at the beginning of the pass, set the Znear value to the Zfar value from the last pass, set Zfar to the maximum z-value, and initialize the NumTransp counter in each sample to zero; 2) test all transparent geometry and identify the front-most transparent surface by finding geometry that is in front of both Znear and Zfar; 3) as geometry is processed, determine the new Zfar value, but don't change the Znear value; 4) count the number of visible transparent surfaces by incrementing NumTransp when geometry that is in front of Znear is encountered; and, 5) at the end of the pass, send this front-most transparent surface down the pipeline.
  • this pass is defined in an unusual way, because, at the end, Zfar is nearer than Znear; but this allows the rule, "set the Znear value to the Zfar value from the last pass, and set Zfar to the maximum z-value" to be true for every transparent pass. If this is confusing, the definition of Znear and Zfar can be swapped, but this changes the definition of the second transparent pass.
  • Subsequent Transparent Passes determine progressively farther geometry, and the maximum number of transparent passes is specified by the MaxTranspPasses parameter.
  • Each of these passes does the following: 1 ) at the beginning of the pass, set the Znear value to the Zfar value from the last pass, set Zfar to the maximum z-value, and the NumTransp counter in each sample is not changed; 2) test all transparent geometry and identify the next-front-most transparent surface by finding the front-most geometry that is between Znear and Zfar, but discard all the transparent geometry if all of the visible transparent layers have been found for this sample (i.e., NumTranspPass > NumTransp); 3) as geometry is processed, determine the new Zfar value, but don't change the Znear value; and, 4) at the end of the pass, send this second-most transparent surface down the pipeline.
  • the front-most opaque geometry is again identified, but this time, the geometry is sent down the pipeline.
  • the efficiency of CUL is increased (i.e., fewer fragments sent down the pipeline) in front-to-back sorted transparency mode, especially when there are lots of visible depth complexity for transparent surfaces. Also, this may enhance image quality by allowing the user to discern the front-most N transparencies, rather than all those in front of the front-most opaque surface.
  • the stamp portion memory block 9018 contains the VSP coverage masks for each stamp in the tile. The maximum number of VSPs a stamp can have is 16. The VSP masks should be updated or dispatched early when a new VSP comes in from Z Cull 9012.
  • the Stamp Portion Mask unit performs the mask update or dispatch strictly depending on the SendNew, KeepOld and SendOld control bits.
  • the update should occur at the same time for a maximum of 16 old VSPs in a stamp because a new VSP can potentially modify the coverage mask of all the old VSPs in the stamp.
  • the Stamp Portion Data unit 9016 contains other information associated with a VSP including but not limited to the Color Pointer.
  • the Stamp Portion Data memory also needs to hold the data for all VSPs contained in a tile. Whenever a new VSP is created, its associated data need to be stored in the Stamp Portion Data memory. Also, whenever an old VSP is dispatched, its data need to be retrieved from the Stamp Portion Data memory.
  • FIG. 14 illustrates a detailed block diagram of Cull block 9000.
  • Cull block 9000 is composed of the following components: Input FIFO 9050, MCCAM Cull 9002, Subrasterizer 9052, Column Selection
  • the Setup Output Cull Packet is a packet of mode bits. This packet includes: 1 ) bits for enabling/disabling the MCCAM Cull and Z Cull processes; 2) a bit, CullFlushAII, that causes a flush of all the VSPs from the Cull block; and 3) the bits: AliasPolys, AliasLines, and AliasPoints, which disable antialiasing for the three types of primitives.
  • the Setup Output Begin Frame Packet tells Cull that a new frame is starting. The next packet will be a Sort Output Begin Tile Packet.
  • the Setup Output Begin Frame Packet contains all the per-frame information that is needed throughout the pipeline.
  • the Setup Output End Frame Packet indicates the frame has ended, and that the current tile's input has been completed.
  • the Setup Output Begin Tile Packet tells the Cull block that the current tile has ended and that the processed data should be flushed down the pipeline. Also, at the same time, the Cull block should start to process the new tile's primitives. If a tile is to be repeated due to the pipeline being in sorted transparency mode, then this requires another Setup Output Begin Tile Packet.
  • the Setup Output Tween Packet can only occur between (hence 'tween) frames, which, of course is between tiles. Cull treats this packet as a black box, and just passes it down the pipeline. This packet has only one parameter, TweenData, which is 144 bits. in addition to the mode packets, the Cull block also receives Setup Output Primitive Packets, as illustrated in Figure 15.
  • the Setup Output Primitive Packets each describe, on a per tile basis, either a triangle, a line, or a point.
  • the data field in Setup Output Primitive Packets contain bits to indicate the primitive type (triangle, line, or point). The interpretation of the rest of the geometry data field depends upon the primitive type. If the input packet is a Setup Output Primitive Packet, a PrimType parameter indicates the primitive type (triangle, line or point). The spatial location of the primitive (including derivatives, etc.) is specified using a unified description. That is, the packet describes the primitive as a quadrilateral (non-screen aligned), no matter whether the primitive is a quadrilateral, triagle, or point, and triangles and points are treated as degenerate cases of the quadralateral. The packet includes a color pointer, used by the Mode
  • the packet also includes several mode bits, many of which can change state on a primitive by primitive basis. The following are considered to be "mode bits", and are input to state machines in Z Cull
  • Input FIFO Figure 16 illustrates a flow chart of a conservative hidden surface removal method using the Cull block 9000 components shown in the Figure 14 detailed block diagram.
  • Input FIFO unit 9050 interfaces with the Setup block 8000.
  • Input FIFO 9050 receives data packets from Setup and stores each packet in a queue, step 9160.
  • the number of FIFO memory locations needed is between about sixteen and about 32, in one embodiment the depth is assumed to be sixteen.
  • the MCCAM Cull unit 9002 uses an MCCAM array 9003 to perform a spatial query on a primitive's bounding box to determine the set of stamps within the bounding box that may be visible.
  • Figure 17A illustrates a sample tile including a primitive
  • MCCAM Cull 9002 uses ZMin to perform z comparisons. MCCAM Cull 9002 stores the maximum z value per stamp of all the primitives that have been processed.
  • MCCAM Cull 9002 compares in parallel ZMin for the primitive with all the ZMaxes for every stamp. Based on this comparison, MCCAM Cull determines (a) whether the whole primitive is hidden, based on all the stamps inside the simple bounding box; or (b) what stamps are potentially visible in that bounding box, step 9164.
  • Figure 17B shows the largest z values (ZMax) for each stamp in the tile.
  • Figure 17C shows the results of the comparison.
  • Stamps where ZMin ⁇ ZMax are indicated with a one, step 9166. These are the potentially visible stamps.
  • MCCAM Cull also identifies each row which has a stamp with ZMin ⁇ ZMax, step 9168. These are the rows that the Stamp Selection Logic unit 9008 needs to process. Stamp Selection Logic unit 9008 skips the rows that are identified with a zero.
  • MCCAM Cull can process one primitive per cycle from the input FIFO 9050. Read operations from the FIFO occur when the FIFO is not empty and either the last primitive removed is completely hidden as determined by MCCAM Cull or the last primitive is being processed by the Subrasterizer unit 9052. In other words, MCCAM Cull does not "work ahead" of the Subrasterizer. Rather, MCCAM Cull only gets the next primitive that the Subrasterizer needs to process, and then waits.
  • Cull block 9000 does not include an MCCAM Cull unit 9002.
  • the Stamp Selection Logic unit 9008 processes all of the rows.
  • Subrasterizer 9052 is the unit that does the edge walking (actually, the computation is not iterative, as the term "walking" would imply). Each cycle, Subrasterizer 9052 obtains a packet from MCCAM Cull 9002. One type of packet received by the Cull block is the Setup Output Primitive Packet, illustrated in
  • Setup Output Primitive Packets include row numbers and row masks generated by MCCAM Cull 9002 which indicate the potentially visible stamps in each row.
  • Subrasterizer 9052 also receives the vertex and slope data it needs to compute the the left most and right most positions of the primitive in each subraster line that contains at least one sample point, XleftSubS, and XrightSubS, .
  • Subrasterizer 9052 decodes the PrimitiveType field in the Setup Output Primitive Packet to determine if a primitive is a triangle, a line or a point, based on this information
  • Subrasterizer 9052 determines whether the primitive is antialiased. Referring to Figure 18, for each row of stamps that MCCAM Cull indicates is potentially visible (using the row selection bits 9271 ), Subrasterizer 9052 simultaneously computes the XleftSub, and
  • XrightSub for each of the sample points in the stamp, in a preferred embodiment there are 16 samples per stamp, step 9170.
  • Each pair of XleftSub, and XrightSub define a set of stamps in the row that is touched by the primitive, which are referred to as a sample row mask.
  • Figure 19 illustrates a set of XleftSub, and XrightSub,.
  • each stamp in the potentially visible rows that is touched by the primitive is indicated by setting the corresponding stamp coverage bit 9272 to a one ('1"), as shown in tile 9270.
  • Subrasterizer 9052 logically OR's the sixteen row masks to get the set of stamps touched by the primitive. Subraster 9052 then ANDs the touched stamps with the stamp selection bits 9278, as shown in tile 9276, to form one touched stamp list, which is shown in tile 9280, step 9172.
  • the Subrasterizer passes a request to MCCAM Cull for each stamp row, and receives a potentially visible stamp list from MCCAM Cull.
  • the visible stamp list is combined with the touched stamp list, to determine the final potentially visible stamp set in a stamp row, step 9174. For each row, the visible stamp set is sent to the Column Selection block 9054 of Stamp Selection Logic unit 9008.
  • the Subrasterizer can process one row of stamps per cycle.
  • the Subrasterizer takes more than one cycle to process the primitive and therefore will request MCCAM to stall the removal of primitives from the Input FIFO.
  • the Subrasterizer itself can be stalled if a request is made by the Column Selection unit.
  • Figure 20 illustrates a stamp 9291 , containing four pixels 9292, 9293, 9294 and 9295. Each pixel is divided into 8 x 8 subraster grid.
  • the grid shown in Figure 20 shows grid lines located at the mid-point of each subraster step.
  • samples are located at the center of a unit grid, as illustrated by samples 0 - 15 in Figure 20 designated by the circled numbers (e.g. ®). Placing the samples in this manner, off grid by one half of a subraster step, avoids the complications of visibility rules that apply to samples on the edge of a polygon.
  • polygons can be defined to go to the edge of a subraster line or pixel boundary, but samples are restricted to positions off of the subraster grid.
  • two samples in adjacent pixels are placed on the same subraster. This simplifies sample processing by reducing the number of XleftSub, and XrightSub, by a factor of two.
  • the Column Selection unit 9054 tells the Z Cull unit 9012 which stamp to process in each clock cycle. If a stamp row contains more than one potentially visible stamp, the Column Selection unit requests that the Subrasterizer stall.
  • the Z Cull unit 9012 contains the Sample Z Buffer unit 9055 and Z Cull Sample State Machines 9057, shown in Figure 14.
  • the Sample Z Buffer unit 9055 stores all the data for each sample in a tile, including the z value for each sample, and all the sample FSM state bits.
  • Z Cull unit 9012 accesses the z values for all 16 sample points in a stamp in parallel and also computes the new primitive's z values at those sample points in parallel.
  • Z Cull unit 9012 determines whether a primitive covers a particular sample point i by comparing the sample point x coordinate, Xsample,, with the XlefS ⁇ b and XrightSub values computed by the
  • the Z Cull unit 9012 then computes the z value of the primitive at those sample points, step 9180, and compares the resulting z values to the corresponding z values stored in the Sample Z Buffer for that stamp, step 9182. Generally if the sample point z value is less than the z value in the Z Buffer then the sample point is considered to be visible.
  • an API can allow programmers to specify the comparison function (>, ⁇ , ⁇ , ⁇ , always, never). Also, the z comparison can be affected by whether alpha test or blending is turned on, and whether the pipeline is in sorted transparency mode.
  • the Z Cull Sample State Machines 9057 includes a per-sample FSM for each sample in a stamp.
  • each stamp consists of 16 samples
  • the sixteen Z Cull Sample State Machines output the control bits: KeepOld, SendOld, NewVSPMask, and SendNew, to indicate how a sample is to be processed, step 9186.
  • the set of NewVSPMask bits (16 of them) constitute a new stamp portion (SP) coverage mask, step 9188.
  • the new stamp portion is dispatched to the New VSP Queue. In the event that the primitive is not visible at all in the stamp (all NewVSPMask bits are FALSE), then nothing is sent to the New VSP Queue. If more than one sample may affect the final sample position final value, then the stamp portions containing a sample for the sample position are early dispatched, step 9192. All of the control bits for the 16 samples in a stamp are provided to Stamp Portion Memory 9018 in parallel.
  • Samples are sent down the pipeline in VSPs, e.g. as part of a group comprising all of the currenlty visible samples in a stamp.
  • VSPs e.g. as part of a group comprising all of the currenlty visible samples in a stamp.
  • one sample within a stamp is dispatched (either early dispatch or end-of- tile dispatch)
  • other samples within the same stamp and the same primitive are also dispatched as a VSP. While this causes more samples to be sent down the pipeline, it generally causes a net decrease in the amount of color computation. This is due to the spatial coherence within a pixel (i.e., samples within the same pixel tend to be either visible together or hidden together) and a tendency for the edges of polygons with alpha test, color test, stencil test, and/or alpha blending to potentially split otherwise spatially coherent stamps.
  • Figures 21A - 21D illustrate an example of the operation of an embodiment of Z Cull 9012.
  • primitive 9312 is the first primitive in tile 9310.
  • Z Cull 9012 therefore updates all the z values touched by the primitive and stores 35 stamp portions into Stamp Portion Memory 9018.
  • Figure 21 B a second primitive 9322 is added to tile 9310.
  • Primitive 9322 has lower z values than primitive 9312.
  • Z-Cull 9012 processes the 27 stamps touched by primitive 9322.
  • Figure 21C illustrates the 54 stamp portions stored in Stamp Portion Memory 9018 after primitive 9322 is processed.
  • the 54 stamp portions are the sum of the stamps touched by primitives 9312 and 9322 minus eight stamp portions from primitive
  • Region 9332 in Figure 21 D indicates the eight stamp portions that are removed, which are the stamp portions wherein the entire component of the stamp portion touched by primitive 9312 is also touched by primitive 9322 which has lesser Z values.
  • Z Cull 9012 maintains one z value for each sample, as well as various state bits.
  • Z Cull 9012 maintains two z values for each sample, the second z value improves the efficiency of the conservative hidden surface removal process.
  • Z Cull 9012 controls Stamp
  • Stamp Portion Memory 9018 can maintain 16 stamp portions per stamp, for a total of 256 stamp portions per tile.
  • Z Cull 9012 outputs the four bit control signal (SendNew, KeepOld and SendOld and NewVSPMask) to Stamp Portion Memory 9018 that controls how the sample is processed.
  • KeepOld indicates that the corresponding sample in Stamp Portion Memory 9018 is not invalidated. That is, if the sample is part of a stamp portion in Stamp Portion Memory 9018, it is not discarded.
  • SendOld is the early dispatch indicator.
  • Stamp Portion Memory 9018 If the sample corresponding to a SendOld bit belongs to a stamp portion in Stamp Portion Memory 9018, then this stamp portion is sent down the pipeline. SendOld is only asserted when KeepOld is asserted.
  • NewVSPMask is asserted, when the Z Cull 9012 process determines this sample is visible (at that point in the processing) and a new stamp portion needs to be created for the new primitive, which is done by Stamp
  • SendNew is asserted when the Z Cull 9012 process determines the sample is visible (at that point in the processing) and needs to be sent down the pipeline.
  • SendNew causes an early dispatch of a stamp portion in the new primitive.
  • Figure 22 illustrates an example of how samples are processed by Z Cull 9012.
  • Primitive 9352 is processed in tile 9350 before primitive 9354.
  • Primitive 9354 has lesser z values than primitive 9352 and is therefore in front of primitive 9352.
  • Z Cull 9012 sets the KeepOld control bits to zero, and the NewVSPMask control bits to one.
  • Figures 23A - 23D illustrate an example of early dispatch.
  • Early dispatch is the sending of geometry down the pipeline before all geometry in the tile has been processed. In sorted transparency mode early dispatch is not used.
  • First a single primitive 9372, illustrated in Figure 23A is processed in tile 9370.
  • Primitive 9370 touches 35 stamps, and these are stored in Stamp Portion Memory 9018.
  • a second primitive, 9382, with lesser z values is then added with the mode bit DoABIend asserted.
  • the DoABIend mode bit indicates that the colors from the overlapping stamp portions should be blended.
  • Z Cull 9012 then processes the 27 stamps touched by primitive 9382.
  • Z Cull 9012 can be designed so that samples from up to N primitives can be stored for each stamp.
  • FIG. 23C illustrates the stamp portions in Stamp Portion Memory 9018 after primitive 9382 is processed.
  • Figure 23D illustrates the 20 visible stamp portions touched by region 9374 that are dispatched early from primitive 9372 because the stamp portion z values were replaced by the lesser z values from primitive 9382.
  • Stamp 9390 includes part of primitive 9382 and part of primitive 9372, both of which are shown in Figure 23B.
  • the samples in region 9392 all are touched by primitive 9382 which has lesser z values than primitive 9372. Therefore, for these seven samples Z Cull 9012 outputs the control signal SendOld.
  • Z Cull 9012 determines that one sample in a stamp should be sent down the pipeline then Z Cull 9012 sends all of the samples in that stamp down the pipeline so as to preserve spatial coherency. This is also minimizes the number of fragments that are sent down the pipeline.
  • this approach is applied at a pixel level, wherein if Z Cull 9012 determines that any sample in a pixel should be sent down the pipeline all of the samples in the pixel are sent down the pipeline.
  • stamp Portion Memory 9018 In a cull process where everything in a scene is an opaque surface, after all the surfaces have been processed, only the stamp portions that are visible are left in Stamp Portion Memory 9018. The known visible stamp portions are then sent down the pipeline. However, when an early dispatch occurs, the early dispatch stamp portions are sent down the pipeline right away.
  • a reference called Zref is generated. In one embodiment, the Zref is placed at the center of the stamp. The values dzJdx. and ⁇ z/ ⁇ y at the Zref point are also computed.
  • Pixel block 15000 does a final z test. As part of the final z test, Pixel block 15000 re-computes the exactly equivalent z values for each sample using the Zref value and the ⁇ z dx and dzldy values using the equation:
  • Z Cull 9012 performs the same computations that Pixel block does to determine the z value for each stamp so as to avoid introducing any artifacts.
  • a small number of bits can be used to express the delta x and delta y values, since the distances are only fractions of a pixel. For example, in one embodiment a 24 bit derivative and 4 bit delta values are used.
  • MCCAM Update unit 9059 determines the maximum of the sixteen updated z values for the sixteen sample points in each stamp and sends it to the MCCAM Cull unit to update the MCCAM array 9003.
  • Z Cull unit 9012 Each clock cycle, Z Cull unit 9012 generates the four sets of four control bits (KeepOld, SendOld, NewVSPMask, and SendNew) per stamp portion.
  • Z Cull 9012 processes one stamp per primitive per cycle, but not all of the stamps processed are visible, only the Visible Stamp Portions (VSPs) are sent into New VSP Queue 9058.
  • the input rate to New VSP Queue 9058 is therefore variable.
  • the SPM Mask and Valid unit 9060 can store one new stamp portion every clock cycle. However, the SPM Mask and Valid unit 9060 requires multiple clocks for a new stamp portion when early dispatch of VSPs occurs.
  • New VSP Queue 9058 stores the new stamp portions, thus allowing Z Cull 9012 to proceed without stalling.
  • One new VSP may cause the dispatch of up to 16 old VSPs, so the removal rate from the New VSP Queue is also variable.
  • New VSP Queue 9058 is only used with early dispatches. The SPM Mask and
  • the New VSP Queue ensures stamp portions are available for Z Cull 9012 when an early dispatch involves more than one VSP. Based upon performance analysis, typically about 450 stamps are expected to be touched in a tile.
  • the depth complexity of a scene refers to the average number of times a pixel in the scene needs to be rendered. With a depth complexity of two, 225
  • VSPs would be expected to be provided as output from Z Cull 9012 per tile. Therefore on average about four VSPs are expected per stamp.
  • a triangle with blend turned on covering a 50 pixel area can touch on average three tiles, and the number of stamps it touches within a tile should be less than eight. Therefore, in one embodiment, the New VSP Queue depth is set to be 32.
  • New VSP Queue 9058 The link between Z Cull unit 9012 and Stamp Portion Memory 9018 through New VSP Queue 9058 is unidirectional. By avoiding using a feedback loop New VSP Queue 9058 is able to process samples in each cycle.
  • the active Stamp Portion Memory (SPM) Mask and Valid unit 9060 stores the VSP coverage masks for the tile. Each VSP entry includes a valid bit to indicate if there is a valid VSP stored there. The valid bits for the VSPs are stored in a separate memory.
  • the Stamp Portion Memory Mask and Valid unit 9060 is doubie buffered (i.e. there are two copies 9060 and 9062) as shown in Figure 14.
  • the Memory Mask and Valid Active State unit 9060 contains VSPs for the current tile while the Memory Mask and Valid Dispatch
  • State unit page 9062 contains VSPs from the previous tile (currently being dispatched).
  • the active state SPM Mask and Valid unit 9060 updates the VSP Mask for the VSPs that already exist in its mask memory and adds the new VSP to the memory content.
  • the active state SPM Mask and Valid unit dispatches VSPs through the active SPM Data unit 9064 to the dispatch queue. The operations performed in the mask update or early dispatch are controlled by the KeepOld, SendOld, SendNew and
  • the SendOld and SendNew mask bits are off.
  • VSP coverage masks are mutually exclusive, therefore if a new VSP has a particular coverage mask bit turned on, the corresponding bit for all the previously processed VSPs in the stamp have to be turned off.
  • the state transition from active to dispatch and vice versa is controlled by mode packets.
  • Receiving a packet signaling the end of a tile (Begin Tile, End Frame, Buffer Clear, or Cull Packet with CullFlushAII set to TRUE) causes the active state Stamp Portion Memory to switch over to dispatch state and vice versa.
  • the page in dispatch state cycles through each stamp and sends all VSPs to the SPM Data unit, which forwards them to the dispatch queue.
  • the Stamp Portion Memory Mask and Valid unit 9060 is triple buffered.
  • the active Stamp Portion Memory Data unit 9064 stores the Zstamp, dz/dx, dz/dy and the Color Pointer for every VSP in the tile.
  • the Stamp Portion Memory Data unit is also double buffered.
  • the Mask and Valid unit 9060 sends new VSP information to the SPM Data unit 9064.
  • the VSP information includes control signals that instruct the SPM Data unit 9064 to either send the new VSP or save the new
  • the SPM Mask and Valid unit control signals also determine which location among the 16 possible slots the new VSP should occupy.
  • the SPM Data unit also gets a list of old VSP locations and the associated VSP Masks that need early dispatch. The SPM Data unit first checks to see if there are any old VSPs that need to be dispatched, If the SPM Data unit finds any, it will read the VSP data from its memory, merge the VSP data with the VSP Mask sent from the SPM Mask and Valid unit, and put the old VSPs into the dispatch queue.
  • the SPM Data unit then checks if the new VSP should also be sent, and if it is affirmative, then it passes the new VSP data to the dispatch queue 9068. If the new VSP should not be sent, then the SPM Data unit writes the new VSP data into its memory.
  • the Dispatch Queue and Dispatch Logic The Dispatch Logic unit 9072 sends one entry's worth of data at a time from one of the two SPM dispatch queues 9068, 9070 to the Mode Injection unit 10000.
  • the Dispatch Logic unit 9072 requests dispatch from the dispatch state SPM unit first. After the dispatch state SPM unit has exhausted all of its
  • the Dispatch Logic unit 9072 requests dispatch from the active state SPM dispatch queue.
  • Alpha Test compares the alpha value of a given pixel to an alpha reference value.
  • the alpha reference value is often used to indicate the transparency value of a pixel.
  • the type of comparison may be specified, so that for example the comparison may be a greater-than operation, a less-than operation, or other arithmetic, algebraic, or logical comparison, and so forth. If the comparison is a greater-than operation, then a pixel's alpha value has to be greater than the reference to pass the alpha test. For instance, if a pixel's alpha value is 0.9, the reference alpha is 0.8, and the comparison is greater-than, then that pixel passes the alpha test. Any pixel not passing the alpha test is discarded.
  • Alpha test is a per-fragment operation and in a preferred embodiment is performed by the Pixel block after all of the fragment coloring calculations, lighting operations and shading operations are completed.
  • Figure 25 illustrates an example of processing samples with alpha test with a CHSR method. This diagram illustrates the rendering of six primitives (Primitives A, B, C, D, E, and F) at different z coordinate locations for a particular sample, rendered in the following order (starting with a "depth clear” and with "depth test” set to less-than): primitives A, B, and C (with "alpha test” disabled); primitive D (with "alpha test” enabled); and primitives E and F (with “alpha test” disabled).
  • Step 1 The depth clear causes the following result in each sample finite state machine: 1 ) z values are initialized to the maximum value; 2) primitive information is cleared; and 3) sample state bits are set to indicate the z value is accurate.
  • Step 2 When primitive A is processed by the sample FSM, the primitive is kept (i.e., it becomes the current best guess for the visible surface), and this causes the sample FSM to store: 1 ) the z value z A as the "near" z value; 2) primitive information needed to color primitive A; and 3) the z value (z A ) is labeled as accurate.
  • Step 3 When primitive B is processed by the sample FSM, the primitive is kept (its z value is less- than that of primitive A), and this causes the sample FSM to store: 1 ) the z value z B as the "near" z value (z A is discarded); 2) primitive information needed to color primitive B (primitive A's information is discarded); and 3) the z value (z B ) is labeled as accurate.
  • Step 4 When primitive C is processed by the sample FSM the primitive is discarded (i.e., it is obscured by the current best guess for the visible surface, primitive B), and the sample FSM data is not changed.
  • Step 5 When primitive D (which has alpha test enabled) is processed by the sample FSM, the primitive's visibility cannot be determined because it is closer than primitive B and because its alpha value is unknown at the time the sample FSM operates. Because a decision cannot be made as to which primitive would end up being visible (either primitive B or primitive D) primitive B is early dispatched down the pipeline (to have its colors generated) and primitive D is kept.
  • the sample FSM stores: 1 ) the "near" z value is z D and the "far” z value is z & 2) primitive information needed to color primitive D (primitive B's information has undergone early dispatch); and 3) the z values are labeled as conservative (because both a near and far are being maintained).
  • the sample FSM can determine that a piece of geometry closer than z D obscures previous geometry, geometry farther than z B is obscured, and geometry between z D and z B is indeterminate and must be assumed to be visible (hence a conservative assumption is made).
  • the sample FSM method considers the depth value of the stored primitive information to be the near depth value.
  • Step 6 When primitive E (which has alpha test disabled) is processed by the sample FSM, the primitive's visibility cannot be determined because it is between the near and far z values (i.e., between z D and z B ). However, primitive E is not sent down the pipeline at this time because it could result in the primitives reaching the z buffered blend (part of the Pixel block in a preferred embodiment) out of correct time order. Therefore, primitive D is sent down the pipeline to preserve the time ordering.
  • the sample FSM stores: 1 ) the "near" z value is z D and the "far” z value is z B (note these have not changed, and z E is not kept); 2) primitive information needed to color primitive E (primitive D's information has undergone early dispatch); and 3) the z values are labeled as conservative
  • Step 7 When primitive F is processed by the sample FSM, the primitive is kept (its z value is less- than that of the near z value), and this causes the sample FSM to store: 1) the z value z F as the "near" z value (z D and z B are discarded); 2) primitive information needed to color primitive F (primitive E's information is discarded); and 3) the z value (z F ) is labeled as accurate.
  • Step 8 When all the geometry that touches the tile has been processed (or, in the case there are no tiles, when all the geometry in the frame has been processed), any valid primitive information is sent down the pipeline. In this case, primitive F's information is sent.
  • 7.1.0.1 Stencil Test In OpenGL ® stencil test conditionally discards a fragment based on the outcome of a comparison between a value stored in a stencil buffer at location (x w , y w ) and a reference value.
  • Several stencil comparison functions are permitted such that whether the stencil test passes can depend upon whether the reference value is less than, less than or equal to, equal to, greater than or equal to, greater than, or not equal to the masked stored value in the stencil buffer.
  • the stencil test fails, the incoming fragment is discarded.
  • the reference value and the comparison value can have multiple bits, typically 8 bits so that 256 different values may be represented.
  • a tag having the stencil bits is also written into the frame buffer. These stencil bits are part of the pipeline state.
  • the type of stencil test to perform can be specified at the time the geometry is rendered.
  • the stencil bits are used to implement various filtering, masking or stenciling operations, to generate, for example, effects such as shadows. If a particular fragment ends up affecting a particular pixel in the frame buffer, then the stencil bits can be written to the frame buffer along with the pixel information.
  • the stencil values are not available to the CHSR method performed in the Cull block. While it is possible for the stencil values to be transferred from the Frame Buffer for use in the CHSR process, this would generally require a long latency path that would reduce performance.
  • APIs such as OpenGL ® the stencil test is performed after alpha test, and the results of alpha test are not known to the
  • the stencil test is defined for a group of primitives.
  • control software sets the CullFlushAII bit in the corresponding Setup Output Cull Packet.
  • CullFlushAII causes all of the VSPs from the Cull block to be sent to Pixel block 15000, and clears the z values in Stamp Portion Memory 9018. This "flushing" is needed because changing the stencil reference value effectively changes the "visibility rules" in the z buffered blend (or Pixel block).
  • Pixel block 15000 compares the stencil values of the samples for a given sample location and determines which samples affect the final frame buffer color based on the stencil test. For example, for one group of samples corresponding to a sample location, the stencil test may be render if the stencil bit is equal to one. Pixel block 15000 then discards each of the samples for that sample in this group that have a stencil bit value not equal to one.
  • Figure 26 has two primitives (primitives A and C) covering four particular samples (with corresponding sample FSMs, referred to as SFSM0 through SFSM3) and an additional primitive (primitive B) covering two of those four samples.
  • the three primitives are rendered in the following order (starting with a depth clear and with depth test set to less-than): primitive A (with stencil test disabled); primitive B (with stencil test enabled and StencilOp set to "REPLACE", see OpenGL ® specification); and primitive C (with stencil test disabled).
  • Step 1 The depth clear causes the following in each of the four sample FSMs in this example: 1 ) z values are initialized to the maximum value; 2) primitive information is cleared; and 3) sample state bits are set to indicate the z value is accurate.
  • Step 2 When primitive A is processed by each sample FSM, the primitive is kept (i.e., it becomes the current best guess for the visible surface), and this causes the four sample FSMs to store: 1 ) their corresponding z values (either z A0 , z A1 , z ⁇ , or z A3 respectively) as the "near" z value; 2) primitive information needed to color primitive A; and 3) the z values in each sample FSM are labeled as accurate.
  • Step 3 When primitive B is processed by the sample FSMs, only samples 1 and 2 are affected, causing SFSMO and SFSM3 to be unaffected and causing SFSM1 and SFSM2 to be updated as follows:
  • Step 4 When primitive C is processed by each sample FSM, the primitive is kept, but the sample
  • FSMs do not all handle the primitive the same way.
  • SFSMO and SFSM3 the state is updated as: 1 ) z co and z C3 become the "near" z values fa Q and Z 3 are discarded); 2) primitive information needed to color primitive C (primitive A's information is discarded); and 3) the z values are labeled as accurate.
  • SFSM1 and SFSM2 the state is updated as: 1) z ⁇ and z ⁇ become the "far" z values (the near z values are kept);
  • Alpha blending is used to combine the colors of two primitives into one color. However, the primitives are still subject to the depth test for the updating of the z values. The amount of color contribution from each of the samples depends upon the transparency values, referred to as the alpha value, of the samples.
  • the blend is performed according to the equation
  • C is the resultant color
  • C s is the source color for an incoming primitive sample
  • ⁇ s is the alpha value of the incoming primitive sample
  • C d is the destination color at the corresponding frame buffer location.
  • Alpha values are defined at the vertices of primitives, and alpha values for samples are interpolated from the values at the vertices.
  • Step 1 The depth clear causes the following in each CHSR sample FSM: 1 ) z values are initialized to the maximum value; 2) primitive information is cleared; and 3) sample state bits are set to indicate the z value is accurate.
  • Step 2 When primitive A is processed by the sample FSM, the primitive is kept (i.e., it becomes the current best guess for the visible surface), and this causes the sample FSM to store: 1 ) the z value z A as the "near" z value; 2) primitive information needed to color primitive A; and 3) the z value is labeled as accurate.
  • Step 3 When primitive B is processed by the sample FSM, the primitive is kept (because its z value is less-than that of primitive A), and this causes the sample FSM to store: 1 ) the z value z B as the "near" z value (z A is discarded); 2) primitive information needed to color primitive B (primitive A's information is sent down the pipeline); and 3) the z value (z B ) is labeled as accurate.
  • Primitive A is sent down the pipeline because, at this point in the rendering process, the color of primitive B is to be blended with primitive A. This preserves the time order of the primitives as they are sent down the pipeline.
  • Step 4 When primitive C is processed by the sample FSM, the primitive is discarded (i.e., it is obscured by the current best guess for the visible surface, primitive B), and the sample FSM data is not changed. Note that if primitives B and C need to be rendered as transparent surfaces, then primitive C should not be hidden by primitive B. This could be accomplished by turning off the depth mask while primitive B is being rendered, but for transparency blending to be correct, the surfaces should be blended in either front-to-back or back-to-front order.
  • the depth mask (see OpenGL ® specification) is disabled, writing to the depth buffer (i.e., saving z values) is not performed; however, the depth test is still performed.
  • the depth mask is disabled for primitive B, then the value z B is not saved in the sample FSM. Subsequently, primitive C would then be considered visible because its z value would be compared to z A .
  • FIG. 28A illustrates part of a Spatial Packet containing three control bits: DoAlphaTest, DoABIend and Transparent.
  • the Transparent bit is set by the Geometry block 3000 and is normally only used in sorted transparency mode. When the Transparent bit is reset the corresponding primitive is only processed in passes for opaque primitives. When the Transparent bit is set the corresponding primitive is only processed in passes for transparent primitives. The Transparent bit is generated in the Geometry block
  • the DoAlphaTest control bit controls whether Alpha test is performed on the samples in the primitive. When the DoAlphaTest control bit is set to a one it means that downstream from Cull block 9000 an alpha test will be performed on each fragment.
  • DoAlphaTest can then be set to zero which indicates to Cull block 9000 that since all the samples are guaranteed to pass alpha test, it can process the samples as if they were not subject to alpha test. Observe that in an embodiment where one z value is stored, a sample being subject to alpha test can cause the stored sample to be made conservative. Therefore, DoAlphaTest being zero allows Cull to identify more samples as accurate and thereby eliminate more samples.
  • a detailed description of the control of the DoAlphaTest control bit is provided in the provisional patent application entitled "Graphics Processor with Deferred Shading," filed August 20, 1998, which is incorporated by reference.
  • the DoABIend control bit generated by the Geometry block 3000, indicates whether a primitive is subject to blending. Blending combines the color values of two samples.
  • the Geometry block 3000 checks the alpha values at each vertex. If, given the alpha values, the BlendEquation and the BlendFunc pipeline state information is defined such that the frame buffer color values cannot affect the final color, then blending is turned off for that primitive using the DoABIend control bit. Observe that if blending was always on, and all primitives were treated as transparent, then a hidden surface removal process before lighting and shading might not not remove any geometry.
  • the following describes the method for evaluating texture data to determine whether blending can be turned off for a render if less than depth test.
  • a render if less than depth test if there are two opaque primitives at the same location, the primitive that is in front is rendered.
  • the present invention can also be used with a render if greater than depth test. Blending is turned off when a primitive is opaque and therefore no geometry behind the primitive will contribute to the corresponding final colors in the frame buffer. Whether a primitive is opaque is determined conservatively in that if there is any uncertainty as to whether the final frame buffer colors will be a blend of the current primitive and other primitives with greater z values, then the primitive is treated as transparent.
  • the first texture type is RGB texture.
  • RGB texture each texel (the equivalent of a pixel in texture space) is defined by a red color component value "R,” a green color component value "G,” and a blue color component value "B.” There are no alpha values in this first texture type.
  • the second texture type describes each texel by R, G and B values as well as by an alpha value.
  • the texture data comprise the values of the R, G, B and alpha components.
  • the texture environment defines how to determine the final color of a pixel based on the relevant texture data and properties of the primitive. For example, the texture environment may define the type of interpolation that is used, as well as the lighting equation and when each operation is performed.
  • Figure 28B illustrates how the alpha values are evaluated to set the DoABIend control bit.
  • Alpha mode register stores the Transparent bits for each of the three vertices of a triangular primitive.
  • Transparent bit defines whether the corresponding vertex is transparent indicated by a one, or opaque indicated by a zero. If all three of the vertices are opaque then blending is turned off, otherwise blending is on.
  • Logic block implements this blending control function. When the AlphaAIIOne control signal is asserted and all three of the transparent bits in the alpha mode register are equal to one, logic block sets
  • DoABIend to a zero to turn off blending.
  • the alpha value can also be inverted so that an alpha value of zero indicates that a vertex is opaque. Therefore, in this mode of operation, when the AlphaAIIZero control signal is asserted and all three of the transparent bits are zero, the logic block sets DoABIend to a zero ("0") to turn off blending.
  • the graphics pipeline operates in either time order mode or in sorted transparency mode.
  • sorted transparency mode the process of reading geometry from a tile is divided into multiple passes.
  • the Sort block outputs guaranteed opaque geometry, and in subsequent passes the Sort block outputs potentially transparent geometry.
  • the time ordering is preserved, and mode data is inserted into its correct time-order location.
  • Sorted transparency mode can be performed in either back-to-front or front-to-back order.
  • the sorted transparency method is performed jointly by the Sort block and the Cull block.
  • back-to-front sorted transparency modes a pixel color is determined by first rendering the front most opaque surface at the sample location. In the next pass the farthest transparent surface, that is in front of the opaque surface is rendered. In the subsequent pass the next farthest transparent surface is rendered, and this process is repeated until all of the samples at the sample location have been rendered or when a predetermined maximum number of samples have been rendered for the sample location.
  • the Sort block sends the opaque primitives.
  • Cull block 9000 stores the z values for the opaque primitive samples in MCCAM array 9003 (shown in Figure 15) (step 2901 ).
  • the Sort block sends transparent primitives to the Cull block in the second and subsequent passes.
  • MCCAM array 9003 and Sample Z Buffer 9055 each store two z values (Zfar and Znear) for each corresponding sample.
  • the Zfar value is the z value of the closest opaque sample.
  • the Znear value is the z value of the sample nearest to, and less than, the z value of the opaque layer.
  • One embodiment includes two MCCAM arrays 9003 and two Sample Z Buffers 9055 so as to store the Zfar and Znear values in separate units.
  • First the z values for the front-most non-transparent samples are stored in the MCCAM array 9003 (step 2902).
  • the front-most non-transparent samples are then dispatched down the pipeline to be rendered (step 2903).
  • a flag bit in every pointer indicates whether the corresponding geometry is transparent or non-transparent.
  • the Znear values for each sample are reset to zero (step 2904) in preparation for the next pass.
  • the z value for each sample point in the current primitive is compared with both the Zfar and the Znear values for that sample point. If the z value is larger than Znear but smaller than Zfar, then the sample is closer to the opaque layer and its z value replaces the current Znear value.
  • the samples corresponding to the new Znear values are then dispatched down the pipeline to be rendered (step 2907), and Zfar for each such sample is set to the value of Znear (step 2908). This process is then repeated in the next pass.
  • Cull block 9000 detects that it has finished processing a tile when for each sample point, there is at most one sample that is in front of Zfar. Transparent layer processing is not finished as long as there are two or more samples in front of Zfar for any sample point in the tile.
  • front-to-back sorted transparency modes the transparent samples are rendered in order, starting at the front most transparent sample and then the next farther transparent sample in each subsequent cycle is rendered.
  • An advantage of using a front-to-back sorted transparency mode is that if a maximum number of layers is defined, then the front most transparent layers are rendered which thereby provides a more accurate final displayed image.
  • the maximum number of layers to render is determined by accumulating the alpha values.
  • the alpha value represents the transparency of the sample location. As each sample is rendered the transparency at that sample location decreases, and the cumulative alpha value increases
  • an alpha value of one is defined as opaque.
  • the maximum cumulative alpha value may be defined to be 0.9, when the cumulative alpha value exceeds 0.9 then no further samples at that sample location are rendered.
  • the first counter is set to one which indicates the first primitive in this group should be rendered.
  • this first counter is incremented, to identify the second primitive as the primitive to be rendered.
  • the second counter maintains a count of the primitive being evaluated within a pass.
  • the third primitive has the sample that should be rendered.
  • the first counter is equal to three and the second counter is equal to one.
  • the first counter value is compared with the second counter value and because the counter values are not equal the sample from the first primitive is not rendered.
  • the second counter is then incremented, but the counters are still not equal so the sample from the second primitive is not rendered.
  • the first and second counter values are equal, therefore the sample from the third primitive is rendered.
  • inventive deferred shading graphics processor DSGP
  • CULL deferred shading graphics processor
  • a method of performing hidden surface removal in a computer graphics pipeline comprising the steps of: selecting a current primitive from a group of primitives, each primitive comprising a plurality of stamps; comparing stamps in the current primitive to stamps from previously evaluated primitives in the group of primitives; selecting a first stamp as a currently potentially visible stamp (CPVS) based on a relationship of depth states of samples in the first stamp with depth states of samples of previously evaluated stamps; comparing the CPVS to a second stamp; discarding the second stamp when no part of the second stamp would affect a final graphics display image based on the stamps that have been evaluated; discarding the CPVS and making the second stamp the CPVS, when the second stamp hides the CPVS; dispatching the CPVS and making the second stamp the CPVS when both the second stamp and the CPVS are at least partially visible in the final graphics display image; and dispatching the second stamp and the CPVS when the visibility of the second stamp and the CPVS depends on parameters evaluated later in the computer graphics pipeline.
  • the method of (1) wherein the step of comparing the CPVS to a second stamp furthing comprises the steps of: comparing depth states of samples in the CPVS to depth states of samples in the second stamp; and evaluating pipeline state values.
  • the depth state comprises one z value per sample, and wherein the z value includes a state bit which is defined to be accurate when the z value represents an actual z value of a currently visible surface and is defined to be conservative when the z value represents a maximum z value.
  • the method of (1 ) further comprising the step of dispatching the second stamp and the CPVS when the second stamp potentially alters the final graphics display image independent of the depth state.
  • the method of (1 ) further comprising the steps of: coloring the dispatched stamps; and performing an exact z buffer test on the dispatched stamps, after the coloring step.
  • the method of (1 ) further comprising the steps of: comparing alpha values of a plurality of samples to a reference alpha value; and performing the step of dispatching the second stamp and the CPVS, independent of alpha values when the alpha values of the plurality of samples are all greater than the reference value.
  • the method of (1 ) further comprising the steps of: determining whether any samples in the current primitive may affect final pixel color values in the final graphics display image; and turning blending off for the current primitive when no samples in the current primitive affect final pixel color values in the final graphics display image.
  • the step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining a maximum z value for a plurality of stamp locations of the current primitive; comparing the maximum z value for a plurality of stamp positions with a minimum z value of the current primitive and setting corresponding stamp selection bits; and identifying as a process row a row of stamps wherein the maximum z value for a stamp position in the row is greater than the minimum z value of the current primitive.
  • the method of (8) wherein the step of determining a maximum z value for a plurality of stamp locations of the current primitive further comprises determining a maximum z value for each stamp in a bounding box of the current primitive.
  • step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining the left most and right most stamps touched by the current primitive in each of the process rows and defining corresponding stamp primitive coverage bits; and combining the stamp primitive coverage bits with the stamp selection bits to generate a final potentially visible stamp set.
  • step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining a set of sample points in a stamp in the final potentially visible stamp set; computing a z value for a plurality of sample points in the set of sample points; and comparing the computed z values with stored z values and outputting sample control signals.
  • the step of computing a z value for a plurality of sample points in the set of sample points further comprises the steps of: creating a reference z value for a stamp; computing partial derivatives for a plurality of sample points in the set of sample points; sending down the pipeline the reference z value and the partial derivatives; and computing a z value for a sample based on the reference z value and partial derivatives.
  • the method of (10) further comprising the steps of: receiving a reference z value and partial derivatives; and re-computing a z value for a sample based on the reference z value and partial derivatives.
  • the method of (10) further comprising the step of dispatching the CPVS when the CPVS can affect stencil values.
  • the method of (13) further comprising the step of dispatching all currently potentially visible stamps when a stencil test changes.
  • the method of (10) further comprising the steps of: storing concurrently samples from a plurality of primitives; and comparing a computed z value for a sample at a first sample location with stored z values of samples at the first sample location from a plurality of primitives.
  • each stamp comprises at least one pixel and wherein the pixels in a stamp are processed in parallel.
  • the method of (20) further comprising the steps of: dividing a display image area into tiles; and rendering the display image in each tile independently.
  • the method of (10) wherein the sample points are located at positions between subraster grid lines.
  • locations of the sample points within each pixel are programmable.
  • the method of (23) further comprising the steps of: programming a first set of sample locations in a plurality of pixels; evaluating stamp visibility using the first set of sample locations; programming a second set of sample locations in a plurality of pixels; and evaluating stamp visibility using the second set of sample locations.
  • the method of (10) further comprising the step of eliminating individual stamps that are determined not to affect the final graphics display image.
  • the method of (10) further comprising the step of turning off blending when alpha values at vertices of the current primitive have values such that frame buffer color values cannot affect a final color of samples in the current primitive.
  • the method of (1 ) wherein the depth state comprises a far z value and a near z value.
  • a hidden surface removal system for a deferred shader computer graphics pipeline comprising: a magnitude comparison content addressable memory Cull unit for identifying a first group of potentially visible samples associated with a current primitive; a Stamp Selection unit, coupled to the magnitude comparison content addressable memory cull unit, for identifying, based on the first group and a perimeter of the primitive, a second group of potentially visible samples associated with the primitive; a Z Cull unit, coupled to the stamp selection unit and the magnitude comparison content addressable memory cull unit, for identifying visible stamp portions by evaluating a pipeline state, and comparing depth states of the second group with stored depth state values; and a Stamp Portion Memory unit, coupled to the Z Cull unit, for storing visible stamp portions based on control signals received from the Z Cull unit, wherein the Stamp Portion Memory unit dispatches stamps having a visibility dependent on parameters evaluated later in the computer graphics pipeline.
  • the Z Cull unit evaluates depth state and pipeline state values, and compares a currently potentially visible stamp (CPVS) to a first stamp; and wherein the Stamp Portion Memory, based on control signals from the Z Cull unit: discards the first stamp when no part of the first stamp would affect a final graphics display image based on the stamps that have been evaluated; discards the CPVS and makes the first stamp the CPVS, when the first stamp hides CPVS; dispatches the CPVS and makes the first stamp the CPVS when both the first stamp and the CPVS are at least partially visible in the final graphics display image; and dispatches the first stamp and the CPVS when the visibility of the first stamp and the CPVS depends on parameters evaluated later in the computer graphics pipeline.
  • CPVS currently potentially visible stamp
  • each of the Z Cull Sample State Machines receive, process and output control signals for samples in parallel.
  • a method of rendering a computer graphics image comprising the steps of: receiving a plurality of primitives to be rendered; selecting a sample location; rendering a front most opaque sample at the selected sample location, and defining the z value of the front most opaque sample as Zfar; comparing z values of a first plurality of samples at the selected sample location; defining to be Znear a first sample, at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the first plurality of samples; rendering the first sample; setting Zfar to the value of Znear; comparing z values of a second plurality of samples at the selected sample location; defining as Znear the z value of a second sample at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the second plurality of samples; and rendering the second sample.
  • the method of 35 further comprising the steps of: when a third plurality of samples at the selected sample location have a common z value which is less than Zfar, and the common z value is the z value nearest to Zfar of the first plurality of samples: rendering a third sample, wherein the third sample is the first sample received of the third plurality of samples; incrementing a first counter value to define a sample render number, wherein the sample render number identifies the sample to be rendered; selecting a fourth sample from the third plurality of samples; incrementing a second counter wherein the second counter defines an evaluation sample number; comparing the sample render number and the evaluation sample number; and rendering a sample when the corresponding evaluation sample number equals the sample render number.

Abstract

Structure (FIG.12) and method for performing conservative hidden surface removal in a graphics processor. Culling (9000) is divided into two steps, a magnitude comparison content addressable memory cull operation (MCCAM Cull) (9002), and a subpixel cull operation. The MCCAM Cull discards primitives that are hidden completely by previously processed geometry. The Subpixel Cull takes the remaining primitives (which are partly or entirely visible), and determines the visible fragments. Performing hidden surface removal includes: selecting a current primitive comprising a plurality of stamps; comparing stamps to stamps from previously evaluated primitives; selecting a first stamp as a currently potentially visible stamp (CPVS) based on the depth of current and previous stamps; comparing the CPVS to a second stamp; discarding the second stamp when it would not affect a final graphics display image; making the second stamp the CPVS, when it hides the CPVS and when both the second stamp and the CPVS are at least partially visible; and dispatching both when their visibility depends on parameters evaluated later in the graphics pipeline.

Description

METHOD AND APPARATUS FOR PERFORMING CONSERVATIVE HIDDEN SURFACE REMOVAL IN A GRAPHICS PROCESSOR WITH DEFERRED SHADING
Inventors
Jerome F. Duluk Jr., Steven L. Dodgen, Emerson S. Fang, Heng-Wei Hsu, Jason R. Redgrave, Sushma Trivedi
RELATED APPLICATIONS This application claims the benefit under 35 USC Section 119(e) of U.S. Provisional Patent
Application Serial No. 60/097,336 filed 20 August 1998 and entitled GRAPHICS PROCESSOR WITH DEFERRED SHADING; claims the benefit under 35 USC Section 120 of U.S. Patent Application Serial No.
09/213,990 filed 17 December 1998 entitled HOW TO DO TANGENT SPACE LIGHTING IN A DEFERRED SHADING ARCHITECTURE; each of which is hereby incorporated by reference.
This application is also related to the following U.S. Patent Applications, each of which are incorporated herein by reference:
Serial No. 09/213,990, filed 17 December 1998, entitled HOW TO DO TANGENT SPACE LIGHTING IN A DEFERRED SHADING ARCHITECTURE (Atty. Doc. No. A-66397);
Serial No. , filed 20 August 1999, entitled APPARATUS AND METHOD FOR PERFORMING SETUP OPERATIONS IN A 3-D GRAPHICS PIPELINE USING UNIFIED PRIMITIVE
DESCRIPTORS (Atty. Doc. No. A-66382); Serial No. , filed 20 August 1999, entitled POST-FILE SORTING SETUP (Atty. Doc.
No. A-66383); Serial No. filed 20 August 1999, entitled TILE RELATIVE Y-VALUES AND SCREEN
RELATIVE X-VALUES (Atty. Doc. No. A-66384); Serial No. , filed 20 August 1999, entitled SYSTEM, APARATUS AND METHOD FOR
SPATIALLY SORTING IMAGE DATA IN A THREE-DIMENSIONAL GRAPHICS PIPELINE (Atty. Doc. No. A-66380);
Serial No. , filed 20 August 1999, entitled SYSTEM, APPARATUS AND METHOD FOR GENERATING GUARANTEED CONSERVATIVE MEMORY ESTIMATE FOR SORTING OBJECT
GEOMETRY IN A THREE-DIMENSIONAL GRAPHICS PIPELINE (Atty. Doc. No. A-66381);
Ser1ai No filed entitled METHOD AND APPARATUS FOR
PERFORMING CONSERVATIVE HIDDEN SURFACE REMOVAL IN A GRAPHICS PROCESSOR WITH DEFERRED SHADING (Attorney Doc. No. A-66386);
Seriai No __, filed entitled DEFERRED SHADING GRAPHICS
PIPELINE PROCESSOR HAVING ADVANCED FEATURES (Atty. Doc. No. A-66364) Serial No. , filed 20 August 1999, entitled SYSTEM, APPARATUS AND METHOD FOR BALANCING RENDERING RESOURCES IN A THREE-DIMENSIONAL GRAPHICS PIPELINE (Atty. Doc.
No. A-66379); Serial No. , filed 20 August 1999, entitled GRAPHICS PROCESSOR WITH PIPELINE
STATE STORAGE AND RETRIEVAL (Atty. Doc. No. A-66378); Serial No , filed 20 August 1999, entitled METHOD AND APPARATUS FOR
GENERATING TEXTURE (Atty. Doc. No. A-66398); Serial No. , filed 20 August 1999, entitled APPARATUS AND METHOD FOR
GEOMETRY OPERATIONS IN A 3D GRAPHICS PIPELINE (Atty. Doc. No. A-66373); Serial No. , filed 20 August 1999, entitled APPARATUS AND METHOD FOR
FRAGMENT OPERATIONS IN A 3D GRAPHICS PIPELINE (Atty. Doc. No. A-66399); and Serial No. filed 20 August 1999, entitled DEFERRED SHADING GRAPHICS PIPELINE
PROCESSOR (Atty. Doc. No. A-66360).
1 FIELD OF THE INVENTION This invention relates to three-dimensional computer graphics computer systems, and more particularly to a system for performing conservative hidden surface removal in a graphics processor with deferred shading.
2 BACKGROUND OF THE INVENTION
Computer graphics is the art and science of generating pictures with a computer. This picture or image generation process is commonly called rendering. Generally, in three-dimensional (3D) computer graphics, geometry that represents surfaces (or volumes) of objects in a scene is translated into pixels stored in a frame buffer, and then displayed on a display device. Real-time display devices, such as CRTs used as computer monitors, refresh the display by continuously displaying the image over and over. This refresh usually occurs row-by-row, where each row is called a raster line or scan line. In this document, raster lines are numbered from bottom to top, but are displayed in order from top to bottom. In a 3D animation, a sequence of images is displayed, giving the illusion of motion in three- dimensional space. Interactive 3D computer graphics allows a user to change his viewpoint or change the geometry in real-time, thereby requiring the rendering system to create new images on-the-fly in real-time.
In 3D computer graphics, each renderable object generally has its own local object coordinate system, and therefore needs to be translated (or transformed) from object coordinates to pixel display coordinates. Conceptually, this is a 4-step process: 1) translation (including scaling for size enlargement or shrink) from object coordinates to world coordinates, which is the coordinate system for the entire scene;
2) translation from world coordinates to eye coordinates, based on the viewing point of the scene; 3) translation from eye coordinates to perspective translated eye coordinates, where perspective scaling
(farther objects appear smaller) has been performed; and 4) translation from perspective translated eye coordinates to pixel coordinates, also called screen coordinates. Screen coordinates are points in three- dimensional space, and can be in either screen-precision (i.e., pixels) or object-precision (high precision numbers, usually floating-point), as described later. These translation steps can be compressed into one or two steps by pre-computing appropriate translation matrices before any translation occurs. Once the geometry is in screen coordinates, it is broken into a set of pixel color values (that is "rasterized") that are stored into the frame buffer. Many techniques are used for generating pixel color values, including Gouraud shading, Phong shading, and texture mapping.
A summary of the prior art rendering process can be found in: "Fundamentals of Three-dimensional Computer Graphics", by Watt, Chapter 5: The Rendering Process, pages 97 to 113, published by Addison-
Wesley Publishing Company, Reading, Massachusetts, 1989, reprinted 1991 , ISBN 0-201-15442-0 (hereinafter referred to as the Watt Reference).
Figure 1 shows a three-dimensional object, a tetrahedron, with its own coordinate axes (xobJ,yobJ,zobJ). The three-dimensional object is translated, scaled, and placed in the viewing point's coordinate system based on (xeye.yeye.z eye)- Tne object is projected onto the viewing plane, thereby correcting for perspective. At this point, the object appears to have become two-dimensional; however, the object's z coordinates are preserved so they can be used later by hidden surface removal techniques. The object is finally translated to screen coordinates, based on (xscreen.yscreen.z screen). where z.creen is going perpendicularly into the page.
Points on the object now have their x and y coordinates described by pixel location (and fractions thereof) within the display screen and their z coordinates in a scaled version of distance from the viewing point.
Because many different portions of geometry can affect the same pixel, the geometry representing the surfaces closest to the scene viewing point must be determined. Thus, for each pixel, the visible surfaces within the volume subtended by the pixel's area determine the pixel color value, while hidden surfaces are prevented from affecting the pixel. Non-opaque surfaces closer to the viewing point than the closest opaque surface (or surfaces, if an edge of geometry crosses the pixel area) affect the pixel color value, while all other non-opaque surfaces are discarded. In this document, the term "occluded" is used to describe geometry which is hidden by other non-opaque geometry. Many techniques have been developed to perform visible surface determination, and a survey of these techniques are incorporated herein by reference to: "Computer Graphics: Principles and Practice", by Foley, van Dam, Feiner, and Hughes, Chapter 15: Visible-Surface Determination, pages 649 to 720, 2nd edition published by Addison-Wesley Publishing Company, Reading, Massachusetts, 1990, reprinted with corrections 1991 , ISBNO-201 -12110-7 (hereinafter referred to as the Foley Reference). In the Foley
Reference, on page 650, the terms "image-precision" and "object-precision" are defined: "Image-precision algorithms are typically performed at the resolution of the display device, and determine the visibility at each pixel. Object-precision algorithms are performed at the precision with which each object is defined, and determine the visibility of each object."
As a rendering process proceeds, most prior art renderers must compute the color value of a given screen pixel multiple times because multiple surfaces intersect the volume subtended by the pixel. The average number of times a pixel needs to be rendered, for a particular scene, is called the depth complexity of the scene. Simple scenes have a depth complexity near unity, while complex scenes can have a depth complexity of ten or twenty. As scene models become more and more complicated, renderers will be required to process scenes of ever increasing depth complexity. Thus, for most renders, the depth complexity of a scene is a measure of the wasted processing. For example, for a scene with a depth complexity of ten, 90% of the computation is wasted on hidden pixels. This wasted computation is typical of hardware renderers that use the simple Z buffer technique (discussed later herein), generally chosen because it is easily built in hardware. Methods more complicated than the Z Buffer technique have heretofore generally been too complex to build in a cost-effective manner. An important feature of the method and apparatus invention presented here is the avoidance of this wasted computation by eliminating hidden portions of geometry before they are rasterized, while still being simple enough to build in cost- effective hardware.
When a point on a surface (frequently a polygon vertex) is translated to screen coordinates, the point has three coordinates: 1) the x coordinate in pixel units (generally including a fraction); 2) the y coordinate in pixel units (generally including a fraction); and 3) the z coordinate of the point in either eye coordinates, distance from the virtual screen, or some other coordinate system which preserves the relative distance of surfaces from the viewing point. In this document, positive z coordinate values are used for the "look direction" from the viewing point, and smaller values indicate a position closer to the viewing point.
When a surface is approximated by a set of planar polygons, the vertices of each polygon are translated to screen coordinates. For points in or on the polygon (other than the vertices), the screen coordinates are interpolated from the coordinates of vertices, typically by the processes of edge walking and span interpolation. Thus, a z coordinate value is generally included in each pixel value (along with the color value) as geometry is rendered.
2.1 Generic 3D Graphics Pipeline Many hardware renderers have been developed, and an example is incorporated herein by reference: "Leo: A System for Cost Effective 3D Shaded Graphics", by Deering and Nelson, pages 101 to 108 of SIGGRAPH93 Proceedings, 1 -6 August 1993, Computer Graphics Proceedings, Annual Conference
Series, published by ACM SIGGRAPH, New York, 1993, Softcover ISBN 0-201-58889-7 and CD-ROM ISBN 0-201-56997-3 (hereinafter referred to as the Deering Reference). The Deering Reference includes a diagram of a generic 3D graphics pipeline (i.e., a renderer, or a rendering system) that it describes as "truly generic, as at the top level nearly every commercial 3D graphics accelerator fits this abstraction", and this pipeline diagram is reproduced here as Figure 2. Such pipeline diagrams convey the process of rendering, but do not describe any particular hardware. This document presents a new graphics pipeline that shares some of the steps of the generic 3D graphics pipeline. Each of the steps in the generic 3D graphics pipeline will be briefly explained here. Processing of polygons is assumed throughout this document, but other methods for describing 3D geometry could be substituted. For simplicity of explanation, triangles are used as the type of polygon in the described methods.
As seen in Figure 2, the first step within the floating-point intensive functions of the generic 3D graphics pipeline after the data input (step 212) is the transformation step (step 214), which was described above. The second step, the clip test, checks the polygon to see if it is at least partially contained in the view volume (sometimes shaped as a frustum) (step 216). If the polygon is not in the view volume, it is discarded; otherwise processing continues. The third step is face determination, where polygons facing away from the viewing point are discarded (step 218). Generally, face determination is applied only to objects that are closed volumes. The fourth step, lighting computation, generally includes the set up for Gouraud shading and/or texture mapping with multiple light sources of various types, but could also be set up for Phong shading or one of many other choices (step 222). The fifth step, clipping, deletes any portion of the polygon that is outside of the view volume because that portion would not project within the rectangular area of the viewing plane (step 224). Generally, polygon clipping is done by splitting the polygon into two smaller polygons that both project within the area of the viewing plane. Polygon clipping is computationally expensive. The sixth step, perspective divide, does perspective correction for the projection of objects onto the viewing plane (step 226). At this point, the points representing vertices of polygons are converted to pixel space coordinates by step seven, the screen space conversion step (step 228). The eighth step (step 230), set up for incremental render, computes the various begin, end, and increment values needed for edge walking and span interpolation (e.g.: x, y, and z coordinates; RGB color; texture map space u and v coordinates; and the like).
Within the drawing intensive functions, edge walking (step 232) incrementally generates horizontal spans for each raster line of the display device by incrementing values from the previously generated span
(in the same polygon), thereby "walking" vertically along opposite edges of the polygon. Similarly, span interpolation (step 234) "walks" horizontally along a span to generate pixel values, including a z coordinate value indicating the pixel's distance from the viewing point. Finally, the z buffered blending also referred to as Testing and Blending (step 236) generates a final pixel color value. The pixel values also include color values, which can be generated by simple Gouraud shading (i.e., interpolation of vertex color values) or by more computationally expensive techniques such as texture mapping (possibly using multiple texture maps blended together), Phong shading (i.e., per-fragment lighting), and/or bump mapping (perturbing the interpolated surface normal). After drawing intensive functions are completed, a double-buffered MUX output look-up table operation is performed (step 238). In this figure the blocks with rounded corners typically represent functions or process operations, while sharp cornered rectangles typically represent stored data or memory. By comparing the generated z coordinate value to the corresponding value stored in the Z Buffer, the z buffered blend either keeps the new pixel values (if it is closer to the viewing point than previously stored value for that pixel location) by writing it into the frame buffer, or discards the new pixel values (if it is farther). At this step, antialiasing methods can blend the new pixel color with the old pixel color. The z buffered blend generally includes most of the per-fragment operations, described below.
The generic 3D graphics pipeline includes a double buffered frame buffer, so a double buffered MUX is also included. An output lookup table is included for translating color map values. Finally, digital to analog conversion makes an analog signal for input to the display device. A major drawback to the generic 3D graphics pipeline is its drawing intensive functions are not deterministic at the pixel level given a fixed number of polygons. That is, given a fixed number of polygons, more pixel-level computation is required as the average polygon size increases. However, the floating-point intensive functions are proportional to the number of polygons, and independent of the average polygon size. Therefore, it is difficult to balance the amount of computational power between the floating-point intensive functions and the drawing intensive functions because this balance depends on the average polygon size. Prior art Z Buffers are based on conventional Random Access Memory (RAM or DRAM), Video RAM (VRAM), or special purpose DRAMs. One example of a special purpose DRAM is presented in
"FBRAM: A new Form of Memory Optimized for 3D Graphics", by Deering, Schlapp, and Lavelle, pages 167 to 174 of SIGGRAPH94 Proceedings, 24-29 July 1994, Computer Graphics Proceedings, Annual
Conference Series, published by ACM SIGGRAPH, New York, 1994, Softcover ISBN 0201607956.
2.2 Pipeline State OpenGL® is a software interface to graphics hardware which consists of several hundred functions and procedures that allow a programmer to specify objects and operations to produce graphical images. The objects and operations include appropriate characteristics to produce color images of three-dimensional objects. Most of OpenGL® (Version 1.2) assumes or requires a that the graphics hardware include a frame buffer even though the object may be a point, line, polygon, or bitmap, and the operation may be an operation on that object. The general features of OpenGL® (just one example of a graphical interface) are described in the reference "The OpenGL® Graphics System: A Specification
(Version 1.2) edited by Mark Segal and Kurt Akeley, Version 1.2, March 1998; and hereby incorporated by reference. Although reference is made to OpenGL®, the invention is not limited to structures, procedures, or methods which are compatible or consistent with OpenGL®, or with any other standard or non-standard graphical interface. Desirably, the inventive structure and method may be implemented in a manner that is consistent with the OpenGL®, or other standard graphical interface, so that a data set prepared for one of the standard interfaces may be processed by the inventive structure and method without modification.
However, the inventive structure and method provides some features not provided by OpenGL®, and even when such generic input/output is provided, the implementation is provided in a different manner.
The phrase "pipeline state" does not have a single definition in the prior-art. The OpenGL® specification, for example, sets forth the type and amount of the graphics rendering machine or pipeline state in terms of items of state and the number of bits and bytes required to store that state information. In the OpenGL® definition, pipeline state tends to include object vertex pertinent information including for example, the vertices themselves the vertex normals, and color as well as "non-vertex" information. When information is sent into a graphics renderer, at least some object geometry information is provided to describe the scene. Typically, the object or objects are specified in terms of vertex information, where an object is modeled, defined, or otherwise specified by points, lines, or polygons (object primitives) made up of one or more vertices. In simple terms, a vertex is a location in space and may be specified for example by a three-space (x,y,z) coordinate relative to some reference origin. Associated with each vertex is other information, such as a surface normal, color, texture, transparency, and the like information pertaining to the characteristics of the vertex. This information is essentially "per-vertex" information.
Unfortunately, forcing a one-to-one relationship between incoming information and vertices as a requirement for per-vertex information is unnecessarily restrictive. For example, a color value may be specified in the data stream for a particular vertex and then not respecified in the data stream until the color changes for a subsequent vertex. The color value may still be characterized as per-vertex data even though a color value is not explicitly included in the incoming data stream for each vertex. Texture mapping presents an interesting example of information or data which could be considered as either per-vertex information or pipeline state information. For each object, one or more texture maps may be specified, each texture map being identified in some manner, such as with a texture coordinate or coordinates. One may consider the texture map to which one is pointing with the texture coordinate as part of the pipeline state while others might argue that it is per-vertex information. Other information, not related on a one-to-one basis to the geometry object primitives, used by the renderer such as lighting location and intensity, material settings, reflective properties, and other overall rules on which the renderer is operating may more accurately be referred to as pipeline state. One may consider that everything that does not or may not change on a per-vertex basis is pipeline state, but for the reasons described, this is not an entirely unambiguous definition. For example, one may define a particular depth test (See later description) to be applied to certain objects to be rendered, for example the depth test may require that the z value be strictly "greater-than" for some objects and "greater-than-or-equal-to" for other objects. These particular depth tests which change from time to time, may be considered to be pipeline state at that time.
Parameters considered to be renderer (pipeline) state in OpenGL® are identified in Section 6.2 of the aforementioned OpenGL® Specification (Version 1.2, at pages 193-217).
Essentially then, there are two types of data or information used by the renderer: (1 ) primitive data which may be thought of as per-vertex data, and (ii) pipeline state data (or simply pipeline state) which is everything else. This distinction should be thought of as a guideline rather than as a specific rule, as there are ways of implementing a graphics renderer treating certain information items as either pipeline state or non-pipeline state.
2.3 Per-Fragment Operations In the generic 3D graphics pipeline, the "z buffered blend" step actually incorporates many smaller
"per-fragment" operational steps. Application Program Interfaces (APIs), such as OpenGL® (Open Graphics Library) and D3D, define a set of per-fragment operations (See Chapter 4 of Version 1.2 OpenGL® Specification). Some exemplary OpenGL® per-fragment operations are briefly reviewed so that any generic similarities and differences between the inventive structure and method and conventional structures and procedures can be more readily appreciated.
Under OpenGL®, a frame buffer stores a set of pixels as a two-dimensional array. Each picture- element or pixel stored in the frame buffer is simply a set of some number of bits. The number of bits per pixel may vary depending on the particular GL implementation or context. Corresponding bits from each pixel in the frame buffer are grouped together into a bitplane; each bitplane containing a single bit from each pixel. The bitplanes are grouped into several logical buffers referred to as the color, depth, stencil, and accumulation buffers. The color buffer in turn includes what is referred to under OpenGL® as the front left buffer, the front right buffer, the back left buffer, the back right buffer, and some additional auxiliary buffers. The values stored in the front buffers are the values typically displayed on a display monitor while the contents of the back buffers and auxiliary buffers are invisible and not displayed. Stereoscopic contexts display both the front left and the front right buffers, while monoscopic contexts display only the front left buffer. In general, the color buffers must have the same number of bitplanes, but particular implementations of context may not provide right buffers, back buffers, or auxiliary buffers at all, and an implementation or context may additionally provide or not provide stencil, depth, or accumulation buffers.
Under OpenGL®, the color buffers consist of either unsigned integer color indices or R, G, B, and, optionally, a number "A" of unsigned integer values; and the number of bitplanes in each of the color buffers, the depth buffer (if provided), the stencil buffer (if provided), and the accumulation buffer (if provided), is fixed and window dependent. If an accumulation buffer is provided, it should have at least as many bit planes per R, G, and B color component as do the color buffers. A fragment produced by rasterization with window coordinates of (xw, yw) modifies the pixel in the frame buffer at that location based on a number of tests, parameters, and conditions. Noteworthy among the several tests that are typically performed sequentially beginning with a fragment and its associated data and finishing with the final output stream to the frame buffer are in the order performed (and with some variation among APIs): 1 ) pixel ownership test; 2) scissor test; 3) alpha test; 4) Color Test; 5) stencil test;
6) depth test; 7) blending; 8) dithering; and 9) logicop. Note that the OpenGL® does not provide for an explicit "color test" between the alpha test and stencil test. Per-Fragment operations under OpenGL® are applied after all the color computations. Each of these tests or operations is briefly described below.
2.3.1 Ownership Test Under OpenGL®, the pixel ownership test determines if the pixel at location (xw, yw) in the frame buffer is currently owned by the GL context. If it is not, the window system decides the fate of the incoming fragment. Possible results are that the fragment is discarded or that some subset of the subsequent per- fragment operations are applied to the fragment. This pixel ownership test allows the window system to properly control the GL's behavior.
Assume that in a computer having a display screen, one or several processes are running and that each process has a window on the display screen. For each process, the associated window defines the pixels the process wants to write or render to. When there are two or more windows, the window associated with one process may be in front of the window associated with another process, behind that window, or both windows may be entirely visible. Since there is only a single frame buffer for the entire display screen or desktop, the pixel ownership test involves determining which process and associated window owns each of the pixels. If a particular process does not "own" a pixel, it fails the pixel ownership test relative to the frame buffer and that pixel is thrown away. Note that under the typical paradigm, the pixel ownership test is run by each process, and that for a give pixel location in the frame buffer, that pixel may pass the pixel ownership test for one of the processes, and fail the pixel ownership test for the other process.
Furthermore, in general, a particular pixel can pass the ownership test for only one process because only one process can own a particular frame buffer pixel at the same time.
In some rendering schemes the pixel ownership test may not be particularly relevant. For example, if the scene is being rendered to an off-screen buffer, and subsequently Block Transferred or "blitted" to the desktop, pixel ownership is not really even relevant. Each process automatically or necessarily passes the pixel ownership test (if it is executed) because each process effectively owns its own off-screen buffer and nothing is in front of that buffer.
If for a particular process, the pixel is not owned by that process, then there is no need to write a pixel value to that location, and all subsequent processing for that pixel may be ignored. In a typical workstation, all the data associated with a particular pixel on the screen is read during rasterization. All information for any polygon that feeds that pixel is read, including information as to the identity of the process that owns that frame buffer pixel, as well as the Z buffer, the color value, the old color value, the alpha value, stencil bits, and so forth. If a process owns the pixel, then the other downstream process are executed (for example, scissor test, alpha test, and the like) On the other hand, if the process does not own the pixel and fails the ownership test for that pixel, the process need not consider that pixel further and that pixel is skipped for subsequent tests.
2.3.2 Scissor Test Under OpenGL®, the scissor test determines if (x^ y lies within a scissor rectangle defined by four coordinate values corresponding to a left bottom (left, bottom) coordinate, a width of the rectangle, and a height of the rectangle. The values are set with the procedure "void Scissor( int left, int bottom, sizei width, sizei height)" under OpenGL®. If left ≤ xw < left+width and bottom ≤ yw < bottom+height, then the scissor test passes; otherwise the scissor test fails and the particular fragment being tested is discarded. Various initial states are provided and error conditions monitored and reported. In simple terms, a rectangle defines a window which may be an on-screen or off-screen window.
The window is defined by an x-left, x-right, y-top, and y-bottom coordinate (even though it may be expressed in terms of a point and height and width dimensions from that point). This scissor window is useful in that only pixels from a polygon fragment that fall in that screen aligned scissor window will change. In the event that a polygon straddles the scissor window, only those pixels that are inside the scissor window may change. When a polygon in an OpenGL® machine comes down the pipeline, the pipeline calculates everything it needs to in order to determine the z value and color of that pixel. Once z value and color are determined, that information is used to determine what information should be placed in the frame buffer
(thereby determining what is displayed on the display screen). Just as with the pixel ownership test, the scissor test provides means for discarding pixels and/or fragments before they actually get to the frame buffer to cause the output to change.
2.3.3 Alpha Test Color is defined by four values, red (R), green (G), blue (B), and alpha (A). The RGB values define the contribution from each of the primary colors, and alpha is related to the transparency. Typically, color is a 32-bit value, 8-bits for each component, though such representation is not limited to 32-bits. Alpha test compares the alpha value of a given pixel to an alpha reference value. The type of comparison may also be specified, so that for example the comparison may be a greater-than operation, a less-than operation, and so forth. If the comparison is a greater-than operation, then the pixel's alpha value has to be greater than the reference to pass the alpha test. Any pixel not passing the alpha test is thrown away or discarded. The OpenGL® Specification describes the manner in which alpha test is implemented in OpenGL®.
Alpha test is a per-fragment operation and after all of the fragment coloring calculations and lighting and shading operations are completed. Each of these per-fragment operations may be thought of as part of the conventional z buffer blending operations.
2.3.4 Color Test Color test is similar to the alpha test described hereinbefore, except that rather than performing the magnitude or logical comparisons between the pixel alpha (A) value and a reference value, the color test performs a magnitude or logical comparison between one or a combination of the R, G, or B color components and reference value(s). The comparison test may be for example, greater-than, less-than, equal-to, greater-than-or-equal-to, "greater-than- and less- thai] c " where c ^nd c are sore predetermined reference values, and so forth. One might for example, specify a reference minimum R value, and a reference maximum R value, such that the color test would be passed only if the pixel R value is between that minimum and maximum. Color test might, for example, be useful to provide blue-screen functionality. The comparison test may also be performed on a single color component or on a combination of color components. Furthermore, although for the alpha test one typically has one value for each component, for the color test there are effectively two values per component, a maximum value and a minimum value.
2.3.5 Stencil Test Under OpenGL®, stencil test conditionally discards a fragment based on the outcome of a comparison between a value stored in a stencil buffer at location (xw, y and a reference value. Several stencil comparison functions are permitted such that whether the stencil test passes can depend upon whether the reference value is less than, less than or equal to, equal to, greater than or equal to, greater than, or not equal to the masked stored value in the stencil buffer. The Under OpenGL®, if the stencil test fails, the incoming fragment is discarded. The reference value and the comparison value can have multiple bits, typically 8 bits so that 256 different values may be represented. When an object is rendered into the frame buffer, a tag having the stencil bits is also written into the frame buffer. These stencil bits are part of the pipeline state. The type of stencil test to perform can be specified at the time the geometry is rendered. The stencil bits are used to implement various filtering, masking or stenciling operations. For example, if a particular fragment ends up affecting a particular pixel in the frame buffer, then the stencil bits can be written to the frame buffer along with the pixel information.
2.3.6 Depth Buffer Test
Under OpenGL®, the depth buffer test discards the incoming fragment if a depth comparison fails. The comparison is enabled or disabled with the generic Enable and Disable commands using the OpenGL® symbolic constant DEPTH_TEST. When depth test is disabled, the depth comparison and subsequent possible updates to the depth buffer value are bypassed and a fragment is passed to the next operation. The stencil bits are also involved and are modified even if the test is bypassed. The stencil value is modified if the depth buffer test passed. If depth test is enabled, the depth comparison takes place and the depth buffer and stencil value may subsequently be modified. The manner in which the depth test is implemented in OpenGL® is described in greater detail in the OpenGL® specification at page 145.
Depth comparisons are implemented in which possible outcomes are as follows: the depth buffer test either never passes or always passes, if the incoming fragment's z„ value is less than, less than or equal to, equal to, greater than, greater than or equal to, or not equal to the depth value stored at the location given by the incoming fragment's (xw, yw) coordinates. If the depth buffer test fails, the incoming fragment is discarded. The stencil value at the fragment's (x^ yw) coordinate is updated according to the function currently in effect for depth buffer test failure. Otherwise, the fragment continues to the next operation and the value of the depth buffer at the fragment's (x*,, yw) location is set to the fragment's z, value. In this case the stencil value is updated according to the function currently in effect for depth buffer test success. The necessary OpenGL® state is an eight-valued integer and a single bit indicating whether depth buffering is enabled or disabled.
2.3.7 Alpha Blending
Under OpenGL®, alpha blending (also referred toas blending) combines the incoming fragment's R, G, B, and A values with the R, G, B, and A values stored in the frame buffer at the incoming fragment's
(XwNw) location. This blending is typically dependent on the incoming fragment's alpha value (A) and that of the corresponding frame buffer stored pixel. In the following discussion, Cs refers to the source color for an incoming fragment, Cd refers to the destination color at the corresponding frame buffer location, and Cc refers to a constant color in-the GL state. Individual RGBA components of these colors are denoted by subscripts of s, d, and c respectively.
Blending is basically an operation that takes color in the frame buffer and the color in the fragment, and blends them together. The manner in which blending is achieved, that is the particular blending function, may be selected from various alternatives for both the source and destination. Blending is described in the OpenGL® specification at page 146-149 and is hereby incorporated by reference. Various blend equations are available under OpenGL®. For example, an additive type blend is available wherein a blend result (C) is obtained by adding the product of a source color (Cs) by a source weighting factor quadruplet (S) to the product of a destination color (Cd) and a destination weighting factor (D) quadruplet, that is C=CSS + CdD. Alternatively, the blend equation may be a subtraction (C=CSS - CdD), a reverse subtraction (C=CdD- CSS), a minimum function (C=min(Cs, Cd)), a maximum function (C=max(Cs, Cd)),. Under OpenGL®, the blending equation is evaluated separately for each color component and its corresponding weighting coefficient. Each of the four R, G, B, A components has its own weighting factor. The blending test (or blending equation) is part of pipeline state and can potentially change for every polygon, but more typically would change only for the object made up or several polygons. In generally, blending is only performed once other tests such as the pixel ownership test and stencil test have been passed so that it is clear that the pixel or fragment under consideration would or could have an effect in the output. 2.3.8 Dithering Under OpenGL®, dithering selects between two color values or indices. In RGBA mode, consider the value of any of the color components as a fixed-point value with m bits to the left of the binary point, where m is the number of bits allocated to that component in the frame buffer; call each such value c. For each c, dithering selects a value d such that d e {max{0, [c]-1 , [c]}. This selection may depend on the xw and yw coordinates of the pixel. In color index mode, the same rule applies with c being a single color index.
The value of c must not be larger than the maximum value representable in the frame buffer for either the component or the index.
Although many dithering algorithms are possible, a dithered value produced by any algorithm must generally depend only the incoming value and the fragment's x and y window coordinates. When dithering is disabled, each color component is truncated to a fixed-point value with as many bits as there are in the corresponding frame buffer component, and the color index is rounded to the nearest integer representable in the color index portion of the frame buffer. The OpenGL® Specification of dithering is described more fully in the OpenGL specification, particularly at pages 149-150, which are incorporated by reference.
2.3.9 Logicop Under OpenGL®, there is a final logical operation applied between the incoming fragment's color or index values and the color or index values stored in the frame buffer at the corresponding location. The result of the logical operation replaces the values in the frame buffer at the fragment's (x, y) coordinates.
Various logical operations may be implemented between source (s) and destination (d), including for example: clear, set, and, noop, xor, or, nor, nand, invert, copy, inverted and, equivalence, reverse or, reverse and, inverted copy, and inverted or. The logicop arguments and corresponding operations, as well as additional details of the OpenGL® logicop implementation, are set forth in the OpenGL® specification at pates 150-151. Logical operations are performed independently for each color index buffer that is selected for writing, or for each red, green, blue, and alpha value of each color buffer that is selected for writing. The required state is an integer indicating the logical operation, and two bits indicating whether the logical operation is enabled or disabled.
2.4 Antialiasing
In this document, pixels are referred to as the smallest individually controllable element of the display device. But, because images are quantized into discrete pixels, spatial aliasing occurs. A typical aliasing artifact is a "staircase" effect caused when a straight line or edge cuts diagonally across rows of pixels.
Some rendering systems reduce aliasing effects by dividing pixels into subpixels, where each sub- pixel can be colored independently. When the image is to be displayed, the colors for all sub-pixels within each pixel are blended together to form an average color for the pixel. A renderer that uses up to 16 sub- pixels per pixel is described in "RealityEngine Graphics", by Akeley, pages 109 to 116 of SIGGRAPH93
Proceedings, 1-6 August 1993, Computer Graphics Proceedings, Annual Conference Series, published by ACM SIGGRAPH, New York, 1993, Softcover ISBN 0-201-58889-7 and CD-ROM ISBN 0-201-56997-3 (hereinafter referred to as the Akeley Reference).
Another prior art antialiasing method is the A-Buffer used to perform blending (this technique is also included in the Akeley Reference), and is described in "The A-buffer, an Antialiased Hidden Surface Method" by L. Carpenter, SIGGRAPH 1984 Conference Proceedings, pp.103-108 (hereinafter referred to as the Carpenter Reference). The A-buffer is an antialiasing technique that reduces aliasing by keeping track of the percent coverage of a pixel by a rendered polygon. The main drawback to this technique is the need to sort polygons front-to-back (or back-to-front) at each pixel in order to get acceptable antialiased polygons.
2.5 Content Addressable Memories Most Content Addressable Memories (CAM) perform a bit-for-bit equality test between an input vector and each of the data words stored in the CAM. This type of CAM frequently provides masking of bit positions in order to eliminate the corresponding bit in all words from affecting the equality test. It is inefficient to perform magnitude comparisons in a equality-testing CAM because a large number of clock cycles is required to do the task. CAMs are presently used in translation look-aside buffers within a virtual memory systems in some computers. CAMs are also used to match addresses in high speed computer networks.
Magnitude comparison CAM (MCCAM) is defined here as any content addressible memory where the stored data are treated as numbers, and arithmetic magnitude comparisons (i.e. less-than, greater-than, less-than-or-equal-to, and the like) are performed on the data in parallel. This is in contrast to ordinary CAM which treats stored data strictly as bit vectors, not as numbers. One exemplary magnitude comparison content addressable memory is described in, and incorporated herein by reference, is U.S. Patent Number 4,996,666, by Jerome F. Duluk Jr., entitled "Content-Addressable Memory System Capable of Fully Parallel
Magnitude Comparisons", granted February 26, 1991 (hereinafter referred to as the Duluk Patent). Structures within the Duluk Patent specifically referenced shall include the prefix "Duluk Patent" (for example, "Duluk Patent MCCAM Bit Circuit"). Other types of magnitude comparison content addressable memories may also be used. The "MCCAM" abbreviation is conveniently used in this description to refer to various types, structures, and methods for magnitude comparison content addressable memory and is not limited to the particular magnitude comparison content addressable memory described in U.S. Patent
Number 4,996,666. The basic internal structure of an MCCAM is a set of memory bits organized into words, where each word can perform one or more arithmetic magnitude comparisons between the stored data and input data. In general, for an MCCAM, when a vector of numbers is applied in parallel to an array of words, all arithmetic comparisons in all words occur in parallel. Such a parallel search comparison operation is called a "query" of the stored data.
The invention described here augments the capability of the MCCAM by adding various features, including the ability to output all the query result bits every clock cycle and to logically "or" together these output query result bits to form additional outputs. 3 SUMMARY OF THE INVENTION The inventive apparatus and method provide conservative hidden surface removal (CHSR) in a deferred shading graphics pipeline (DSGP). The pipeline renders primitives, and the invention is described relative to a set of renderable primitives that include: 1 ) triangles, 2) lines, and 3) points. Polygons with more than three vertices are divided into triangles in the Geometry block (described hereinafter), but the DSGP pipeline could be easily modified to render quadrilaterals or polygons with more sides. Therefore, since the pipeline can render any polygon once it is broken up into triangles, the inventive renderer effectively renders any polygon primitive. The invention advantageously takes into account whether and in what part of the display screen a given primitive may appear or have an effect. To identify what part of a 3D window on the display screen a given primitive may affect, the pipeline divides the 3D window being drawn into a series of smaller regions, called tiles and stamps. The pipeline performs deferred shading, in which pixel colors are not determined until after hidden-surface removal. The use of a Magnitude Comparison Content
Addressable Memory (MCCAM) advantageously allows the pipeline to perform hidden geometry culling efficiently.
Implementation of the inventive Conservative Hidden Surface Removal procedure, advantageously maintains compatibility with other standard APIs, such as OpenGL®, including their support of dynamic rule changes for the primitives (e.g. changing the depth test or stencil test during a scene). In embodiments of the inventive deferred shader, the conventional rendering paradigm, wherein non-deferred shaders typically execute a sequence of rules for every geometry item and then check the final rendered result, is broken. The inventive structure and method anticipate or predict what geometry will actually affect the final values in the frame buffer without having to make or generate all the colors for every pixel inside of every piece of geometry. In principle, the spatial position of the geometry is examined, and a determination is made for any particular sample, the one geometry item that affects the final color in the z buffer, and then generates only that color.
In one embodiment, the CHSR processes each primitive in time order and, for each sample that a primitive touches, CHSR makes conservative decision based on the various Application Program Interface
(API) state variables, such as depth test and alpha test. One of the advantageous features of the CHSR process is that color computation does not need to be done during hidden surface removal even though non-depth-dependent tests from the API, such as alpha test, color test, and stencil test can be performed by the DSGP pipeline. The CHSR process can be considered a finite state machine (FSM) per sample.
Hereinafter, each per-sample FSM is called a sample finite state machine. Each sample FSM maintains per-sample data including: (1) z coordinate information; (2) primitive information (any information needed to generate the primitive's color at that sample or pixel, or a pointer to such information); and (3) one or more sample state bits (for example, these bits could designate the z value or z values to be accurate or conservative). While multiple z values per sample can be easily used, multiple sets of primitive information per sample would be expensive. Hereinafter, it is assumed that the sample FSM maintains primitive information for one primitive. Each sample FSM may also maintain transparency information, which is used for sorted transparencies.
The DSGP can operate in two distinct modes: 1 ) time order mode, and 2) sorted transparency mode. Time order mode is designed to preserve, within any particular tile, the same temporal sequence of primitives. In time order mode, time order of vertices and modes are preserved within each tile, where a tile is a portion of the display window bounded horizontally and vertically. By time order preserved, we mean that for a given tile, vertices and modes are read in the same order as they are written. In sorted transparency mode, the process of reading geometry from a tile is divided into multiple passes. In the first pass, the opaque geometry(i.e., geometry that can completely hide more distant geometry) is processed, and in subsequent passes, potentially transparent geometry is processed. Within each sorted transparency mode pass, the time ordering is preserved, and mode data is inserted in its correct time-order location. Sorted transparency mode can spatially sort (on a sample-by-sample basis) the geometry into either back- to-front or front-to-back order, thereby providing a mechanism for the visible transparent objects to be blended in spatial order (rather than time order), resulting in a more correct rendering. In a preferred embodiment, the sorted transparency method is performed jointly by the Sort block and the Cull block. The inventive structure and method may be implemented in various embodiments. In one aspect, the invention provides structure and method for performing hidden surface removal wherein the structure is advantageously implemented as a computer graphics pipeline and wherein the inventive hidden surface removal method includes the following steps or procedures. First, an object primitive (current primitive) is selected from a group of primitives, each primitive comprising a plurality of stamps. Next, stamps in the current primitive are compared to stamps from previously evaluated primitives in the group of primitives, and a first stamp is selected from the current primitive by the stamp selection process as a current stamp (CS), and optionally by the SAM for performance reasons. CS is compared to a second stamp or a CPVS selected from previously evaluated stamps that have not been discarded. The second stamp is discarded when no part of the second stamp would affect a final graphics display image based on the comparison with the CS. If part, but not all, of the second stamp would not affect the final image based on the comparison with the CS, then the part of second stamp that would not affect the final image is deleted from the second stamp. The CS is discarded when no part of the second stamp would affect a final graphics display image based on the comparison with the second stamp. If part, but not all, of the CS would not affect the final image based on the comparison with the second stamp, then the part of CS that would not affect the final image is deleted from the CS. When all stamps in all primitives within a region of the display screen have been evaluated, the stamps that have not been discarded have their pixels, or samples, colored by the part of the pipeline downstream from these first steps in performing hidden surface removal. In one embodiment, the set of non-discarded stamps can be limited to one stamp per sample. In this embodiment, when the second stamp and the CS include the same sample and both can not be discarded, the second stamp is dispatched and the CS is kept in the list of non-discarded stamps. Also for this alternate embodiment, when the visibility of the second stamp and the CS depends on parameters evaluated later in the computer graphics pipeline, the second stamp and the CS are dispatched. As an alternate embodiment, the selection of the first stamp by for example the SAM and the stamp selection process, as a current stamp (CS) is based on a relationship test of depth states of samples in the first stamp with depth states of samples of previously evaluated stamps; and an aspect of the inventive apparatus simultaneously performs the relationship test on a multiplicity of stamps.
In another aspect of the inventive structure and method for performing hidden surface removal, a set of currently potentially visible stamps (CPVSs) is maintained separately from the set of current depth values (CDVs), wherein the inventive hidden surface removal method includes the following steps or procedures. First, an object primitive (current primitive) is selected from a group of primitives, each primitive comprising a plurality of stamps. Next, a first stamp from the current primitive is selected as a currently stamp (CS). Next, a currently potentially visible stamp (CPVS) is selected from the set of CPVSs such that the CPVS overlaps the CS. For each sample that is overlapped by both the selected CPVS and the CS, the depth value of the CS is compared to the corresponding value in the set of CDVs, and this comparison operation takes into account the pipeline state and updates the CDVs. Samples in the selected CPVS that are determined to be not visible are deleted for the selected CPVS. If all samples in the selected CPVS are deleted, the selected CPVS is deleted from the set of CPVS's. If any sample in the CS is determined to be visible, the CS is added to the set of the CPVS's with only its visible samples included. If for any sample both the CS and selected CPVS are visible, then at least those visible samples in the selected CPVS are sent down the pipeline for color computations. If the visibility of a sample included in both the CS and CPVS depend on parameters evaluate later in the computer graphics pipeline, at least those samples are sent down the pipeline for color computations. The invention provides structure and method for processing in parallel all CPVS's that overlap the CS. Furthermore, the parallel processing is pipelined such that a CS can be processed at the rate of one CS per clock cycle. Also multiple CS's can be processed in parallel.
In another aspect, the invention provides structure and method for a hidden surface removal system for a deferred shader computer graphics pipeline, wherein the pipeline includes a Magnitude Comparison
Content Addressable Memory (MCCAM) Cull unit for identifying a first group of potentially visible samples associated with a current primitive; a Stamp Selection unit, coupled to the MCCAM cull unit, for identifying, based on the first group and a perimeter of the primitive, a second group of potentially visible samples associated with the primitive; a Z-Cull unit, coupled to the stamp selection unit and the MCCAM cull unit, for identifying visible stamp portions by evaluating a pipeline state, and comparing depth states of the second group with stored depth state values; and a Stamp Portion Memory unit, coupled to the Z-Cull unit, for storing visible stamp portions based on control signals received from the Z-Cull unit, wherein the Stamp Portion Memory unit dispatches stamps having a visibility dependent on parameters evaluated later in the computer graphics pipeline. In yet another aspect, the invention provides structure and method of rendering a graphics image including the steps of: receiving a plurality of primitives to be rendered; selecting a sample location; rendering a front most opaque sample at the selected sample location, and defining the z value of the front most opaque sample as Zfar; comparing z values of a first plurality of samples at the selected sample location; defining to be Znear a first sample, at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the first plurality of samples; rendering the first sample; setting Zfar to the value of Znear; comparing z values of a second plurality of samples at the selected sample location; defining as Znear the z value of a second sample at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the second plurality of samples; and rendering the second sample. BRIEF DESCRIPTION OF THE DRAWINGS Additional objects and features of the invention will be more readily apparent from the following detailed description and appended claims when taken in conjunction with the drawings, in which: Figure 1 is a diagrammatic illustration showing a tetrahedron, with its own coordinate axes, a viewing point's coordinate system, and screen coordinates Figure 2 is a diagrammatic illustration showing a conventional generic renderer for a 3D graphics pipeline. Figure 3 is a diagrammatic illustration showing a first embodiment of the inventive 3-Dimensional
Deferred Shading Graphics Pipeline. Figure 4 is a diagrammatic illustration showing a second embodiment of the inventive 3-Dimensional
Deferred Shading Graphics Pipeline. Figure 5 is a diagrammatic illustration showing a third embodiment of the inventive 3-Dimensional
Deferred Shading Graphics Pipeline. Figure 6 is a diagrammatic illustration showing a fourth embodiment of the inventive 3-Dimensional
Deferred Shading Graphics Pipeline. Figure 7 is a diagrammatic illustration showing a fifth embodiment of the inventive 3-Dimensional
Deferred Shading Graphics Pipeline. Figure 8 is a diagrammatic illustration showing a sixth embodiment of the inventive 3-Dimensional
Deferred Shading Graphics Pipeline. Figure 9 is a diagramatic illustration showing an exemplary flow of data through blocks of an embodiment of the pipeline. Figure 10 is a diagrammatic illustration showing an embodiment of the inventive 3-Dimensional graphics pipeline including information passed between the blocks. Figure 11 is a diagramatic illustration showing the manner in which an embodiment of the Cull block produces fragments from a partially obscured triangle. Figure 12 illustrates a block diagram of the Cull block according to one embodiment of the present invention. Figure 13 illustrates the relationships between tiles, pixels, and stamp portions in an embodiment of the invention. Figure 14 illustrates a detailed block diagram of the Cull block according to one embodiment of the present invention. Figure 15 illustrates a Setup Output Primitive Packet according to one embodiment of the present invention. Figure 16 illustrates a flow chart of a conservative hidden surface removal method according to one embodiment of the present invention. Figure 17A illustrates a sample tile including a primitive and a bounding box.
Figure 17B shows the largest z values (ZMax) for each stamp in the tile. Figure 17C shows the results of the z value comparisons between the ZMin for the primitive and the ZMaxes for every stamp. Figure 18 illustrates an example of a stamp selection process of the conservative hidden surface removal method according to one embodiment of the present invention. Figure 19 illustrates an example showing a set of the left most and right most positions of a primitive in each subraster line that contains at least one sample point,
Figure 20 illustrates a stamp containing four pixels. Figures 21 A - 21 D illustrate an example of the operation of the Z Cull unit.
Figure 22 illustrates an example of how samples are processed by the Z Cull unit. Figures 23A - 23D illustrate an example of early dispatch.
Figure 24 illustrates a sample level example of early dispatch processing. Figure 25 illustrates an example of processing samples with alpha test with a CHSR method according to one embodiment of the present invention. Figure 26 illustrates aspects of stencil testing relative to rendering operations for an embodiment of CHSR. Figure 27 illustrates aspects of alpha blending relative to rendering operations for an embodiment of CHSR. Figure 28A illustrates part of a Spatial Packet containing three control bits: DoAlphaTest, DoABIend and Transparent. Figure 28B illustrates how the alpha values are evaluated to set the DoABIend control bit.
Figure 29 illustrates a flow chart of a sorted transparency mode CHSR method according to one embodiment of the present invention.
4 DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION
The invention is directed to a system for performing Conservative Hidden Surface Removal (CHSR) Deferred Shading Graphics Processor (DSGP) graphics processor. This present invention includes numerous embodiments of the DSGP pipeline. The DSGP pipeline can be connected to a computer via a variety of possible interfaces, including but not limited to for example, an Advanced Graphics Port (AGP) and/or a PCI bus interface, amongst the possible interface choices. VGA and video output are generally also included. Embodiments of the invention supports both OpenGL® and Direct3D APIs. The OpenGL® specification, entitled "The OpenGL® Graphics System: A Specification (Version 1.2)" by Mark Segal and Kurt Akeley, edited by Jon Leech, is included incorporated by reference.
We first describe features that may be implemented in one or more embodiments of the CHSR and then describe several exemplary embodiments or versions of a Deferred Shading Graphics Pipeline.
4.0.1 CHSR and Greater-than Depth Test
Implementation of the Conservative Hidden Surface Removal procedure, advantageously maintains compatibility with other standard APIs, such as OpenGL®. One complication of many APIs is their ability to change the depth test. Typically 3D rendering assumes visible objects are closer than obscured objects (i.e., at lesser z values), and this is accomplished by selecting a "less-than" depth test (i.e., an object is visible if its z value is "less-than" other geometry). However, most APIs support other depth tests, which may change within a frame, such as: greater-than, less-than, greater-than-or-equal-to, equal, less-than-or- equal-to, less-than, not-equal, and the like algebraic, magnitude, and logical relationships. This essentially dynamically "changes the rules" for what is visible, and as a result, the time order of primitives with different rendering rules must be taken into account. In the inventive conservative hidden surface removal system, different or additional procedures are advantageously implemented, to maintain compatibility with other standard APIs when a "greater-than" depth test is used. Analogous changes may advantageously be employed if the depth test is greater-than- or-equal-to, or another functional relationship that would otherwise result in the anomalies described. Further observe that conventional non-deferred shaders execute a sequence of rules for every geometry item and then check the final rendered result. By comparison, in embodiments of the inventive deferred shader, that conventional paradigm is broken. The inventive structure and method anticipate or predict what geometry will actually affect the final values in the frame buffer without having to make or generate all the colors for every pixel inside of every piece of geometry. In principle, the spatial position of the geometry is examined, and a determination is made for any particular sample, the one geometry item that affects the final color in the z buffer, and then generate only that color.
4.0.2 Additional Considerations for the CHSR Process
Samples are processed in parallel, and generally all the samples in all the pixels within a stamp are processed in parallel. Hence, if one stamp can be processed per clock cycle (and there are 4 pixels per stamp and 4 samples per pixel), then 16 samples are processed per clock cycle. A "stamp" defines the number of pixels and samples processed at one time. This per-stamp processing is generally pipelined, with pipeline stalls injected if a stamp needs to be processed again before the same stamp (from a previous primitive) has completed (that is, unless out-of-order stamp processing can be handled).
If no early dispatches are needed, then only end-of-tile dispatches are needed. This is the case when all the geometry in a tile is opaque and there are no stencil tests or operations and there are no alpha tested primitives that could be visible. The primitive information in each sample FSM can be replaced by a pointer into a memory where all the primitive information is stored. As described in later in a preferred embodiment, the Color Pointer is used to point to a primitive's information in Polygon Memory.
As an alternative, only the far z value could be kept (the near z value is not kept), thereby reducing data storage, but requiring the sample state bits to remain "conservative" after primitive F and also causing primitive E to be sent down the pipeline because it would not be known whether primitive E is in front or behind primitive F.
As an alternative to maintaining both a near z value and a far z value, only the far z value could be kept, thereby reducing data storage, but requiring the sample state bits to remain "conservative" when they could have been labeled "accurate", and also causing additional samples to be sent down the pipeline. In the first CHSR example above (the one including alpha test), the sample state bits would remain
"conservative" after primitive F, and also, primitive E would be sent down the pipeline because it would not be known whether primitive E is in front or behind primitive F due to the lack of the near z value.
Processing samples in groups by stamp improves efficiency by allowing sample FSMs to operate in parallel on a stamp-by-stamp basis. Stamps are also used to reduce the number of data packets transmitted down the pipeline. That is, when one sample within a stamp is dispatched (either early dispatch or end-of-tile dispatch), other samples within the same stamp and the same primitive are also dispatched (such a joint dispatch is hereinafter called a Visible Stamp Portion, or VSP). In the second CHSR example above (the one including stencil test), if all four samples were in the same stamp, then the early dispatching of samples 1 and 2 would cause early dispatching of samples 0 and 3. While this causes more samples to be sent down the pipeline, it generally causes a net decrease in the amount of color computation. This is due to the spatial coherence within a pixel (i.e., samples within the same pixel tend to be either visible together or hidden together) and a tendency for the edges of polygons with alpha test, color test, stencil test, and/or alpha blending to potentially split otherwise spatially coherent stamps. That is, sending additional samples down the pipeline when they do not appreciably increase the computational load is more than offset by reducing the total number of VSPs that need to be sent. In the second CHSR example above, if all the samples are in the same stamp, then the same number of VSPs would be generated.
In the case of alpha test, if alpha values for a primitive arise only from the alpha values at the vertices (not from other places such as texturing), then a simplified alpha test can be done for entire primitives. That is, the vertex processing block (which is the Geometry block in a preferred embodiment) can determine when any interpolation of the vertex alpha values would be guaranteed to pass the alpha test, and for that primitive, disable the alpha test. This cannot be done if the alpha values cannot be determined before CHSR is performed. If a frame does not start with depth clear, then the sample FSMs are set to their most conservative state (with near z values at the minimum and far z values at the maximum).
4.0.3 Two Modes of DSGP Operation
The DSGP can operate in two distinct modes: 1 ) time order mode, and 2) sorted transparency mode. Time order mode is described above, and is designed to preserve, within any particular tile, the same temporal sequence of primitives. The Sorted Transparency mode is described immediately below. In a preferred embodiment, the control of the pipeline operating mode is done in the Sort block 6000.
Sort block 6000 is located in the pipeline between a Mode Extraction block (MEX) 4000 and Setup (STP) block 8000. Sort block operates primarily to take geometry scattered around the display window and sort it into tiles. Sort block also manages the Sort Memory, which stores all the geometry from the entire scene before it is rasterized, along with some mode information. Sort memory comprises a double-buffered list of vertices and modes. One page collects a scene's geometry (vertex by vertex and mode by mode), while the other page is sending its geometry (primitive by primitive and mode by mode) down the rest of the pipeline. When a page in sort memory is being written, vertices and modes are written sequentially into the sort memory as they are received by the sort block. When a page is read from sort memory, the read is done on a tile-by-tile basis, and the read process operates in two modes: (1 ) time order mode, and (2) sorted transparency mode.
4.0.4 Time Order Mode In time order mode, time order of vertices and modes are preserved within each tile, where a tile is a portion of the display window bounded horizontally and vertically. By time order preserved, we mean that for a given tile, vertices and modes are read in the same order as they are written. 4.0.5 Sorted Transparency Mode
In sorted transparency mode, the process of reading geometry from a tile is divided into multiple passes. In the first pass, the Sort block outputs guaranteed opaque geometry, and in subsequent passes the Sort block outputs potentially transparent geometry. Within each sorted transparency mode pass, the time ordering is preserved, and mode data is inserted in its correct time-order location. Sorted transparency mode can be performed in either back-to-front or front-to-back order. In a preferred embodiment, the sorted transparency method is performed jointly by the Sort block and the Cull block.
4.0.6 Multiple-step Hidden Surface Removal
Conventionally hidden surfaces are removed using either an "exact" hidden surface removal procedure, or using z buffers. In one embodiment of the inventive structure and method, a two-step approach is implemented wherein a (i) "conservative" hidden surface removal is followed by (ii) a Z buffer based procedure. In a different embodiment, a three-step approach is implemented: (i) a particular spatial
Cull procedure, (ii) conservative hidden surface removal, and (iii) z buffer.
4.1 Versions of the Deferred Shading Graphics Pipeline Several versions or embodiments of the Deferred Shading Graphics Pipeline are described here, and embodiments having various combinations of features may also be implemented. Furthermore, features of the invention may be implemented independently of other features. Most of the important features described above can be applied to all versions of the DSGP pipeline.
4.1.1 Tiles, Stamps, Samples, and Fragments Each frame (also called a scene or user frame) of 3D graphics primitives is rendered into a 3D window on the display screen. A window consists of a rectangular grid of pixels, and the window is divided into tiles (hereinafter tiles are assumed to be 16x16 pixels, but could be any size). If tiles are not used, then the window is considered to be one tile. Each tile is further divided into stamps (hereinafter stamps are assumed to be 2x2 pixels, thereby resulting in 64 stamps per tile, but stamps could be any size within a tile).
Each pixel includes one or more of samples, where each sample has its own color values and z value (hereinafter, pixels are assumed to include four samples, but any number could be used). A fragment is the collection of samples covered by a primitive within a particular pixel. The term "fragment" is also used to describe the collection of visible samples within a particular primitive and a particular pixel.
4.1.2 Deferred Shading
In conventional Z buffer rendering, the renderer calculates the color value (RGB or RGBA) and z value for each pixel of each primitive, then compares the z value of the new pixel with the current z value in the Z-buffer. If the z value comparison indicates the new pixel is "in front of the existing pixel in the frame buffer, the new pixel overwrites the old one; otherwise, the new pixel is thrown away.
Z buffer rendering typically results in a great deal of wasted processing effort if the scene contains many hidden surfaces. In complex scenes, the renderer may calculate color values for ten or twenty times as many pixels as are visible in the final picture. This means the computational cost of any per-pixel operation — such as Phong shading or texture-mapping — is multiplied by ten or twenty. Therefore an improved method of rendering that reduces Z buffer rendering inefficiency is needed. In a pipeline that performs deferred shading, hidden surface removal (HSR) is completed before any pixel coloring is done. The objective of a deferred shading pipeline is to generate pixel colors for only those primitives that appear in the final image (i.e., exact HSR). Eliminating samples before the resource intensive coloring process is performed on the samples improves the efficiency of the rendering process. In deferred shading the primitives should generally be accumulated before HSR begins. For a frame with only opaque primitives, the HSR process determines the single visible primitive at each sample within all the pixels. Once the visible primitive is determined for a sample, then the primitive's color at that sample location is determined. Additional efficiency can be achieved by determining a single per-pixel color for all the samples within the same pixel, rather than computing per-sample colors.
For a frame with at least some alpha blending of primitives (generally due to transparency), there are some samples that are colored by two or more primitives. This means the HSR process must determine a set of visible primitives per sample. Alpha blending is described in greater detail in the aforementioned OpenGL® specification. in some APIs, such as OpenGL®, the HSR process can be complicated by other operations (that is by operations other than depth test) that can discard primitives. These other operations include: pixel ownership test, scissor test, alpha test, color test, and stencil test (as described elsewhere in this specification). Some of these operations discard a primitive based on its color (such as alpha test), which is not determined in a deferred shading pipeline until after the HSR process (this is because alpha values are often generated by the texturing process, included in pixel fragment coloring). For example, a primitive that would normally obscure a more distant primitive (generally at a greater z value) can be discarded by alpha test, thereby causing it to not obscure the more distant primitive. A HSR process that does not take alpha test into account could mistakenly discard the more distant primitive. Hence, there may be an inconsistency between deferred shading and alpha test (similarly, with color test and stencil test); that is, pixel coloring is postponed until after hidden surface removal, but hidden surface removal can depend on pixel colors. Simple solutions to this problem include: 1) eliminating non-depth-dependent tests from the
API, such as alpha test, color test, and stencil test, but this potential solution might prevent existing programs from executing properly on the deferred shading pipeline; and 2) having the HSR process do some color generation, only when needed, but this potential solution would complicate the data flow considerably. Therefore, neither of these choices is attractive. A third alternative, called conservative hidden surface removal (CHSR), is one of the important innovations provided by the inventive structure and method.
Another complication in many APIs is their ability to change the depth test. The standard way of thinking about 3D rendering assumes visible objects are closer than obscured objects (i.e., at lesser z values), and this is accomplished by selecting a "less-than" depth test (i.e., an object is visible if its z value is "less-than" other geometry). However, most APIs support other depth tests such as: greater-than, less- than, greater-than-or-equal-to, equal, less-than-or-equal-to, less-than, not-equal, and the like algebraic, magnitude, and logical relationships. This essentially "changes the rules" for what is visible. This complication is compounded by an API allowing the application program to change the depth test within a frame. Different geometry may be subject to drastically different rules for visibility. Hence, a deferred shading pipeline must preserve the time ordering of primitives, and the correct pipeline state (for example, the depth test) must be associated with each primitive.
4.1.3 Deferred Shading Graphics Pipeline, First Embodiment (Version 1)
A conventional 3D graphics pipeline is illustrated in Figure 2. The following describes a first embodiment of the inventive 3D Deferred Shading Graphics Pipeline Version 1 (hereinafter "DSGPvl"), as shown in Figure 3. It will be observed that the inventive pipeline (Figure 3) has been obtained from the generic conventional pipeline (Figure 2) by replacing the drawing intensive functions 231 with: (1 ) a scene memory 250 for storing the pipeline state and primitive data describing each primitive, called scene memory in the figure; (2) an exact hidden surface removal process 251 ; (3) a fragment coloring process 252; and
(4) a blending process 253. The scene memory 250 stores the primitive data for a frame, along with their attributes, and also stores the various settings of pipeline state throughout the frame. Primitive data includes vertex coordinates, texture coordinates, vertex colors, vertex normals, and the like In DSGPvl , primitive data also includes the data generated by the setup for incremental render, which includes spatial, color, and edge derivatives.
When all the primitives in a frame have been processed by the floating-point intensive functions 213 and stored into the scene memory 250, then the HSR process commences. The scene memory 250 can be double buffered, thereby allowing the HSR process to perform computations on one frame while the floating-point intensive functions perform computations on the next frame. The scene memory can also be triple buffered. The scene memory could also be a scratchpad for the HSR process, storing intermediate results for the HSR process, allowing the HSR process to start before all primitive have been stored into the scene memory. In the scene memory, every primitive is associated with the pipeline state information that was valid when the primitive was input to the pipeline. The simplest way to associate the pipeline state with each primitive is to include the entire pipeline state within each primitive. However, this would introduce a very large amount of redundant information because much of the pipeline state does not change between most primitives (especially when the primitives are in the same object). The preferred way to store information in the scene memory is to keep separate lists: one list for pipeline state settings and one list for primitives. Furthermore, the pipeline state information can be split into a multiplicity of sub-lists, and additions to each sub-list occurs only when part of the sub-list changes. The preferred way to store primitives is done by storing a series of vertices, along with the connectivity information to re-create the primitives. This preferred way of storing primitives eliminates redundant vertices that would otherwise occur in polygon meshes and line strips.
The HSR process described relative to DSGPvl is required to be an exact hidden surface removal (EHSR) because it is the only place in the DSGPvl where hidden surface removal is done. The exact hidden surface removal (EHSR) process 251 determines precisely which primitives affect the final color of the pixels in the frame buffer. This process accounts for changes in the pipeline state, which introduces various complexities into the process. Most of these complications stem from the per-fragment operations (ownership test, scissor test, alpha test, and the like), as described above. These complications are solved by the innovative conservative hidden surface removal (CHSR) process, described later, so that exact hidden surface removal is not required.
The fragment coloring process generates colors for each sample or group of samples within a pixel. This can include: Gouraud shading, texture mapping, Phong shading, and various other techniques for generating pixel colors. This process is different from edge walk 232 and span interpolation 234 because this process must be able to efficiently generate colors for subsections of primitives. That is, a primitive may be partially visible, and therefore, colors need to be generated for only some of its pixels, and edge walk and span interpolation assume the entire primitive must be colored. Furthermore, the HSR process may generate a multiplicity of visible subsections of a primitive, and these may be interspersed in time amongst visible subsections of other primitives. Hence, the fragment coloring process 252 should be capable of generating color values at random locations within a primitive without needing to do incremental computations along primitive edges or along the x-axis or y-axis.
The blending process 253 of the inventive pipeline combines the fragment colors together to generate a single color per pixel. In contrast to the conventional z buffered blend process 236, this blending process 253 does not include Z buffer operations because the exact hidden surface removal process 251 as already determined which primitives are visible at each sample. The blending process 253 may keep separate color values for each sample, or sample colors may be blended together to make a single color for the entire pixel. If separate color values are kept per sample and are stored separately into the Frame buffer 240 , then final pixel colors are generated from sample colors during the scan out process as data is sent to the digital to analog converter 242.
4.1.4 Deferred Shading Graphics Pipeline, Second Embodiment (Version 2)
As described above for DSGPvl , the scene memory 250 stores: (1 ) primitive data; and (2) pipeline state. In a second embodiment of the Deferred Shading Graphics Pipeline 260 (Version 2)
(DSGPv2),illustrated in Figure 4, this scene memory 250 is split into two parts: a spatial memory 261 part and polygon memory 262 part. The split of the data is not simply into primitive data and pipeline state data.
In DSGPv2, the part of the pipeline state data needed for HSR is stored into spatial memory 261 , while the rest is stored into polygon memory 262. Examples of pipeline state needed for HSR include (as defined, for example, in the OpenGL® Specification) are DepthFunc, DepthMask, StencilEnable, etc.
Examples of pipeline state not needed for HSR include: BlendEquation, BlendFunc, stipple pattern, etc. While the choice or identification of a particular blending function (for example, choosing R=RSAS + Ro(1-A_)) is not needed for HSR, the HSR process must account for whether the primitive is subject to blending, which generally means the primitive is treated as not being able to fully occlude prior geometry. Similarly, the HSR process must account for whether the primitive is subject to scissor test, alpha test, color test, stencil test, and other per-fragment operations.
Primitive data is also split. The part of the primitive data needed for HSR is stored into spatial memory 261 , and the rest of the primitive data is stored into polygon memory 262. The part of primitive data needed for HSR includes vertex locations and spatial derivatives (i.e., dz/dx, dz/dy, dx/dy for edges, etc.). The part of primitive data not needed for HSR includes vertex colors, texture coordinates, color derivatives, etc. If per-fragment lighting is performed in the pipeline, the entire lighting equation is applied to every fragment. But in a deferred shading pipeline, only visible fragments require lighting calculations. In this case, the polygon memory may also include vertex normals, vertex eye coordinates, vertex surface tangents, vertex binormals, spatial derivatives of all these attributes, and other per-primitive lighting information. During the HSR process, a primitive's spatial attributes are accessed repeatedly, especially if the
HSR process is done on a per-tile basis. Splitting the scene memory 250 into spatial memory 261 and polygon memory 262 has the advantage of reducing total memory bandwidth.
The output from setup for incremental render 230 is input to the spatial data separation process 263, which stores all the data needed for HSR into spatial memory 261 and the rest of the data into polygon memory 262. The EHSR process 264 receives primitive spatial data (e.g., vertex screen coordinates, spatial derivatives, etc.) and the part of the pipeline state needed for HSR (including all control bits for the per-fragment testing operations). When visible fragments are output from the EHSR process 264, the data matching process 265 matches the vertex state and pipeline state with visible fragments, and tile information is stored in tile buffers 266. The remainder of the pipeline is primarily concerned with the scan out process including sample to/from pixel conversion 267, reading and writing to the frame buffer, double buffered MUX output look-up, and digital to analog (D/A) conversion of the data stored in the frame buffer to the actual analog display device signal values.
4.1.5 Deferred Shading Graphics Pipeline, Third Embodiment (Version 3) In a third embodiment of the Deferred Shading Graphics Pipeline (Version 3) (DSGPv3), illustrated in Figure 5, the scene memory 250 is still split into two parts (a spatial memory 261 and polygon memory 262) and in addition the setup for incremental render 230 is replaced by a spatial setup which occurs after data separation and prior to exact hidden surface removal. The remainder of the pipeline structure and processes are unchanged from those already described relative to the first embodiment.
4.1.6 Deferred Shading Graphics Pipeline, Fourth Embodiment (Version 4)
In a fourth embodiment of the Deferred Shading Graphics Pipeline (Version 4) (DSGPv4), illustrated in Figure 6, the exact hidden surface removal of the third embodiment (Figure 5) is replaced by a conservative hidden surface removal structure and procedure and a down-stream z buffered blend replaces the blending procedure.
4.1.7 Deferred Shading Graphics Pipeline, Fifth Embodiment (Version 5)
In a fifth embodiment of the Deferred Shading Graphics Pipeline (Version 5) (DSGPvδ), illustrated in Figure 7, exact hidden surface removal is used as in the third embodiment, however, the tiling is added, and a tile sorting procedure is added after data separation, and the read is by tile prior to spatial setup. In addition, the polygon memory of the first three embodiments is replaced with a state memory. 4.1.8 Deferred Shading Graphics Pipeline, Sixth Embodiment (Version 6)
In a sixth embodiment of the Deferred Shading Graphics Pipeline (Version 6) (DSGPv6), illustrated in Figure 8, the exact hidden surface removal of the fifth embodiment (Figure 7) is replaced with a conservative hidden surface removal, and the downstream blending of the fifth embodiment is replaced with exact z buffered blending (Testing & Blending).
4.1.9 Other possible Embodiments (Versions)
It should be noted that although several exemplary embodiments of the inventive Graphics Pipeline have been shown and described relative to Figures 3 - 8, persons of ordinary skill in the art in light of the description provided here will readily appreciate that the inventive structures and procedures may be implemented in different combinations and permutations to provide other embodiments of the invention, and that the invention is not limited to the particular combinations specifically identified here.
5 System Level Overview The pipeline takes data from the host computer's I/O bus, processes it, and sends it to the computer's display. The pipeline is divided into twelve blocks, plus three memory stores and the frame buffer. Figure 9 shows the flow of data through the pipeline 1000. The blocks that make up the pipeline are discussed below. Figure 10 shows a method flow diagram of pipeline 1000 which also illustrates the information that is passed between the blocks.
5.1 Command Fetch and Decode (CFD)
Command Fetch and Decode (CFD) 2000 handles communication with the host computer through the I/O bus. It converts its input into a series of packets, which it passes to the Geometry block. Most of the input stream consists of geometrical data — lines, points, and polygons. The descriptions of these geometrical objects can include colors, surface normals, texture coordinates, etc. The input stream also contains rendering information, such as lighting, blending modes, and buffer functions.
5.2 Geometry (GEO) The Geometry block 3000 handles four major tasks: transforms, decomposition of all polygons into triangles, clipping, and per-vertex lighting calculations needed for Gouraud shading. First, the Geometry block transforms incoming graphics primitives into a uniform coordinate space
("world space"). Then it clips the primitives to the viewing volume, or frustum. In addition to the six planes that define the viewing volume (left, right, top, bottom, front and back), the DSGP pipeline provides six user- definable clipping planes. After clipping, the Geometry block breaks polygons with more than three vertices into sets of triangles, to simplify processing.
Finally, if there is any Gouraud shading in the frame, the Geometry block calculates the vertex colors that the Fragment block uses to perform the shading. 5.3 Mode Extraction (MEX) The Mode Extraction block 4000 separates the data stream into two parts: 1 ) vertices, and
2) everything else. Vertices are sent to the Sort block. The "everything else" — lights, colors, texture coordinates, etc. — is stored in a special buffer called the Polygon Memory, where it can be retrieved by the Mode Injection block. The Polygon Memory is double buffered, so the Mode Injection block can read data for one frame, while the Mode Extraction block is storing data for the next frame. The mode data stored in the Polygon Memory falls into three major categories: per-frame data (such as lighting), per-primitive data (such as material properties) and per-vertex data (such as color). The Mode Extraction and Mode injection blocks further divide these categories to optimize efficiency. For each vertex, the Mode Extraction block sends the Sort block a packet containing the vertex data and a pointer into the Polygon Memory. The pointer is called the Color Pointer, although it is used to retrieve all sorts of other information in addition to color. The packet also contains fields indicating whether the vertex represents a point, the endpoint of a line, or the corner of a triangle. The vertices are sent in a strict time sequential order, the same order in which they were fed into the pipeline. The packet also specifies whether the current vertex forms the last one in a given primitive (i.e., "completes" the primitive). In the case of triangle strips or fans, and line strips or loops, the vertices are shared between adjacent primitives. In this case, the packets indicate how to identify the other vertices in each primitive.
5.4 Sort (SRT) The Sort block 6000 receives vertices from the Mode Extraction block and sorts the resulting points, lines, and triangles by tile. In the double-buffered Sort Memory 7000, it maintains a list of vertices representing the graphic primitives, and a set of Tile Pointer Lists-one list for each tile in the frame. When
Sort block 6000 receives a vertex that completes a primitive (such as the third vertex in a triangle), it checks to see which tiles the primitive touches. For each tile a primitive touches, the Sort block adds a pointer to the vertex to that tile's Tile Pointer List. When the Sort block has finished sorting all the geometry in a frame, it sends the data to Setup block 8000. Each Sort block output packet represents a complete primitive. Sort sends its output in tile-by- tile order: all of the primitives that touch a given tile, then all of the primitives that touch the next tile, and so on. Note that this means that Sort may send the same primitive many times, once for each tile it touches.
5.5 Setup (STP) The Setup block 8000 calculates spatial derivatives for lines and triangles. Setup block 8000 processes one tile's worth of data, one primitive at a time. When Setup block 8000 is done with a primitive, it sends the data on to the Cull block. The Setup block also breaks stippled lines into separate line segments
(each a rectangular region), and computes the minimum z value for each primitive within the tile. The Setup block outputs Setup Output Primitive Packets, each of which represents one primitive: a triangle, line segment or point.
5.6 Cull (CUL) Cull block 9000 performs the conservative hidden surface removal process. The Cull block 9000 processing is divided into two steps: Magnitude Comparison Content Addressable Memory (MCCAM) Cull, and Subpixel Cull. The Cull block accepts data one tile's worth at a time. The MCCAM Cull discards primitives that are hidden completely by previously processed geometry. The Subpixel Cull takes the remaining primitives (which are partly or entirely visible), and determines the visible fragments. The Subpixel Cull outputs one stamp's worth of fragments at a time, called a Visible Stamp Portion (VSP). Figure 11 shows an example of how the Cull block produces fragments from a partially obscured triangle. Each pixel in a VSP is divided up into a number of samples to determine how much of the pixel is covered by a given fragment. The Pixel block uses this information when it blends the fragments to produce the final color for the pixel.
5.7 Mode Injection (MU)
The Mode Injection block 10000 retrieves mode information — such as colors, material properties, and so on— from the Polygon Memory 5000 and passes it downstream as required. To save bandwidth, the individual downstream blocks cache recently used mode information. The Mode Injection block keeps track of what information is cached downstream, and only sends information as necessary.
5.8 Fragment (FRG)
The Fragment block 11000 primarily performs interpolation. Fragment block 11000 interpolates color values for Gouraud shading, surface normals for Phong shading and texture coordinates for texture mapping. Fragment block 11000 also interpolates surface tangents for use in the bump mapping algorithm, if bump maps are in use. In addition, the Fragment block performs perspective corrected interpolation using barycentric coefficients.
5.9 Texture (TEX) The Texture block 12000 applies texture maps to the pixel fragments. Texture maps are stored in the Texture Memory 13000. Unlike the other memory elements described previously, the Texture Memory is single-buffered. It is loaded from the host computer's memory using the AGP interface. A single polygon can use up to four textures. Textures are mip-mapped. That is, each texture comprises a series of texture maps at different levels of detail, each map representing the appearance of the texture at a given distance from the eye point. To produce a texture value for a given pixel fragment, the Texture block performs tri-linear interpolation from the texture maps, to approximate the correct level of detail. The Texture block also performs other interpolation methods, such as anisotropic interpolation.
The Texture block supplies interpolated texture values (generally as RGBA color values) to the Phong block on a per-fragment basis. Bump maps represent a special kind of texture map. Instead of a color, each texel of a bump map contains a height field gradient. 5.10 Phong (PHG) The Phong block 14000 performs Phong shading for each pixel fragment. It uses the material and lighting information supplied by the Mode Injection block, the texture colors from the Texture block, and the surface normal generated by the Fragment block to determine the fragment's apparent color. If bump mapping is in use, the Phong block uses the interpolated height field gradient from the Texture block to perturb the fragment's surface normal before shading.
5.11 Pixel (PIX)
The Pixel block 15000 receives VSPs, where each fragment has an independent color value. The Pixel block performs pixel ownership test, scissor test, alpha test, stencil operations, depth test, blending, dithering and logic operations on each sample in each pixel (see OpenGL® Spec 1.1 , Section 4.1 , "Per- Fragment Operations," p. 109). When the Pixel block has accumulated a tile's worth of finished pixels, it blends the samples within each pixel (thereby performing antialiasing of pixels) and sends them to the Backend, to be stored in the Frame Buffer 17000.
To process fragments having common sample locations, the Pixel block blends the fragments and the background color to produce the final pixels. The Pixel block weights each fragment according to how much of the pixel it covers-or to be more precise, by the number of samples it covers. In addition to this blending process, the Pixel block performs stencil testing, alpha blending, and antialiasing of pixels. When the Pixel block accumulates a tile's worth of finished pixels, it sends them to the Backend, to be stored in the Frame Buffer.
5.12 Backend (BKE)
The Backend 16000 receives a tile's worth of pixels at a time from the Pixel block, and stores them into the Frame Buffer 17000. The Backend also sends a tile's worth of pixels back to the Pixel block, because specific Frame Buffer values can survive from frame to frame. For example, stencil bit values can be constant over many frames, but can be used in all those frames.
In addition to controlling the Frame Buffer, the Backend performs 2D drawing and sends the finished frame to the output devices. The Backend provides the interface between the Frame Buffer and the computer monitor and video output.
5.13 Communication Between Blocks in the Pipeline The inventive structure and method provides for packetized communication between the functional blocks of the pipeline. The nature of these packets is described in the individual functional block sections.
6 Cull Block Overview Figure 12 illustrates a block diagram of Cull block 9000. The Cull block is responsible for: 1 ) pre- shading hidden surface removal; and 2) breaking down primitive geometry entities (triangles, lines and points) to stamp based geometry entities called Visible Stamp Portions (VSPs). The Cull block does, in general, a conservative culling of hidden surfaces. To facilitate the conservative hidden surface removal process Cull block 9000 does not handle some "fragment operations" such as alpha test and stencil test. Z Cull 9012 can store two depth values per sample, but Z Cull 9012 only stores the attributes of one primitive per sample. Thus, whenever a sample requires blending colors from two pieces of geometry, the
Cull block sends the first primitive (using time order) down the pipeline, even though there may be later geometry that hides both pieces of the blended geometry.
The Cull block receives input in the form of packets from the Setup block 8000. One type of packet received by the Cull block is a mode packet. Mode packets provide the Cull block control information including the start of a new tile, a new frame, and the end of a frame. Cull block 9000 also receives Setup Output Primitive Packets. The Setup Output Primitive Packets each describe, on a per tile basis, either a triangle, a line or a point. The data field in Setup Output Primitive Packets contain bits to indicate the primitive type (triangle, line or point). The interpretation of the rest of the geometry data field depends upon the primitive type. A non-geometry data field contains the Color Pointer and mode bits that control the culling mode that can be changed on a per primitive bases. Mode packets include mode bits that indicate whether alpha test is on, whether Z buffer write is enabled, whether culling is conservative or accurate, whether depth test is on, whether blending is on, whether a primitive is anti-aliased and other control information. Sort block 6000 bins the incoming geometry entities to tiles. Setup block 8000 pre-processes the primitives to provide more detailed geometric information for the Cull block to do the hidden surface removal. Setup block 8000 pre-calculates the slope value for all the edges, the bounding box of the primitive within the tile, minimum depth value (front most) of the primitive within the tile, and other relevant data. Prior to Sort, Mode Extraction block 4000 has already extracted the color, light, texture and related mode data, the Cull block only gets the mode data that is relevant to the Cull block and a pointer, called Color Pointer, that points to color, light and texture data stored in Polygon Memory 5000.
The Cull block performs two main functions. The primary function is to remove geometry that is guaranteed to not affect the final results in Frame Buffer 17000 (i.e., a conservative form of hidden surface removal). The second function is to break primitives into units of Visible Stamp Portions (VSP). A stamp portion is the intersection of a primitive with a given stamp. A VSP is a visible portion of a geometry entity within a stamp. In one embodiment, each stamp is comprised of four pixels, and each pixel has four predetermined sample points. Thus each stamp has 16 predetermined sample points. The stamp portion
"size" is then given by the number and the set of sample points covered by a primitive in a given stamp. The Cull block sends one VSP at a time to the Mode Injection block 10000. Mode Injection block
10000 reconnects the VSP with its color, light and texture data and sends it to Fragment 11000 and later stages in the pipeline.
The Cull block processes primitives one tile at a time. However, for the current frame, the pipeline is in one of two modes: 1 ) time order mode; or 2) sorted transparency mode. In time order mode, the time order of vertices and modes are preserved within each tile, and the tile is processed in a single pass through the data. That is, for a given tile, vertices and modes are read in the same order as they are written, but are skipped if they do not affect the current tile. In sorted transparency mode, the processing of each tile is divided into multiple passes, where, in the first pass, guaranteed opaque geometry is processed (the Sort block only sends non-transparent geometry for this pass). In subsequent passes, potentially transparent geometry is processed (the Sort block repeatedly sends all the transparent geometry for each pass). Within each pass, the time ordering is preserved, and mode data is inserted in its correct time-order location.
In time order mode, when there is only "simple opaque geometry" (i.e. no scissor testing, alpha testing, color testing, stencil testing, blending, or logicop) in a tile, the Cull block will process all the primitives in the tile before dispatching any VSPs to Mode Injection. This is because the Cull block hidden surface removal method can unambiguously determine, for each sample, the single primitive that covers (i.e., colors) that sample. The case of "simple opaque geometry" is a typically infrequent special case. In time order mode, when the input geometry is not limited to "simple opaque geometry" within a tile, this may cause early dispatch of VSPs (an entire set of VSPs or selected VSPs). However, without exception all the VSPs of a given tile are dispatched before any of the VSPs of a different tile can be dispatched. In general, early dispatch is performed when more than one piece of geometry could possibly affect the final tile values (determined by Pixel block 15000) for any sample.
In sorted transparency mode, each tile is processed in multiple passes (assuming there is at least some transparent geometry in the tile). In each pass, there is no early dispatch of VSPs.
If the input packet is a Setup Output Primitive Packet, a PrimType parameter indicates the primitive type (triangle, line or point). The spatial location of the primitive (including derivatives, etc.) is done using a "unified description". That is, the packet describes the primitive as a quadrilateral (not screen aligned), and triangles and points are degenerate cases. This "unified description" is described in more detail in the provisional patent application entitled "Graphics Processor with Deferred Shading," filed August 20, 1998, which is hereby incorporated by reference. The packet includes a color pointer, used by Mode Injection.
The packet also includes several mode bits, many of which can change primitive by primitive. The following are considered to be "mode bits", and are input to state machines in Z Cull 9012: CullFlushOverlap,
DoAlphaTest; DoABIend, DepthFunc, DepthTestEnabled, DepthTestMask, and NoColor. In addition to Setup Output Primitive Packets, Cull block 9000 receives the following packet types:
Setup Output Clear Packet, Setup Output Cull Packet, Setup Output Begin Frame Packet, Setup Output End Frame Packet, Setup Output Begin Tile Packet, and Setup Output Tween Packet. Each of these packet types is described in detail in the Detailed Description of Cull Block section. But, collectively, these packets are referred to as "mode packets."
In operation, when Cull block 9000 receives a primitive, Cull attempts to eliminate it by querying the Magnitude Comparison Content Addressable Memory (MCCAM) Cull 9002, shown in Figure 12, with the primitive's bounding box. If MCCAM Cull 9002 indicates that a primitive is completely hidden within the tile, then the primitive is eliminated. If MCCAM Cull 9002 cannot reject the primitive completely, it will generate a stamp list, each stamp in the list may contain a portion of the primitive that may be visible. This list of potentially visible stamps is sent to the Stamp Selection Logic 9008 of Cull block 9000. Stamp Selection
Logic 9008 uses the geometry data of the primitive to determine the set of stamps within each stamp row of the tile that are actually touched by the primitive. Combined with the stamp list produced by MCCAM Cull
9002, the Stamp Selection Logic unit dispatches one potentially visible stamp 9006 at a time to the Z Cull block 9012. Each stamp is divided into a grid of 16 by 16 sub-pixels. Each horizontal grid line is called a subraster line. Each of the 16 sample points per stamp has to fall (for antialiased primitives) at the center of one of the 256 possible sub-pixel locations. Each pixel has four sample points within its boundary, as shown with stamp 9212 in Figure 13A. (Figure 13B and Figure 13C illustrate the manner in which the Stamp Portion is input into the Z-Cull process and as stored in SPM, respectively.) Sample locations within pixels can be made programmable. With programmable sample locations, multiple processing passes can be made with different sample locations thereby increasing the effective number of samples per pixel. For example, four passes could be performed with four different sets of sample locations, thereby increasing the effective number of samples per pixel to fourteen.
The display image is divided into tiles to more efficiently render the image. The tile size as a fraction of the display size can be defined based upon the graphics pipeline hardware resources.
The process of determining the set of stamps within a stamp row that is touched by a primitive involves calculating the left most and right most positions of the primitive in each subraster line that contains at least one sample point. These left most and right most subraster line positions are referred to as XleftSubS, and XrightSubS, which stands for x left most subraster line for sample i and x right most subraster line for sample i respectively. Samples are numbered from 0 to 15. The determination of XleftSubS, and XrightSubS, is typically called the edge walking process. If a point on an edge (xO, yO) is known, then the value of x1 corresponding to the y position of y1 can easily be determined by:
dx x1 = xO + (y1 -yO) * — dy
In addition to the stamp number, the set of 16 pairs of XleftSubS, and XrightSubS, is also sent by the Stamp Selection Logic unit to Z Cull 9012.
Z Cull unit 9012 receives one stamp number (or StamplD) at a time. Each stamp number contains a portion of a primitive that may be visible as determined by MCCAM Cull 9002. The set of 16 pairs of
XleftSubS, and XrightSubS, are used to determine which of the 16 sample points are covered by the primitive. Sample i is covered if Xsample,, the x coordinate value of sample i satisfies:
XleftSubS, ≤ Xsample, < XrightSubS,
For each sample that is covered, the primitive's z value is computed at that sample point. At the same time, the current z values and z states for all 16 sample points are read from the Sample Z buffer
9055. Each sample point can have a z state of "conservative" or "accurate". Alpha test, and other tests, are performed by pipeline stages after Cull block 9000. Therefore, for example, a primitive that may appear to affect the final color in the frame buffer based on depth test, may in fact be eliminated by alpha test before the depth test is performed, and thus the primitive does not affect the final color in the frame buffer. To account for this, the Cull block 9000 uses conservative z values. A conservative z value defines the outer limit of a z value for a sample based on the geometry that has been processed up to that point. A conservative z value means that the actual z value is either at that point or at a smaller z value. Thus the conservative z is the maximum z value that the point can have. If the depth test is render if greater than, then the conservative z value is a minimum z value. Conversely, if the depth test is render if less than, then the conservative z value is a maximum z value. For a render if less than depth test, any sample for a given sample location, with a z value less than the conservative z is thus a conservative pass because it is not known at that point in the processes whether it will pass. An accurate z value is a value such that the surface which that z represents is the actual z value of the surface. With an accurate z it is known that the z value represents a surface that is known to be visible and anything in front of it is visible and everything behind it is obscured, at that point in the process.
The status of a sample is maintained by a state machine, and as the process continues the status of a sample may switch between accurate and conservative. In one embodiment, a single conservative z value is used. In another embodiment, two z values are maintained for each sample location, a near z value (Znear) and a far z value (Zfar). The far z value is a conservative z value, and the near z value is an optimistic z value. Using two z values allows samples to be determined to be accurate again after being labeled as conservative. This improves the efficiency of the pipeline because an accurate z value can be used to eliminate more geometry than a conservative z value. For example, if a sample is received that is subject to alpha test, in the Cull block it is not known whether the sample will be eliminated due to alpha test.
In an embodiment where only one z value is stored, the z value may have to be made conservative if the position of the sample subject to alpha test would pass the depth test. The sample that is subject to alpha test is then sent down the pipeline. Since, the sample subject to alpha test is not kept, the z value of the stored sample cannot later be converted back to accurate. By contrast, in an embodiment where two z values are stored, the sample subject to alpha test can, depending on its relative position, be stored as the Zfar/Znear sample. Subsequent samples can then be compared with the sample subject to alpha test as well as the second stored sample. If the Cull block determines, based on the depth test, that one of the subsequent samples, such as an opaque sample in front of the sample subject to alpha test, renders the sample subject to alpha test not visible, then that subsequent sample can be labeled as accurate. In OpenGL® primitives are processed in groups. The beginning and ending of a group of pimitives are identified by the commands, begin and end respectively. The depth test is defined independently for each group of primitives. The depth test is one component of the pipeline state.
Each sample point has a Finite State Machine (FSM) independent of other samples. The z state combined with the mode bits received by Cull drive the sample FSMs. The sample FSMs control the comparison on a per sample basis between the primitive's z value and the Z Cull 9012 z value. The result of the comparison is used to determine whether the new primitive is visible or hidden at each sample point that the primitive covers. The maximum of the 16 sample points' z value is used to update the MCCAM Cull 9002.
A sample's FSM also determines how the Sample Z Buffer in Z Cull 9012 should be updated for that sample, and whether the sample point of the new VSP should be dispatched early. In addition, the sample FSM determines if any old VSP that may contain the sample point should be destroyed or should be dispatched early. For each sample Z Cull 9012 generates four control bits that describe how the sample should be processed, and sends them to the Stamp Portion Mask unit 9014. These per sample control bits are: SendNew, KeepOld, SendOld, and NewVSPMask. If the primitive contains a sample point that is visible, then a NewVSPMask control bit is asserted which causes Stamp Portion Memory (SPM) 9018 to generate a new VSP coverage mask. The remaining three control bits determine how SPM 9018 updates the VSP coverage mask for the primitive. In sorted transparency mode, geometry is spatially sorted on a per-sample basis, and, within each sample, is rendered in either back-to-front or front-to-back order. In either case, only geometry that is determined to be in front of the front-most opaque geometry needs to be send down the pipeline, and this determination is done in Cull 9012. In back-to-front sorted transparency mode, transparent primitives are rasterized in spatial order starting with the layer closest to the front most opaque layer instead of the regular mode of time order rasterization. Two z values are used for each sample location, Zfar and Znear. In sorted transparency mode the transparent primitives go through Z Cull unit 9012 several times. In the first pass, Sort block 6000, illustrated in Figure 9, sends only the opaque primitives. The z values are updated as described above.
The z values for opaque primitives are referred to as being of type Zfar. At the end of the pass, the opaque VSPs are dispatched. The second time Sort block 6000 only sends the transparent primitives for the tile to Cull block 9000. Initially the Znear portion of the Sample Z Buffer are preset to the smallest z value possible. A sample point with a z value behind Zfar is hidden, but a z value in front of Zfar and behind Znear is closer to the opaque layer and therefore replaces the current Znear's z value. This pass determines the z value of the layer that is closest to the opaque layer. The VSPs representing the closest to opaque layer are dispatched. The roles of Znear and Zfar are then switched, and Z Cull receives the second pass of transparent primitives. This process continues until Z Cull determines that it has processed all possible layers of transparent primitives. Z Cull in sorted transparent mode is also controlled by the sample finite state machines.
In back-to-front sorted transparency mode, for any particular tile, the number of transparent passes is equal to the number of visible transparent surfaces. The passes can be done as: a) The Opaque Pass (there is only one Opaque Pass) does the following: the front-most opaque geometry is identified (labeled Zfar) and sent down the pipeline. b) The first Transparent Pass does the following: 1) at the beginning of the pass, keep the Zfar value from the Opaque Pass, and set Znear to zero; 2) identifies the back-most transparent surface between Znear (initialized to zero at the start of the pass) and Zfar; 2) determine the new Znear value; and, 3) at the end of the pass, send this back-most transparent surface down the pipeline. c) The subsequent passes (second Transparent Pass, etc.) do the following: 1) at the beginning of the pass, set the Zfar value to the Znear value from the last pass, and set Znear to zero; 2) identify the next farthest transparent surface between Znear and Zfar; 3) determine the new
Znear value; and, 4) at the end of the pass, send this backmost transparent surface down the pipeline.
In front-to-back sorted transparency mode, for any particular tile, the number of transparent passes can be limited to a preselected maximum, even if the number of visible transparent surfaces at a sample is greater. The passes can be done as: a) In the First Opaque Pass (there are two opaque passes, the other one is the Last Opaque Pass), the front-most opaque geometry is identified (labeled Zfar), but this geometry is not sent down the pipeline, because, only the z-value is valuable in this pass. This Zfar value is the boundary between visible transparent layers and hidden transparent layers. This pass is done with the time order mode sample FSM. b) The next pass, the first Transparent Pass, renders the front-most transparent geometry and also counts the number of visible transparencies at each sample location. This pass does the following: 1 ) at the beginning of the pass, set the Znear value to the Zfar value from the last pass, set Zfar to the maximum z-value, and initialize the NumTransp counter in each sample to zero; 2) test all transparent geometry and identify the front-most transparent surface by finding geometry that is in front of both Znear and Zfar; 3) as geometry is processed, determine the new Zfar value, but don't change the Znear value; 4) count the number of visible transparent surfaces by incrementing NumTransp when geometry that is in front of Znear is encountered; and, 5) at the end of the pass, send this front-most transparent surface down the pipeline. NOTE: concpetually, this pass is defined in an unusual way, because, at the end, Zfar is nearer than Znear; but this allows the rule, "set the Znear value to the Zfar value from the last pass, and set Zfar to the maximum z-value" to be true for every transparent pass. If this is confusing, the definition of Znear and Zfar can be swapped, but this changes the definition of the second transparent pass. c) Subsequent Transparent Passes determine progressively farther geometry, and the maximum number of transparent passes is specified by the MaxTranspPasses parameter. Each of these passes does the following: 1 ) at the beginning of the pass, set the Znear value to the Zfar value from the last pass, set Zfar to the maximum z-value, and the NumTransp counter in each sample is not changed; 2) test all transparent geometry and identify the next-front-most transparent surface by finding the front-most geometry that is between Znear and Zfar, but discard all the transparent geometry if all of the visible transparent layers have been found for this sample (i.e., NumTranspPass > NumTransp); 3) as geometry is processed, determine the new Zfar value, but don't change the Znear value; and, 4) at the end of the pass, send this second-most transparent surface down the pipeline. d) For the Last Opaque Pass, the front-most opaque geometry is again identified, but this time, the geometry is sent down the pipeline. This pass does the following: 1 ) at the beginning of the pass, set Zfar to the maximum z-value (Znear is not used), and the NumTransp counter in each sample is not changed; 2) test all opaque geometry and identify the front-most geometry, using the time order mode sample FSM; 3) as geometry is processed, determine the new Zfar value, but discard the geometry if SkipOpaquelfMaxTransp is TRUE and the maximum number of transparent layers was found (i.e., MaxTranspPasses = NumTransp); and 4) at the end of the pass, send this front-most opaque surface down the pipeline.
The efficiency of CUL is increased (i.e., fewer fragments sent down the pipeline) in front-to-back sorted transparency mode, especially when there are lots of visible depth complexity for transparent surfaces. Also, this may enhance image quality by allowing the user to discern the front-most N transparencies, rather than all those in front of the front-most opaque surface. The stamp portion memory block 9018 contains the VSP coverage masks for each stamp in the tile. The maximum number of VSPs a stamp can have is 16. The VSP masks should be updated or dispatched early when a new VSP comes in from Z Cull 9012. The Stamp Portion Mask unit performs the mask update or dispatch strictly depending on the SendNew, KeepOld and SendOld control bits. The update should occur at the same time for a maximum of 16 old VSPs in a stamp because a new VSP can potentially modify the coverage mask of all the old VSPs in the stamp. The Stamp Portion Data unit 9016 contains other information associated with a VSP including but not limited to the Color Pointer. The Stamp Portion Data memory also needs to hold the data for all VSPs contained in a tile. Whenever a new VSP is created, its associated data need to be stored in the Stamp Portion Data memory. Also, whenever an old VSP is dispatched, its data need to be retrieved from the Stamp Portion Data memory.
7 Detailed Description of Cull Block
Figure 14 illustrates a detailed block diagram of Cull block 9000. Cull block 9000 is composed of the following components: Input FIFO 9050, MCCAM Cull 9002, Subrasterizer 9052, Column Selection
9054, MCCAM Update 9059, Sample Z buffer 9055, New VSP Queue 9058, Stamp Portion Memory Masks 9060 and 9062, Stamp Portion Memory Data units 9064 and 9066, Dispatch Queues 9068 and 9070, and
Dispatch Logic 9072.
7.0.1 Mode and Data Packets The operation of the Cull components is determined by the packets received by the Cull block. The following describes the mode packets: A Setup Output Clear Packet indicates some type of buffer clear is to be performed.
However, buffer clears that occur at the beginning of a user frame (and not subject to scissor test) are included in a Begin Tile packet.
The Setup Output Cull Packet is a packet of mode bits. This packet includes: 1 ) bits for enabling/disabling the MCCAM Cull and Z Cull processes; 2) a bit, CullFlushAII, that causes a flush of all the VSPs from the Cull block; and 3) the bits: AliasPolys, AliasLines, and AliasPoints, which disable antialiasing for the three types of primitives.
The Setup Output Begin Frame Packet tells Cull that a new frame is starting. The next packet will be a Sort Output Begin Tile Packet. The Setup Output Begin Frame Packet contains all the per-frame information that is needed throughout the pipeline. The Setup Output End Frame Packet indicates the frame has ended, and that the current tile's input has been completed. The Setup Output Begin Tile Packet tells the Cull block that the current tile has ended and that the processed data should be flushed down the pipeline. Also, at the same time, the Cull block should start to process the new tile's primitives. If a tile is to be repeated due to the pipeline being in sorted transparency mode, then this requires another Setup Output Begin Tile Packet. Hence, if a particular tile needs an opaque pass and four transparent passes, then a total of five begin tile packets are sent from the Setup block. This packet specifies the location of the tile within the window. The Setup Output Tween Packet can only occur between (hence 'tween) frames, which, of course is between tiles. Cull treats this packet as a black box, and just passes it down the pipeline. This packet has only one parameter, TweenData, which is 144 bits. in addition to the mode packets, the Cull block also receives Setup Output Primitive Packets, as illustrated in Figure 15. The Setup Output Primitive Packets each describe, on a per tile basis, either a triangle, a line, or a point. More particularly, the data field in Setup Output Primitive Packets contain bits to indicate the primitive type (triangle, line, or point). The interpretation of the rest of the geometry data field depends upon the primitive type. If the input packet is a Setup Output Primitive Packet, a PrimType parameter indicates the primitive type (triangle, line or point). The spatial location of the primitive (including derivatives, etc.) is specified using a unified description. That is, the packet describes the primitive as a quadrilateral (non-screen aligned), no matter whether the primitive is a quadrilateral, triagle, or point, and triangles and points are treated as degenerate cases of the quadralateral. The packet includes a color pointer, used by the Mode
Injection unit. The packet also includes several mode bits, many of which can change state on a primitive by primitive basis. The following are considered to be "mode bits", and are input to state machines in Z Cull
9012: CullFlushOverlap, DoAlphaTest; DoABIend, DepthFunc, DepthTestEnabled, DepthTestMask, and NoColor.
The Cull components are described in greater detail in the following sections.
7.0.2 Input FIFO Figure 16 illustrates a flow chart of a conservative hidden surface removal method using the Cull block 9000 components shown in the Figure 14 detailed block diagram. Input FIFO unit 9050 interfaces with the Setup block 8000. Input FIFO 9050 receives data packets from Setup and stores each packet in a queue, step 9160. The number of FIFO memory locations needed is between about sixteen and about 32, in one embodiment the depth is assumed to be sixteen.
7.0.3 MCCAM Cull
The MCCAM Cull unit 9002 uses an MCCAM array 9003 to perform a spatial query on a primitive's bounding box to determine the set of stamps within the bounding box that may be visible. The Setup block
8000 determines the bounding box for each primitive, and determines the minimum z value of the primitive inside the current tile, which is referred to as ZMin. Figure 17A illustrates a sample tile including a primitive
9254 and a bounding box 9252 in MCCAM. MCCAM Cull 9002 uses ZMin to perform z comparisons. MCCAM Cull 9002 stores the maximum z value per stamp of all the primitives that have been processed.
MCCAM Cull 9002 then compares in parallel ZMin for the primitive with all the ZMaxes for every stamp. Based on this comparison, MCCAM Cull determines (a) whether the whole primitive is hidden, based on all the stamps inside the simple bounding box; or (b) what stamps are potentially visible in that bounding box, step 9164. Figure 17B shows the largest z values (ZMax) for each stamp in the tile. Figure 17C shows the results of the comparison. Stamps where ZMin ≤ ZMax are indicated with a one, step 9166. These are the potentially visible stamps. MCCAM Cull also identifies each row which has a stamp with ZMin ≤ ZMax, step 9168. These are the rows that the Stamp Selection Logic unit 9008 needs to process. Stamp Selection Logic unit 9008 skips the rows that are identified with a zero.
MCCAM Cull can process one primitive per cycle from the input FIFO 9050. Read operations from the FIFO occur when the FIFO is not empty and either the last primitive removed is completely hidden as determined by MCCAM Cull or the last primitive is being processed by the Subrasterizer unit 9052. In other words, MCCAM Cull does not "work ahead" of the Subrasterizer. Rather, MCCAM Cull only gets the next primitive that the Subrasterizer needs to process, and then waits.
In an alternative embodiment, Cull block 9000 does not include an MCCAM Cull unit 9002. In this embodiment, the Stamp Selection Logic unit 9008 processes all of the rows.
7.0.4 Subrasterizer within the Stamp Selection Logic Subrasterizer 9052 is the unit that does the edge walking (actually, the computation is not iterative, as the term "walking" would imply). Each cycle, Subrasterizer 9052 obtains a packet from MCCAM Cull 9002. One type of packet received by the Cull block is the Setup Output Primitive Packet, illustrated in
Figure 15. Setup Output Primitive Packets include row numbers and row masks generated by MCCAM Cull 9002 which indicate the potentially visible stamps in each row. Subrasterizer 9052 also receives the vertex and slope data it needs to compute the the left most and right most positions of the primitive in each subraster line that contains at least one sample point, XleftSubS, and XrightSubS, . Subrasterizer 9052 decodes the PrimitiveType field in the Setup Output Primitive Packet to determine if a primitive is a triangle, a line or a point, based on this information Subrasterizer 9052 determines whether the primitive is antialiased. Referring to Figure 18, for each row of stamps that MCCAM Cull indicates is potentially visible (using the row selection bits 9271 ), Subrasterizer 9052 simultaneously computes the XleftSub, and
XrightSub, for each of the sample points in the stamp, in a preferred embodiment there are 16 samples per stamp, step 9170. Each pair of XleftSub, and XrightSub, define a set of stamps in the row that is touched by the primitive, which are referred to as a sample row mask. For example, Figure 19 illustrates a set of XleftSub, and XrightSub,.
Referring to Figure 18, each stamp in the potentially visible rows that is touched by the primitive is indicated by setting the corresponding stamp coverage bit 9272 to a one ('1"), as shown in tile 9270.
Subrasterizer 9052 logically OR's the sixteen row masks to get the set of stamps touched by the primitive. Subraster 9052 then ANDs the touched stamps with the stamp selection bits 9278, as shown in tile 9276, to form one touched stamp list, which is shown in tile 9280, step 9172. The Subrasterizer passes a request to MCCAM Cull for each stamp row, and receives a potentially visible stamp list from MCCAM Cull. The visible stamp list is combined with the touched stamp list, to determine the final potentially visible stamp set in a stamp row, step 9174. For each row, the visible stamp set is sent to the Column Selection block 9054 of Stamp Selection Logic unit 9008. The Subrasterizer can process one row of stamps per cycle. If a primitive contains more than one row of stamps then the Subrasterizer takes more than one cycle to process the primitive and therefore will request MCCAM to stall the removal of primitives from the Input FIFO. The Subrasterizer itself can be stalled if a request is made by the Column Selection unit.
Figure 20 illustrates a stamp 9291 , containing four pixels 9292, 9293, 9294 and 9295. Each pixel is divided into 8 x 8 subraster grid. The grid shown in Figure 20 shows grid lines located at the mid-point of each subraster step. In one embodiment, samples are located at the center of a unit grid, as illustrated by samples 0 - 15 in Figure 20 designated by the circled numbers (e.g. ®). Placing the samples in this manner, off grid by one half of a subraster step, avoids the complications of visibility rules that apply to samples on the edge of a polygon. In this embodiment, polygons can be defined to go to the edge of a subraster line or pixel boundary, but samples are restricted to positions off of the subraster grid. In a further embodiment, two samples in adjacent pixels are placed on the same subraster. This simplifies sample processing by reducing the number of XleftSub, and XrightSub, by a factor of two.
7.0.5 Column Selection within Stamp Selection Logic Tthe Column Selection unit 9054, shown in Figure 14, tells the Z Cull unit 9012 which stamp to process in each clock cycle. If a stamp row contains more than one potentially visible stamp, the Column Selection unit requests that the Subrasterizer stall.
7.0.6 Z Cull
The Z Cull unit 9012 contains the Sample Z Buffer unit 9055 and Z Cull Sample State Machines 9057, shown in Figure 14. The Sample Z Buffer unit 9055 stores all the data for each sample in a tile, including the z value for each sample, and all the the sample FSM state bits. To enable the Z Cull Sample State Machines 9057 to process one stamp per cycle, Z Cull unit 9012 accesses the z values for all 16 sample points in a stamp in parallel and also computes the new primitive's z values at those sample points in parallel.
Z Cull unit 9012 determines whether a primitive covers a particular sample point i by comparing the sample point x coordinate, Xsample,, with the XlefSμb and XrightSub values computed by the
Subrasterizer. Sample i is covered if and only if XlefSub, ≤ Xsample, < XrightSub,, step 9178. Z Cull unit 9012 then computes the z value of the primitive at those sample points, step 9180, and compares the resulting z values to the corresponding z values stored in the Sample Z Buffer for that stamp, step 9182. Generally if the sample point z value is less than the z value in the Z Buffer then the sample point is considered to be visible. However, an API can allow programmers to specify the comparison function (>, ≥, <, ≤, always, never). Also, the z comparison can be affected by whether alpha test or blending is turned on, and whether the pipeline is in sorted transparency mode. The Z Cull Sample State Machines 9057 includes a per-sample FSM for each sample in a stamp.
In an embodiment where each stamp consists of 16 samples, there are 16 Z Cull Sample State Machines 9057 that each determine in parallel how to update the z value and sample state for the sample in the Z buffer it controls, and what action to take on the previously processed VSPs that overlap the sample point. Also in sorted transparency mode the Z Cull Sample State Machines determine whether to perform another pass through the transparent primitives. Based on the results of the comparison between the z value of the primitive at the sample points and the corresponding z values stored in the Sample Z Buffer for that stamp, the current Cull mode bits and the states of the sample state machines, the Sample Z Buffer is updated, step 9184. For each sample, the sixteen Z Cull Sample State Machines output the control bits: KeepOld, SendOld, NewVSPMask, and SendNew, to indicate how a sample is to be processed, step 9186. The set of NewVSPMask bits (16 of them) constitute a new stamp portion (SP) coverage mask, step 9188. The new stamp portion is dispatched to the New VSP Queue. In the event that the primitive is not visible at all in the stamp (all NewVSPMask bits are FALSE), then nothing is sent to the New VSP Queue. If more than one sample may affect the final sample position final value, then the stamp portions containing a sample for the sample position are early dispatched, step 9192. All of the control bits for the 16 samples in a stamp are provided to Stamp Portion Memory 9018 in parallel.
Samples are sent down the pipeline in VSPs, e.g. as part of a group comprising all of the currenlty visible samples in a stamp. When one sample within a stamp is dispatched (either early dispatch or end-of- tile dispatch), other samples within the same stamp and the same primitive are also dispatched as a VSP. While this causes more samples to be sent down the pipeline, it generally causes a net decrease in the amount of color computation. This is due to the spatial coherence within a pixel (i.e., samples within the same pixel tend to be either visible together or hidden together) and a tendency for the edges of polygons with alpha test, color test, stencil test, and/or alpha blending to potentially split otherwise spatially coherent stamps. That is, sending additional samples down the pipeline when they do not appreciably increase the computational load is more than offset by reducing the total number of VSPs that need to be sent. Figures 21A - 21D illustrate an example of the operation of an embodiment of Z Cull 9012. As illustrated in Figure 21A primitive 9312 is the first primitive in tile 9310. Z Cull 9012 therefore updates all the z values touched by the primitive and stores 35 stamp portions into Stamp Portion Memory 9018. In
Figure 21 B a second primitive 9322 is added to tile 9310. Primitive 9322 has lower z values than primitive 9312. Z-Cull 9012 processes the 27 stamps touched by primitive 9322. Figure 21C illustrates the 54 stamp portions stored in Stamp Portion Memory 9018 after primitive 9322 is processed. The 54 stamp portions are the sum of the stamps touched by primitives 9312 and 9322 minus eight stamp portions from primitive
9312 that are completely removed. Region 9332 in Figure 21 D indicates the eight stamp portions that are removed, which are the stamp portions wherein the entire component of the stamp portion touched by primitive 9312 is also touched by primitive 9322 which has lesser Z values. In one embodiment, Z Cull 9012 maintains one z value for each sample, as well as various state bits. In another embodiment, Z Cull 9012 maintains two z values for each sample, the second z value improves the efficiency of the conservative hidden surface removal process. Z Cull 9012 controls Stamp
Portion Memory 9018, but z values and state bits are not associated with stamp portions. Stamp Portion Memory 9018 can maintain 16 stamp portions per stamp, for a total of 256 stamp portions per tile.
Z Cull 9012 outputs the four bit control signal (SendNew, KeepOld and SendOld and NewVSPMask) to Stamp Portion Memory 9018 that controls how the sample is processed. KeepOld indicates that the corresponding sample in Stamp Portion Memory 9018 is not invalidated. That is, if the sample is part of a stamp portion in Stamp Portion Memory 9018, it is not discarded. SendOld is the early dispatch indicator.
If the sample corresponding to a SendOld bit belongs to a stamp portion in Stamp Portion Memory 9018, then this stamp portion is sent down the pipeline. SendOld is only asserted when KeepOld is asserted.
NewVSPMask is asserted, when the Z Cull 9012 process determines this sample is visible (at that point in the processing) and a new stamp portion needs to be created for the new primitive, which is done by Stamp
Portion Memory 9018 when it receives the signal. SendNew is asserted when the Z Cull 9012 process determines the sample is visible (at that point in the processing) and needs to be sent down the pipeline. SendNew causes an early dispatch of a stamp portion in the new primitive.
Figure 22 illustrates an example of how samples are processed by Z Cull 9012. Primitive 9352 is processed in tile 9350 before primitive 9354. Primitive 9354 has lesser z values than primitive 9352 and is therefore in front of primitive 9352. For the seven samples in oval region 9356 Z Cull 9012 sets the KeepOld control bits to zero, and the NewVSPMask control bits to one.
Figures 23A - 23D illustrate an example of early dispatch. Early dispatch is the sending of geometry down the pipeline before all geometry in the tile has been processed. In sorted transparency mode early dispatch is not used. First a single primitive 9372, illustrated in Figure 23A is processed in tile 9370. Primitive 9370 touches 35 stamps, and these are stored in Stamp Portion Memory 9018. A second primitive, 9382, with lesser z values is then added with the mode bit DoABIend asserted. The DoABIend mode bit indicates that the colors from the overlapping stamp portions should be blended. Z Cull 9012 then processes the 27 stamps touched by primitive 9382. Z Cull 9012 can be designed so that samples from up to N primitives can be stored for each stamp. In one embodiment samples from only one primitive are stored for each stamp. Figure 23C illustrates the stamp portions in Stamp Portion Memory 9018 after primitive 9382 is processed. Figure 23D illustrates the 20 visible stamp portions touched by region 9374 that are dispatched early from primitive 9372 because the stamp portion z values were replaced by the lesser z values from primitive 9382.
Figure 24 illustrates a sample level example of early dispatch processing. Stamp 9390 includes part of primitive 9382 and part of primitive 9372, both of which are shown in Figure 23B. The samples in region 9392 all are touched by primitive 9382 which has lesser z values than primitive 9372. Therefore, for these seven samples Z Cull 9012 outputs the control signal SendOld. In one embodiment, if Z Cull 9012 determines that one sample in a stamp should be sent down the pipeline then Z Cull 9012 sends all of the samples in that stamp down the pipeline so as to preserve spatial coherency. This is also minimizes the number of fragments that are sent down the pipeline. In another embodiment this approach is applied at a pixel level, wherein if Z Cull 9012 determines that any sample in a pixel should be sent down the pipeline all of the samples in the pixel are sent down the pipeline. In a cull process where everything in a scene is an opaque surface, after all the surfaces have been processed, only the stamp portions that are visible are left in Stamp Portion Memory 9018. The known visible stamp portions are then sent down the pipeline. However, when an early dispatch occurs, the early dispatch stamp portions are sent down the pipeline right away. For each stamp a reference called Zref is generated. In one embodiment, the Zref is placed at the center of the stamp. The values dzJdx. and θz/δy at the Zref point are also computed. These three values are sent down the pipeline to Pixel block 15000. Pixel block 15000 does a final z test. As part of the final z test, Pixel block 15000 re-computes the exactly equivalent z values for each sample using the Zref value and the Θz dx and dzldy values using the equation:
Figure imgf000043_0001
Computing the z values rather than sending the 16 z values in every stamp down the pipeline significantly reduces the bandwith used. Furthermore, only the z values of potentially visible samples are determined.
To ensure that Z Cull 9012 and Pixel block 15000 use exactly the same z values, Z Cull 9012 performs the same computations that Pixel block does to determine the z value for each stamp so as to avoid introducing any artifacts. To improve the computational efficiency a small number of bits can be used to express the delta x and delta y values, since the distances are only fractions of a pixel. For example, in one embodiment a 24 bit derivative and 4 bit delta values are used.
7.0.7 MCCAM Update MCCAM Update unit 9059, shown in Figure 14, determines the maximum of the sixteen updated z values for the sixteen sample points in each stamp and sends it to the MCCAM Cull unit to update the MCCAM array 9003.
7.0.8 New VSP Queue
Each clock cycle, Z Cull unit 9012 generates the four sets of four control bits (KeepOld, SendOld, NewVSPMask, and SendNew) per stamp portion. Thus Z Cull 9012 processes one stamp per primitive per cycle, but not all of the stamps processed are visible, only the Visible Stamp Portions (VSPs) are sent into New VSP Queue 9058. The input rate to New VSP Queue 9058 is therefore variable. Under "ideal" circumstances, the SPM Mask and Valid unit 9060 can store one new stamp portion every clock cycle. However, the SPM Mask and Valid unit 9060 requires multiple clocks for a new stamp portion when early dispatch of VSPs occurs. When VSPs are dispatched early, New VSP Queue 9058 stores the new stamp portions, thus allowing Z Cull 9012 to proceed without stalling. One new VSP may cause the dispatch of up to 16 old VSPs, so the removal rate from the New VSP Queue is also variable. In one embodiment, New VSP Queue 9058 is only used with early dispatches. The SPM Mask and
Valid unit handles one VSP at a time. The New VSP Queue ensures stamp portions are available for Z Cull 9012 when an early dispatch involves more than one VSP. Based upon performance analysis, typically about 450 stamps are expected to be touched in a tile. The depth complexity of a scene refers to the average number of times a pixel in the scene needs to be rendered. With a depth complexity of two, 225
VSPs would be expected to be provided as output from Z Cull 9012 per tile. Therefore on average about four VSPs are expected per stamp. A triangle with blend turned on covering a 50 pixel area can touch on average three tiles, and the number of stamps it touches within a tile should be less than eight. Therefore, in one embodiment, the New VSP Queue depth is set to be 32.
The link between Z Cull unit 9012 and Stamp Portion Memory 9018 through New VSP Queue 9058 is unidirectional. By avoiding using a feedback loop New VSP Queue 9058 is able to process samples in each cycle.
7.0.9 SPM Mask and Valid The active Stamp Portion Memory (SPM) Mask and Valid unit 9060 stores the VSP coverage masks for the tile. Each VSP entry includes a valid bit to indicate if there is a valid VSP stored there. The valid bits for the VSPs are stored in a separate memory. The Stamp Portion Memory Mask and Valid unit 9060 is doubie buffered (i.e. there are two copies 9060 and 9062) as shown in Figure 14. The Memory Mask and Valid Active State unit 9060 contains VSPs for the current tile while the Memory Mask and Valid Dispatch
State unit page 9062 contains VSPs from the previous tile (currently being dispatched). As a new VSP is removed from the New VSP Queue, the active state SPM Mask and Valid unit 9060 updates the VSP Mask for the VSPs that already exist in its mask memory and adds the new VSP to the memory content. When color blending or other conditions occur that require early dispatch, the active state SPM Mask and Valid unit dispatches VSPs through the active SPM Data unit 9064 to the dispatch queue. The operations performed in the mask update or early dispatch are controlled by the KeepOld, SendOld, SendNew and
NewVSPMask control bits generated in Z Cull 9012. In sorted transparency mode, the SendOld and SendNew mask bits are off. VSP coverage masks are mutually exclusive, therefore if a new VSP has a particular coverage mask bit turned on, the corresponding bit for all the previously processed VSPs in the stamp have to be turned off.
The state transition from active to dispatch and vice versa is controlled by mode packets. Receiving a packet signaling the end of a tile (Begin Tile, End Frame, Buffer Clear, or Cull Packet with CullFlushAII set to TRUE) causes the active state Stamp Portion Memory to switch over to dispatch state and vice versa. The page in dispatch state cycles through each stamp and sends all VSPs to the SPM Data unit, which forwards them to the dispatch queue. In an alternative embodiment, the Stamp Portion Memory Mask and Valid unit 9060 is triple buffered.
7.0.9.1 The SPM Data
The active Stamp Portion Memory Data unit 9064 stores the Zstamp, dz/dx, dz/dy and the Color Pointer for every VSP in the tile. The Stamp Portion Memory Data unit is also double buffered. The SPM
Mask and Valid unit 9060 sends new VSP information to the SPM Data unit 9064. The VSP information includes control signals that instruct the SPM Data unit 9064 to either send the new VSP or save the new
VSP to its memory. If the new VSP should be saved, the SPM Mask and Valid unit control signals also determine which location among the 16 possible slots the new VSP should occupy. In addition, for the case of early dispatch, the SPM Data unit also gets a list of old VSP locations and the associated VSP Masks that need early dispatch. The SPM Data unit first checks to see if there are any old VSPs that need to be dispatched, If the SPM Data unit finds any, it will read the VSP data from its memory, merge the VSP data with the VSP Mask sent from the SPM Mask and Valid unit, and put the old VSPs into the dispatch queue.
The SPM Data unit then checks if the new VSP should also be sent, and if it is affirmative, then it passes the new VSP data to the dispatch queue 9068. If the new VSP should not be sent, then the SPM Data unit writes the new VSP data into its memory.
7.0.10 The Dispatch Queue and Dispatch Logic The Dispatch Logic unit 9072 sends one entry's worth of data at a time from one of the two SPM dispatch queues 9068, 9070 to the Mode Injection unit 10000. The Dispatch Logic unit 9072 requests dispatch from the dispatch state SPM unit first. After the dispatch state SPM unit has exhausted all of its
VSPs, the Dispatch Logic unit 9072 requests dispatch from the active state SPM dispatch queue. 7.1 Alpha Test Alpha test compares the alpha value of a given pixel to an alpha reference value. The alpha reference value is often used to indicate the transparency value of a pixel. The type of comparison may be specified, so that for example the comparison may be a greater-than operation, a less-than operation, or other arithmetic, algebraic, or logical comparison, and so forth. If the comparison is a greater-than operation, then a pixel's alpha value has to be greater than the reference to pass the alpha test. For instance, if a pixel's alpha value is 0.9, the reference alpha is 0.8, and the comparison is greater-than, then that pixel passes the alpha test. Any pixel not passing the alpha test is discarded.
Alpha test is a per-fragment operation and in a preferred embodiment is performed by the Pixel block after all of the fragment coloring calculations, lighting operations and shading operations are completed. Figure 25 illustrates an example of processing samples with alpha test with a CHSR method. This diagram illustrates the rendering of six primitives (Primitives A, B, C, D, E, and F) at different z coordinate locations for a particular sample, rendered in the following order (starting with a "depth clear" and with "depth test" set to less-than): primitives A, B, and C (with "alpha test" disabled); primitive D (with "alpha test" enabled); and primitives E and F (with "alpha test" disabled). Note from the illustration that zA>Zc>zB>zE>zD>zF, such that primitive A is at the greatest z coordinate distance. Also note that alpha test is enabled for primitive D, but disabled for each of the other primitives. The steps for rendering these six primitives under a conservative hidden surface removal process with alpha test are as follows: Step 1 : The depth clear causes the following result in each sample finite state machine: 1 ) z values are initialized to the maximum value; 2) primitive information is cleared; and 3) sample state bits are set to indicate the z value is accurate.
Step 2: When primitive A is processed by the sample FSM, the primitive is kept (i.e., it becomes the current best guess for the visible surface), and this causes the sample FSM to store: 1 ) the z value zA as the "near" z value; 2) primitive information needed to color primitive A; and 3) the z value (zA) is labeled as accurate.
Step 3: When primitive B is processed by the sample FSM, the primitive is kept (its z value is less- than that of primitive A), and this causes the sample FSM to store: 1 ) the z value zB as the "near" z value (zA is discarded); 2) primitive information needed to color primitive B (primitive A's information is discarded); and 3) the z value (zB) is labeled as accurate.
Step 4: When primitive C is processed by the sample FSM the primitive is discarded (i.e., it is obscured by the current best guess for the visible surface, primitive B), and the sample FSM data is not changed. Step 5: When primitive D (which has alpha test enabled) is processed by the sample FSM, the primitive's visibility cannot be determined because it is closer than primitive B and because its alpha value is unknown at the time the sample FSM operates. Because a decision cannot be made as to which primitive would end up being visible (either primitive B or primitive D) primitive B is early dispatched down the pipeline (to have its colors generated) and primitive D is kept. When processing of primitive D has been completed, the sample FSM stores: 1 ) the "near" z value is zD and the "far" z value is z& 2) primitive information needed to color primitive D (primitive B's information has undergone early dispatch); and 3) the z values are labeled as conservative (because both a near and far are being maintained). In this condition, the sample FSM can determine that a piece of geometry closer than zD obscures previous geometry, geometry farther than zB is obscured, and geometry between zD and zB is indeterminate and must be assumed to be visible (hence a conservative assumption is made). When a sample FSM is in the conservative state and it contains valid primitive information, the sample FSM method considers the depth value of the stored primitive information to be the near depth value.
Step 6: When primitive E (which has alpha test disabled) is processed by the sample FSM, the primitive's visibility cannot be determined because it is between the near and far z values (i.e., between zD and zB). However, primitive E is not sent down the pipeline at this time because it could result in the primitives reaching the z buffered blend (part of the Pixel block in a preferred embodiment) out of correct time order. Therefore, primitive D is sent down the pipeline to preserve the time ordering. When processing of primitive E has been completed, the sample FSM stores: 1 ) the "near" z value is zD and the "far" z value is zB (note these have not changed, and zE is not kept); 2) primitive information needed to color primitive E (primitive D's information has undergone early dispatch); and 3) the z values are labeled as conservative
(because both a near and far are being maintained). Step 7: When primitive F is processed by the sample FSM, the primitive is kept (its z value is less- than that of the near z value), and this causes the sample FSM to store: 1) the z value zF as the "near" z value (zD and zB are discarded); 2) primitive information needed to color primitive F (primitive E's information is discarded); and 3) the z value (zF) is labeled as accurate. Step 8: When all the geometry that touches the tile has been processed (or, in the case there are no tiles, when all the geometry in the frame has been processed), any valid primitive information is sent down the pipeline. In this case, primitive F's information is sent. This is the end-of-tile (or end-of-frame) dispatch, and not an early dispatch. In summary in this CHSR process example involving alpha test, primitives A through F are processed, and primitives B, D, and F are sent down the pipeline. The Pixel block resolves the visibility of B, D, and F in the final z buffer blending stage. In this example, only the color primitive F is used for the sample.
7.1.0.1 Stencil Test In OpenGL® stencil test conditionally discards a fragment based on the outcome of a comparison between a value stored in a stencil buffer at location (xw, yw) and a reference value. Several stencil comparison functions are permitted such that whether the stencil test passes can depend upon whether the reference value is less than, less than or equal to, equal to, greater than or equal to, greater than, or not equal to the masked stored value in the stencil buffer. In OpenGL®, if the stencil test fails, the incoming fragment is discarded. The reference value and the comparison value can have multiple bits, typically 8 bits so that 256 different values may be represented. When an object is rendered into Frame Buffer 17000, a tag having the stencil bits is also written into the frame buffer. These stencil bits are part of the pipeline state. The type of stencil test to perform can be specified at the time the geometry is rendered.
The stencil bits are used to implement various filtering, masking or stenciling operations, to generate, for example, effects such as shadows. If a particular fragment ends up affecting a particular pixel in the frame buffer, then the stencil bits can be written to the frame buffer along with the pixel information.
In a preferred embodiment of the CHSR process, all stencil operations are done near the end of the pipeline in the Pixel block in a preferred embodiment. Therefore, the stencil values are stored in the
Frame Buffer and as a result the stencil values are not available to the CHSR method performed in the Cull block. While it is possible for the stencil values to be transferred from the Frame Buffer for use in the CHSR process, this would generally require a long latency path that would reduce performance. In APIs such as OpenGL®, the stencil test is performed after alpha test, and the results of alpha test are not known to the
CHSR process. Furthermore, renderers typically maintain stencil values over many frames (as opposed to depth values that are generally cleared at the start of each frame). Hence, the CHSR process utilizes a conservative approach to dealing with stencil operations. If a primitive can affect the stencil values in the frame buffer, then the VSPs in the primitive are always sent down the pipeline by the Cull block asserting the control bit CullFlushOverlap, shown in Figure 15. Primitves that can affect the stencil values are sent down the pipeline because stencil operations are performed by pipeline stages after Cull block 9000 (see
OpenGL® specification). A CullFlushOverlap condition sets the sample FSM to its most conservative state.
Generally the stencil test is defined for a group of primitives. When Cull block 9000 processes the first sample in a primtive with a new stencil test, control software sets the CullFlushAII bit in the corresponding Setup Output Cull Packet. CullFlushAII causes all of the VSPs from the Cull block to be sent to Pixel block 15000, and clears the z values in Stamp Portion Memory 9018. This "flushing" is needed because changing the stencil reference value effectively changes the "visibility rules" in the z buffered blend (or Pixel block). Pixel block 15000 compares the stencil values of the samples for a given sample location and determines which samples affect the final frame buffer color based on the stencil test. For example, for one group of samples corresponding to a sample location, the stencil test may be render if the stencil bit is equal to one. Pixel block 15000 then discards each of the samples for that sample in this group that have a stencil bit value not equal to one.
As an example of the CHSR process dealing with stencil test (see OpenGL® specification), consider the diagrammatic illustration of Figure 26, which has two primitives (primitives A and C) covering four particular samples (with corresponding sample FSMs, referred to as SFSM0 through SFSM3) and an additional primitive (primitive B) covering two of those four samples. The three primitives are rendered in the following order (starting with a depth clear and with depth test set to less-than): primitive A (with stencil test disabled); primitive B (with stencil test enabled and StencilOp set to "REPLACE", see OpenGL® specification); and primitive C (with stencil test disabled). The steps are as follows: Step 1 : The depth clear causes the following in each of the four sample FSMs in this example: 1 ) z values are initialized to the maximum value; 2) primitive information is cleared; and 3) sample state bits are set to indicate the z value is accurate.
Step 2: When primitive A is processed by each sample FSM, the primitive is kept (i.e., it becomes the current best guess for the visible surface), and this causes the four sample FSMs to store: 1 ) their corresponding z values (either zA0, zA1, z^, or zA3 respectively) as the "near" z value; 2) primitive information needed to color primitive A; and 3) the z values in each sample FSM are labeled as accurate. Step 3: When primitive B is processed by the sample FSMs, only samples 1 and 2 are affected, causing SFSMO and SFSM3 to be unaffected and causing SFSM1 and SFSM2 to be updated as follows:
1 ) the far z values are set to the maximum value and the near z values are set to the minimum value; 2) primitive information for primitives A and B are sent down the pipeline; and 3) sample state bits are set to indicate the z values are conservative. Step 4: When primitive C is processed by each sample FSM, the primitive is kept, but the sample
FSMs do not all handle the primitive the same way. In SFSMO and SFSM3, the state is updated as: 1 ) zco and zC3 become the "near" z values faQ and Z3 are discarded); 2) primitive information needed to color primitive C (primitive A's information is discarded); and 3) the z values are labeled as accurate. In SFSM1 and SFSM2, the state is updated as: 1) z^ and z^ become the "far" z values (the near z values are kept);
2) primitive information needed to color primitive C; and 3) the z values remain labeled as conservative. In summary in this CHSR process example involving stencil test, primitives A through C are processed, and all the primitives are sent down the pipeline, but not all the samples. In a preferred embodiment, the Pixel blocks performs final z buffered blending operations to process the unresolved visibility issues. Multiple samples were shown in this example to illustrate that CullFlushOverlap "flushes" selected samples while leaving others unaffected.
7.1.0.2 Alpha Blending
Alpha blending is used to combine the colors of two primitives into one color. However, the primitives are still subject to the depth test for the updating of the z values. The amount of color contribution from each of the samples depends upon the transparency values, referred to as the alpha value, of the samples. The blend is performed according to the equation
- C = Cs s + Cd(1- s)
where C is the resultant color, Cs is the source color for an incoming primitive sample, αs is the alpha value of the incoming primitive sample, and Cd is the destination color at the corresponding frame buffer location. Alpha values are defined at the vertices of primitives, and alpha values for samples are interpolated from the values at the vertices. As an example of the CHSR process dealing with alpha blending, consider Figure 27, which has four primitives (primitives A, B, C, and D) for a particular sample, rendered in the following order (starting with a depth clear and with depth test set to less-than): primitive A (with alpha blending disabled); primitives B and C (with alpha blending enabled); and primitive D (with alpha blending disabled). The steps are as follows:
Step 1 : The depth clear causes the following in each CHSR sample FSM: 1 ) z values are initialized to the maximum value; 2) primitive information is cleared; and 3) sample state bits are set to indicate the z value is accurate. Step 2: When primitive A is processed by the sample FSM, the primitive is kept (i.e., it becomes the current best guess for the visible surface), and this causes the sample FSM to store: 1 ) the z value zA as the "near" z value; 2) primitive information needed to color primitive A; and 3) the z value is labeled as accurate. Step 3: When primitive B is processed by the sample FSM, the primitive is kept (because its z value is less-than that of primitive A), and this causes the sample FSM to store: 1 ) the z value zB as the "near" z value (zA is discarded); 2) primitive information needed to color primitive B (primitive A's information is sent down the pipeline); and 3) the z value (zB) is labeled as accurate. Primitive A is sent down the pipeline because, at this point in the rendering process, the color of primitive B is to be blended with primitive A. This preserves the time order of the primitives as they are sent down the pipeline.
Step 4: When primitive C is processed by the sample FSM, the primitive is discarded (i.e., it is obscured by the current best guess for the visible surface, primitive B), and the sample FSM data is not changed. Note that if primitives B and C need to be rendered as transparent surfaces, then primitive C should not be hidden by primitive B. This could be accomplished by turning off the depth mask while primitive B is being rendered, but for transparency blending to be correct, the surfaces should be blended in either front-to-back or back-to-front order.
If the depth mask (see OpenGL® specification) is disabled, writing to the depth buffer (i.e., saving z values) is not performed; however, the depth test is still performed. In this example, if the depth mask is disabled for primitive B, then the value zB is not saved in the sample FSM. Subsequently, primitive C would then be considered visible because its z value would be compared to zA.
In summary of this example CHSR process example involving alpha blending, primitives A through D are processed, and all the primitives are sent down the pipeline, but not in all the samples. In a preferred embodiment, the Pixel blocks performs final z buffered blending operations to process the unresolved visibility issues. Multiple samples were shown in this example to illustrate that CullFlushOverlap dispatches selected samples without affecting other samples.
7.1.0.3 Control Bits Figure 28A illustrates part of a Spatial Packet containing three control bits: DoAlphaTest, DoABIend and Transparent. The Transparent bit is set by the Geometry block 3000 and is normally only used in sorted transparency mode. When the Transparent bit is reset the corresponding primitive is only processed in passes for opaque primitives. When the Transparent bit is set the corresponding primitive is only processed in passes for transparent primitives. The Transparent bit is generated in the Geometry block
3000 and is used by the Sort block 6000 to determine whether a particular primitive should be included in an opaque pass or a transparent pass; but, the Cull block 9000 knows the type of pass (i.e. opaque or transparent) by looking at the Begin Tile packet, so there is no need to send the Transparent bit to the Cull block 9000. The DoAlphaTest control bit controls whether Alpha test is performed on the samples in the primitive. When the DoAlphaTest control bit is set to a one it means that downstream from Cull block 9000 an alpha test will be performed on each fragment. When the alpha values of all of the samples in a stamp exceed a predetermined value, then even though an application program indicates that an alpha test should be performed, a functional block upstream from Cull block 9000 may determine that none of the samples can fail alpha test. DoAlphaTest can then be set to zero which indicates to Cull block 9000 that since all the samples are guaranteed to pass alpha test, it can process the samples as if they were not subject to alpha test. Observe that in an embodiment where one z value is stored, a sample being subject to alpha test can cause the stored sample to be made conservative. Therefore, DoAlphaTest being zero allows Cull to identify more samples as accurate and thereby eliminate more samples. A detailed description of the control of the DoAlphaTest control bit is provided in the provisional patent application entitled "Graphics Processor with Deferred Shading," filed August 20, 1998, which is incorporated by reference.
The DoABIend control bit, generated by the Geometry block 3000, indicates whether a primitive is subject to blending. Blending combines the color values of two samples.
In one embodiment, the Geometry block 3000 checks the alpha values at each vertex. If, given the alpha values, the BlendEquation and the BlendFunc pipeline state information is defined such that the frame buffer color values cannot affect the final color, then blending is turned off for that primitive using the DoABIend control bit. Observe that if blending was always on, and all primitives were treated as transparent, then a hidden surface removal process before lighting and shading might not not remove any geometry.
The following describes the method for evaluating texture data to determine whether blending can be turned off for a render if less than depth test. With a render if less than depth test, if there are two opaque primitives at the same location, the primitive that is in front is rendered. The present invention can also be used with a render if greater than depth test. Blending is turned off when a primitive is opaque and therefore no geometry behind the primitive will contribute to the corresponding final colors in the frame buffer. Whether a primitive is opaque is determined conservatively in that if there is any uncertainty as to whether the final frame buffer colors will be a blend of the current primitive and other primitives with greater z values, then the primitive is treated as transparent. For example, given an appropriately defined texture environment, if the alpha values at all of the vertices of a primitive are equal to one then blending can be turned off for that primitive because that primitive can be treated as opaque. Therefore, the culling method can be applied and more distant geometry can be eliminated. Whether blending can be turned off for a primitive depends upon the texture type, the texture data, and the texture environment. In one embodiment there are two texture types. The first texture type is RGB texture. In RGB texture each texel (the equivalent of a pixel in texture space) is defined by a red color component value "R," a green color component value "G," and a blue color component value "B." There are no alpha values in this first texture type. The second texture type describes each texel by R, G and B values as well as by an alpha value. The texture data comprise the values of the R, G, B and alpha components. The texture environment defines how to determine the final color of a pixel based on the relevant texture data and properties of the primitive. For example, the texture environment may define the type of interpolation that is used, as well as the lighting equation and when each operation is performed. Figure 28B illustrates how the alpha values are evaluated to set the DoABIend control bit. Alpha mode register stores the Transparent bits for each of the three vertices of a triangular primitive. The
Transparent bit defines whether the corresponding vertex is transparent indicated by a one, or opaque indicated by a zero. If all three of the vertices are opaque then blending is turned off, otherwise blending is on. Logic block implements this blending control function. When the AlphaAIIOne control signal is asserted and all three of the transparent bits in the alpha mode register are equal to one, logic block sets
DoABIend to a zero to turn off blending. The alpha value can also be inverted so that an alpha value of zero indicates that a vertex is opaque. Therefore, in this mode of operation, when the AlphaAIIZero control signal is asserted and all three of the transparent bits are zero, the logic block sets DoABIend to a zero ("0") to turn off blending.
7.1.0.4 Sorted Transparency Mode
The graphics pipeline operates in either time order mode or in sorted transparency mode. In sorted transparency mode, the process of reading geometry from a tile is divided into multiple passes. In the first pass, the Sort block outputs guaranteed opaque geometry, and in subsequent passes the Sort block outputs potentially transparent geometry. Within each sorted transparency mode pass, the time ordering is preserved, and mode data is inserted into its correct time-order location. Sorted transparency mode can be performed in either back-to-front or front-to-back order. In a preferred embodiment, the sorted transparency method is performed jointly by the Sort block and the Cull block. In back-to-front sorted transparency modes a pixel color is determined by first rendering the front most opaque surface at the sample location. In the next pass the farthest transparent surface, that is in front of the opaque surface is rendered. In the subsequent pass the next farthest transparent surface is rendered, and this process is repeated until all of the samples at the sample location have been rendered or when a predetermined maximum number of samples have been rendered for the sample location.
The following provides a more detailed description of the back-to-front sorted transparency mode rendering method. This method is used with a render if less than depth test. Referring to Figure 29, in the first pass the Sort block sends the opaque primitives. Cull block 9000 stores the z values for the opaque primitive samples in MCCAM array 9003 (shown in Figure 15) (step 2901 ). The Sort block sends transparent primitives to the Cull block in the second and subsequent passes. In sorted transparency mode MCCAM array 9003 and Sample Z Buffer 9055 each store two z values (Zfar and Znear) for each corresponding sample. The Zfar value is the z value of the closest opaque sample. The Znear value is the z value of the sample nearest to, and less than, the z value of the opaque layer. One embodiment includes two MCCAM arrays 9003 and two Sample Z Buffers 9055 so as to store the Zfar and Znear values in separate units. First the z values for the front-most non-transparent samples are stored in the MCCAM array 9003 (step 2902). The front-most non-transparent samples are then dispatched down the pipeline to be rendered (step 2903). In one embodiment, a flag bit in every pointer indicates whether the corresponding geometry is transparent or non-transparent. The Znear values for each sample are reset to zero (step 2904) in preparation for the next pass. During each transparent pass the z value for each sample point in the current primitive is compared with both the Zfar and the Znear values for that sample point. If the z value is larger than Znear but smaller than Zfar, then the sample is closer to the opaque layer and its z value replaces the current Znear value. The samples corresponding to the new Znear values are then dispatched down the pipeline to be rendered (step 2907), and Zfar for each such sample is set to the value of Znear (step 2908). This process is then repeated in the next pass. Cull block 9000 detects that it has finished processing a tile when for each sample point, there is at most one sample that is in front of Zfar. Transparent layer processing is not finished as long as there are two or more samples in front of Zfar for any sample point in the tile.
In front-to-back sorted transparency modes the transparent samples are rendered in order, starting at the front most transparent sample and then the next farther transparent sample in each subsequent cycle is rendered. An advantage of using a front-to-back sorted transparency mode is that if a maximum number of layers is defined, then the front most transparent layers are rendered which thereby provides a more accurate final displayed image. In one embodiment, the maximum number of layers to render is determined by accumulating the alpha values. The alpha value represents the transparency of the sample location. As each sample is rendered the transparency at that sample location decreases, and the cumulative alpha value increases
(where an alpha value of one is defined as opaque). For example, the maximum cumulative alpha value may be defined to be 0.9, when the cumulative alpha value exceeds 0.9 then no further samples at that sample location are rendered. There are two counters in Sample Z Buffer 9055, shown in Figure 15, for every sample. When two samples from different primitives at the same sample location have the same z value, the samples are rendered in the time order that they arrived. The counters are used to determine which sample should be rendered based on the time order. The first counter identifies the primitive that is to be processed in the current pass. For example, in a case where there are five primitives all having a sample in a given sample location with the same z value, in the first pass the first counter is set to one which indicates the first primitive in this group should be rendered. In the second pass this first counter is incremented, to identify the second primitive as the primitive to be rendered. The second counter maintains a count of the primitive being evaluated within a pass. In the five primitive example, in the third pass, the third primitive has the sample that should be rendered. At the start of the first pass the first counter is equal to three and the second counter is equal to one. The first counter value is compared with the second counter value and because the counter values are not equal the sample from the first primitive is not rendered. The second counter is then incremented, but the counters are still not equal so the sample from the second primitive is not rendered. In the third pass, the first and second counter values are equal, therefore the sample from the third primitive is rendered.
Characteristics of Particular Exemplary Embodiments
We now highlight particular embodiments of the inventive deferred shading graphics processor (DSGP). In one aspect (CULL) the inventive DSGP provides structure and method for performing conservative hidden surface removal. Numerous embodiments are shown and described, including but not limited to:
(1 ) A method of performing hidden surface removal in a computer graphics pipeline comprising the steps of: selecting a current primitive from a group of primitives, each primitive comprising a plurality of stamps; comparing stamps in the current primitive to stamps from previously evaluated primitives in the group of primitives; selecting a first stamp as a currently potentially visible stamp (CPVS) based on a relationship of depth states of samples in the first stamp with depth states of samples of previously evaluated stamps; comparing the CPVS to a second stamp; discarding the second stamp when no part of the second stamp would affect a final graphics display image based on the stamps that have been evaluated; discarding the CPVS and making the second stamp the CPVS, when the second stamp hides the CPVS; dispatching the CPVS and making the second stamp the CPVS when both the second stamp and the CPVS are at least partially visible in the final graphics display image; and dispatching the second stamp and the CPVS when the visibility of the second stamp and the CPVS depends on parameters evaluated later in the computer graphics pipeline.
(2)The method of (1) wherein the step of comparing the CPVS to a second stamp furthing comprises the steps of: comparing depth states of samples in the CPVS to depth states of samples in the second stamp; and evaluating pipeline state values. (3) The method of (1) wherein the depth state comprises one z value per sample, and wherein the z value includes a state bit which is defined to be accurate when the z value represents an actual z value of a currently visible surface and is defined to be conservative when the z value represents a maximum z value. (4) The method of (1 ) further comprising the step of dispatching the second stamp and the CPVS when the second stamp potentially alters the final graphics display image independent of the depth state. (5) The method of (1 ) further comprising the steps of: coloring the dispatched stamps; and performing an exact z buffer test on the dispatched stamps, after the coloring step. (6) The method of (1 ) further comprising the steps of: comparing alpha values of a plurality of samples to a reference alpha value; and performing the step of dispatching the second stamp and the CPVS, independent of alpha values when the alpha values of the plurality of samples are all greater than the reference value. (7) The method of (1 ) further comprising the steps of: determining whether any samples in the current primitive may affect final pixel color values in the final graphics display image; and turning blending off for the current primitive when no samples in the current primitive affect final pixel color values in the final graphics display image. (8) The method of claim 1 wherein the step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining a maximum z value for a plurality of stamp locations of the current primitive; comparing the maximum z value for a plurality of stamp positions with a minimum z value of the current primitive and setting corresponding stamp selection bits; and identifying as a process row a row of stamps wherein the maximum z value for a stamp position in the row is greater than the minimum z value of the current primitive. (9) The method of (8) wherein the step of determining a maximum z value for a plurality of stamp locations of the current primitive further comprises determining a maximum z value for each stamp in a bounding box of the current primitive. (10) The method of (8) wherein the step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining the left most and right most stamps touched by the current primitive in each of the process rows and defining corresponding stamp primitive coverage bits; and combining the stamp primitive coverage bits with the stamp selection bits to generate a final potentially visible stamp set. (11 ) The method of (10) wherein the step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining a set of sample points in a stamp in the final potentially visible stamp set; computing a z value for a plurality of sample points in the set of sample points; and comparing the computed z values with stored z values and outputting sample control signals. (12) The method of (10) wherein the step of comparing the computed z values with stored z values, further comprises the steps of: storing a first sample at a first sample location as a Zfar sample, if a first depth state of the first sample is the maximum depth state of a visible sample at the first sample location; comparing a second sample to the first sample; and storing the second sample if the second sample is currently potentially visible as a Zopt sample, and discarding the second sample when the Zfar sample hides the second sample. (13) The method of (10) wherein when it is determined that one sample in a stamp should be dispatched down the pipeline, all samples in the stamp are dispatched down the pipeline. (14) The method of (10) wherein when it is determined that one sample in a pixel should be dispatched down the pipeline, all samples in the pixel are dispatched down the pipeline. (15) The method of (10) wherein the step of computing a z value for a plurality of sample points in the set of sample points further comprises the steps of: creating a reference z value for a stamp; computing partial derivatives for a plurality of sample points in the set of sample points; sending down the pipeline the reference z value and the partial derivatives; and computing a z value for a sample based on the reference z value and partial derivatives. (16) The method of (10) further comprising the steps of: receiving a reference z value and partial derivatives; and re-computing a z value for a sample based on the reference z value and partial derivatives. (17) The method of (10) further comprising the step of dispatching the CPVS when the CPVS can affect stencil values. The method of (13) further comprising the step of dispatching all currently potentially visible stamps when a stencil test changes. (19) The method of (10) further comprising the steps of: storing concurrently samples from a plurality of primitives; and comparing a computed z value for a sample at a first sample location with stored z values of samples at the first sample location from a plurality of primitives. (20) The method of (10) wherein each stamp comprises at least one pixel and wherein the pixels in a stamp are processed in parallel. (21) The method of (20) further comprising the steps of: dividing a display image area into tiles; and rendering the display image in each tile independently. (22) The method of (10) wherein the sample points are located at positions between subraster grid lines. (23) The method of (20) wherein locations of the sample points within each pixel are programmable. (24) The method of (23) further comprising the steps of: programming a first set of sample locations in a plurality of pixels; evaluating stamp visibility using the first set of sample locations; programming a second set of sample locations in a plurality of pixels; and evaluating stamp visibility using the second set of sample locations. (25) The method of (10) further comprising the step of eliminating individual stamps that are determined not to affect the final graphics display image. (26) The method of (10) further comprising the step of turning off blending when alpha values at vertices of the current primitive have values such that frame buffer color values cannot affect a final color of samples in the current primitive. (27) The method of (1 ) wherein the depth state comprises a far z value and a near z value.
(28) A hidden surface removal system for a deferred shader computer graphics pipeline comprising: a magnitude comparison content addressable memory Cull unit for identifying a first group of potentially visible samples associated with a current primitive; a Stamp Selection unit, coupled to the magnitude comparison content addressable memory cull unit, for identifying, based on the first group and a perimeter of the primitive, a second group of potentially visible samples associated with the primitive; a Z Cull unit, coupled to the stamp selection unit and the magnitude comparison content addressable memory cull unit, for identifying visible stamp portions by evaluating a pipeline state, and comparing depth states of the second group with stored depth state values; and a Stamp Portion Memory unit, coupled to the Z Cull unit, for storing visible stamp portions based on control signals received from the Z Cull unit, wherein the Stamp Portion Memory unit dispatches stamps having a visibility dependent on parameters evaluated later in the computer graphics pipeline. (29) The hidden surface removal system of (28) wherein the stored depth state values are stored separately from the visible stamp portions. (30) The hidden surface removal system of
(28) wherein the Z Cull unit evaluates depth state and pipeline state values, and compares a currently potentially visible stamp (CPVS) to a first stamp; and wherein the Stamp Portion Memory, based on control signals from the Z Cull unit: discards the first stamp when no part of the first stamp would affect a final graphics display image based on the stamps that have been evaluated; discards the CPVS and makes the first stamp the CPVS, when the first stamp hides CPVS; dispatches the CPVS and makes the first stamp the CPVS when both the first stamp and the CPVS are at least partially visible in the final graphics display image; and dispatches the first stamp and the CPVS when the visibility of the first stamp and the CPVS depends on parameters evaluated later in the computer graphics pipeline. (31 ) The hidden surface removal system of (28) wherein the MCCAM Cull unit: determines a maximum z value for a plurality of stamp locations of the current primitive; compares the maximum z value for a plurality of stamp positions with a minimum z value of the current primitive and sets corresponding stamp selection bits; and identifies as a process row a row of stamps wherein the maximum z value for a stamp position in the row is greater than the minimum z value of the current primitive. (32) The hidden surface removal system of (31 ) wherein the Stamp Selection unit: determines the leftmost and right most stamps touched by the current primitive in each of the process rows and defines corresponding stamp primitive coverage bits; and combines the stamp primitive coverage bits with the stamp selection bits to generate a final potentially visible stamp set. (33)
The hidden surface removal system of (32) wherein the Z Cull unit: determines a set of sample points in a stamp in the final potentially visible stamp set; computes a z value for a plurality of sample points in the set of sample points; and compares the computed z values with stored z values and outputs control signals. (34) The hidden surface removal system of (33) wherein the Z Cull unit comprises a plurality of Z Cull
Sample State Machines, each of the Z Cull Sample State Machines receive, process and output control signals for samples in parallel.
(35) A method of rendering a computer graphics image comprising the steps of: receiving a plurality of primitives to be rendered; selecting a sample location; rendering a front most opaque sample at the selected sample location, and defining the z value of the front most opaque sample as Zfar; comparing z values of a first plurality of samples at the selected sample location; defining to be Znear a first sample, at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the first plurality of samples; rendering the first sample; setting Zfar to the value of Znear; comparing z values of a second plurality of samples at the selected sample location; defining as Znear the z value of a second sample at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the second plurality of samples; and rendering the second sample. (36) The method of 35 further comprising the steps of: when a third plurality of samples at the selected sample location have a common z value which is less than Zfar, and the common z value is the z value nearest to Zfar of the first plurality of samples: rendering a third sample, wherein the third sample is the first sample received of the third plurality of samples; incrementing a first counter value to define a sample render number, wherein the sample render number identifies the sample to be rendered; selecting a fourth sample from the third plurality of samples; incrementing a second counter wherein the second counter defines an evaluation sample number; comparing the sample render number and the evaluation sample number; and rendering a sample when the corresponding evaluation sample number equals the sample render number. Alternative Embodiments The foregoing descriptions of specific embodiments of the present invention have been presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed, and obviously many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, to thereby enable others skilled in the art to best use the invention and various embodiments with various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the claims appended hereto and their equivalents.
All publications, patents, and patent applications mentioned in this specification are herein incorporated by reference to the same extent as if each individual publication or patent application was specifically and individually indicated to be incorporated by reference. U.S. Provisional Patent App. Serial No. 60/097,336 entitled Graphics Processor with Deferred Shading filed August 20, 1998 is hereby specifically incorporated by reference.

Claims

WHAT IS CLAIMED IS;
1. A method of performing hidden surface removal in a computer graphics pipeline comprising the steps of: selecting a current primitive from a group of primitives, each primitive comprising a plurality of stamps; comparing stamps in the current primitive to stamps from previously evaluated primitives in the group of primitives; selecting a first stamp as a currently potentially visible stamp (CPVS) based on a relationship of depth states of samples in the first stamp with depth states of samples of previously evaluated stamps; comparing the CPVS to a second stamp; discarding the second stamp when no part of the second stamp would affect a final graphics display image based on the stamps that have been evaluated; discarding the CPVS and making the second stamp the CPVS, when the second stamp hides the
CPVS; dispatching the CPVS and making the second stamp the CPVS when both the second stamp and the CPVS are at least partially visible in the final graphics display image; and dispatching the second stamp and the CPVS when the visibility of the second stamp and the CPVS depends on parameters evaluated later in the computer graphics pipeline.
2. The method of claim 1 wherein the step of comparing the CPVS to a second stamp furthing comprises the steps of: comparing depth states of samples in the CPVS to depth states of samples in the second stamp; and evaluating pipeline state values.
3. The method of claim 1 wherein the depth state comprises one z value per sample, and wherein the z value includes a state bit which is defined to be accurate when the z value represents an actual z value of a currently visible surface and is defined to be conservative when the z value represents a maximum z value.
4. The method of claim 1 further comprising the step of dispatching the second stamp and the CPVS when the second stamp potentially alters the final graphics display image independent of the depth state.
5. The method of claim 1 further comprising the steps of: coloring the dispatched stamps; and performing an exact z buffer test on the dispatched stamps, after said coloring step.
6. The method of claim 1 further comprising the steps of: comparing alpha values of a plurality of samples to a reference alpha value; and performing said step of dispatching the second stamp and the CPVS, independent of alpha values when said alpha values of said plurality of samples are all greater than said reference value.
7. The method of claim 1 further comprising the steps of: determining whether any samples in said current primitive may affect final pixel color values in said final graphics display image; and turning blending off for said current primitive when no samples in said current primitive affect final pixel color values in said final graphics display image.
8. The method of claim 1 wherein the step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining a maximum z value for a plurality of stamp locations of the current primitive; comparing the maximum z value for a plurality of stamp positions with a minimum z value of the current primitive and setting corresponding stamp selection bits; and identifying as a process row a row of stamps wherein the maximum z value for a stamp position in the row is greater than the minimum z value of the current primitive.
9. The method of claim 8 wherein the step of determining a maximum z value for a plurality of stamp locations of the current primitive further comprises determining a maximum z value for each stamp in a bounding box of the current primitive.
10. The method of claim 8 wherein the step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining the left most and right most stamps touched by the current primitive in each of the process rows and defining corresponding stamp primitive coverage bits; and combining the stamp primitive coverage bits with the stamp selection bits to generate a final potentially visible stamp set.
11. The method of claim 10 wherein the step of comparing stamps in the current primitive to stamps from previously evaluated primitives further comprises the steps of: determining a set of sample points in a stamp in the final potentially visible stamp set; computing a z value for a plurality of sample points in the set of sample points; and comparing the computed z values with stored z values and outputting sample control signals.
12. The method of claim 10 wherein the step of comparing the computed z values with stored z values, further comprises the steps of: storing a first sample at a first sample location as a Zfar sample, if a first depth state of said first sample is the maximum depth state of a visible sample at said first sample location; comparing a second sample to said first sample; and storing said second sample if said second sample is currently potentially visible as a Zopt sample, and discarding said second sample when said Zfar sample hides said second sample.
13. The method of claim 10 wherein when it is determined that one sample in a stamp should be dispatched down the pipeline, all samples in the stamp are dispatched down the pipeline.
14. The method of claim 10 wherein when it is determined that one sample in a pixel should be dispatched down the pipeline, all samples in the pixel are dispatched down the pipeline.
15. The method of claim 10 wherein the step of computing a z value for a plurality of sample points in the set of sample points further comprises the steps of: creating a reference z value for a stamp; computing partial derivatives for a plurality of sample points in the set of sample points; sending down the pipeline the reference z value and the partial derivatives; and computing a z value for a sample based on the reference z value and partial derivatives.
16. The method of claim 10 further comprising the steps of: receiving a reference z value and partial derivatives; and re-computing a z value for a sample based on the reference z value and partial derivatives.
17. The method of claim 10 further comprising the step of dispatching the CPVS when the CPVS can affect stencil values.
18. The method of claim 13 further comprising the step of dispatching all currently potentially visible stamps when a stencil test changes.
19. The method of claim 10 further comprising the steps of: storing concurrently samples from a plurality of primitives; and comparing a computed z value for a sample at a first sample location with stored z values of samples at the first sample location from a plurality of primitives.
20. The method of claim 10 wherein each stamp comprises at least one pixel and wherein the pixels in a stamp are processed in parallel.
21. The method of claim 20 further comprising the steps of: dividing a display image area into tiles; and rendering the display image in each tile independently.
22. The method of claim 10 wherein the sample points are located at positions between subraster grid lines.
23. The method of claim 20 wherein locations of the sample points within each pixel are programmable.
24. The method of claim 23 further comprising the steps of: programming a first set of sample locations in a plurality of pixels; evaluating stamp visibility using said first set of sample locations; programming a second set of sample locations in a plurality of pixels; and evaluating stamp visibility using said second set of sample locations.
25. The method of claim 10 further comprising the step of eliminating individual stamps that are determined not to affect the final graphics display image.
26. The method of claim 10 further comprising the step of turning off blending when alpha values at vertices of the current primitive have values such that frame buffer color values cannot affect a final color of samples in the current primitive.
27. The method of claim 1 wherein the depth state comprises a far z value and a near z value.
28. A hidden surface removal system for a deferred shader computer graphics pipeline comprising: a Magnitude Comparison Content Addressable Memory (MCCAM) Cull unit for identifying a first group of potentially visible samples associated with a current primitive; a Stamp Selection unit, coupled to the MCCAM cull unit, for identifying, based on the first group and a perimeter of the primitive, a second group of potentially visible samples associated with the primitive; a Z Cull unit, coupled to the stamp selection unit and the MCCAM cull unit, for identifying visible stamp portions by evaluating a pipeline state, and comparing depth states of the second group with stored depth state values; and a Stamp Portion Memory unit, coupled to the Z Cull unit, for storing visible stamp portions based on control signals received from the Z Cull unit, wherein the Stamp Portion Memory unit dispatches stamps having a visibility dependent on parameters evaluated later in the computer graphics pipeline.
29. The hidden surface removal system of claim 28 wherein the stored depth state values are stored separately from the visible stamp portions.
30. The hidden surface removal system of claim 28 wherein the Z Cull unit evaluates depth state and pipeline state values, and compares a currently potentially visible stamp (CPVS) to a first stamp; and wherein said Stamp Portion Memory, based on control signals from the Z Cull unit: discards the first stamp when no part of the first stamp would affect a final graphics display image based on the stamps that have been evaluated; discards the CPVS and makes the first stamp the CPVS, when the first stamp hides CPVS; dispatches the CPVS and makes the first stamp the CPVS when both the first stamp and the CPVS are at least partially visible in the final graphics display image; and dispatches the first stamp and the CPVS when the visibility of the first stamp and the CPVS depends on parameters evaluated later in the computer graphics pipeline.
31. The hidden surface removal system of claim 28 wherein the MCCAM Cull unit: determines a maximum z value for a plurality of stamp locations of the current primitive; compares the maximum z value for a plurality of stamp positions with a minimum z value of the current primitive and sets corresponding stamp selection bits; and identifies as a process row a row of stamps wherein the maximum z value for a stamp position in the row is greater than the minimum z value of the current primitive.
32. The hidden surface removal system of 31 wherein the Stamp Selection unit: determines the leftmost and right most stamps touched by the current primitive in each of the process rows and defines corresponding stamp primitive coverage bits; and combines the stamp primitive coverage bits with the stamp selection bits to generate a final potentially visible stamp set.
33. The hidden surface removal system of claim 32 wherein the Z Cull unit: determines a set of sample points in a stamp in the final potentially visible stamp set; computes a z value for a plurality of sample points in the set of sample points; and compares the computed z values with stored z values and outputs control signals.
34. The hidden surface removal system of claim 33 wherein the Z Cull unit comprises a plurality of Z
Cull Sample State Machines, each of said Z Cull Sample State Machines receive, process and output control signals for samples in parallel.
35. A method of rendering a computer graphics image comprising the steps of: receiving a plurality of primitives to be rendered; selecting a sample location; rendering a front most opaque sample at the selected sample location, and defining the z value of the front most opaque sample as Zfar; comparing z values of a first plurality of samples at the selected sample location; defining to be Znear a first sample, at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the first plurality of samples; rendering the first sample; setting Zfar to the value of Znear; comparing z values of a second plurality of samples at the selected sample location; defining as Znear the z value of a second sample at the selected sample location, having a z value which is less than Zfar and which is nearest to Zfar of the second plurality of samples; and rendering the second sample.
36. The method of claim 35 further comprising the steps of: when a third plurality of samples at the selected sample location have a common z value which is less than Zfar, and the common z value is the z value nearest to Zfar of the first plurality of samples: rendering a third sample, wherein the third sample is the first sample received of the third plurality of samples; incrementing a first counter value to define a sample render number, wherein the sample render number identifies the sample to be rendered; selecting a fourth sample from said third plurality of samples; incrementing a second counter wherein said second counter defines an evaluation sample number; comparing said sample render number and said evaluation sample number; and rendering a sample when the corresponding evaluation sample number equals the sample render number.
PCT/US1999/019190 1998-08-20 1999-08-20 Performing hidden surface removal in a graphics processor with deferred shading WO2000011613A2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU57825/99A AU5782599A (en) 1998-08-20 1999-08-20 Method and apparatus for performing conservative hidden surface removal in a graphics processor with deferred shading

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US9733698P 1998-08-20 1998-08-20
US60/097,336 1998-08-20
US09/213,990 US6771264B1 (en) 1998-08-20 1998-12-17 Method and apparatus for performing tangent space lighting and bump mapping in a deferred shading graphics processor
US09/213,990 1998-12-17

Publications (3)

Publication Number Publication Date
WO2000011613A2 true WO2000011613A2 (en) 2000-03-02
WO2000011613A3 WO2000011613A3 (en) 2000-06-29
WO2000011613A9 WO2000011613A9 (en) 2000-08-03

Family

ID=26793137

Family Applications (6)

Application Number Title Priority Date Filing Date
PCT/US1999/019241 WO2000011604A2 (en) 1998-08-20 1999-08-20 Apparatus and method for geometry operations in a 3d-graphics pipeline
PCT/US1999/019190 WO2000011613A2 (en) 1998-08-20 1999-08-20 Performing hidden surface removal in a graphics processor with deferred shading
PCT/US1999/019363 WO2000011605A2 (en) 1998-08-20 1999-08-20 Fragment operations in a 3d-graphics pipeline
PCT/US1999/019036 WO2000011614A2 (en) 1998-08-20 1999-08-20 Tangent space lighting in a deferred shading architecture
PCT/US1999/018971 WO2000030040A1 (en) 1998-08-20 1999-08-20 Advanced deferred shading graphics pipeline processor
PCT/US1999/019254 WO2000019377A1 (en) 1998-08-20 1999-08-20 Graphics processor with deferred shading

Family Applications Before (1)

Application Number Title Priority Date Filing Date
PCT/US1999/019241 WO2000011604A2 (en) 1998-08-20 1999-08-20 Apparatus and method for geometry operations in a 3d-graphics pipeline

Family Applications After (4)

Application Number Title Priority Date Filing Date
PCT/US1999/019363 WO2000011605A2 (en) 1998-08-20 1999-08-20 Fragment operations in a 3d-graphics pipeline
PCT/US1999/019036 WO2000011614A2 (en) 1998-08-20 1999-08-20 Tangent space lighting in a deferred shading architecture
PCT/US1999/018971 WO2000030040A1 (en) 1998-08-20 1999-08-20 Advanced deferred shading graphics pipeline processor
PCT/US1999/019254 WO2000019377A1 (en) 1998-08-20 1999-08-20 Graphics processor with deferred shading

Country Status (6)

Country Link
US (5) US6771264B1 (en)
EP (2) EP1105844A1 (en)
JP (3) JP3657519B2 (en)
KR (2) KR100485241B1 (en)
AU (6) AU5687599A (en)
WO (6) WO2000011604A2 (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6532013B1 (en) 2000-05-31 2003-03-11 Nvidia Corporation System, method and article of manufacture for pixel shaders for programmable shading
US6664963B1 (en) 2000-05-31 2003-12-16 Nvidia Corporation System, method and computer program product for programmable shading using pixel shaders
US6690372B2 (en) 2000-05-31 2004-02-10 Nvidia Corporation System, method and article of manufacture for shadow mapping
US6697064B1 (en) 2001-06-08 2004-02-24 Nvidia Corporation System, method and computer program product for matrix tracking during vertex processing in a graphics pipeline
US6704025B1 (en) 2001-08-31 2004-03-09 Nvidia Corporation System and method for dual-depth shadow-mapping
US6734861B1 (en) 2000-05-31 2004-05-11 Nvidia Corporation System, method and article of manufacture for an interlock module in a computer graphics processing pipeline
US6778181B1 (en) 2000-12-07 2004-08-17 Nvidia Corporation Graphics processing system having a virtual texturing array
EP1450308A1 (en) * 2001-11-27 2004-08-25 Sony Computer Entertainment Inc. Image processing apparatus and constituent parts thereof, rendering method
US6844880B1 (en) 1999-12-06 2005-01-18 Nvidia Corporation System, method and computer program product for an improved programmable vertex processing model with instruction set
US6870540B1 (en) * 1999-12-06 2005-03-22 Nvidia Corporation System, method and computer program product for a programmable pixel processing model with instruction set
US7006101B1 (en) 2001-06-08 2006-02-28 Nvidia Corporation Graphics API with branching capabilities
US7009605B2 (en) 2002-03-20 2006-03-07 Nvidia Corporation System, method and computer program product for generating a shader program
US7009615B1 (en) 2001-11-30 2006-03-07 Nvidia Corporation Floating point buffer system and method for use during programmable fragment processing in a graphics pipeline
US7023437B1 (en) 1998-07-22 2006-04-04 Nvidia Corporation System and method for accelerating graphics processing using a post-geometry data stream during multiple-pass rendering
US7162716B2 (en) 2001-06-08 2007-01-09 Nvidia Corporation Software emulator for optimizing application-programmable vertex processing
US7170513B1 (en) 1998-07-22 2007-01-30 Nvidia Corporation System and method for display list occlusion branching
US7209140B1 (en) 1999-12-06 2007-04-24 Nvidia Corporation System, method and article of manufacture for a programmable vertex processing model with instruction set
US7286133B2 (en) 2001-06-08 2007-10-23 Nvidia Corporation System, method and computer program product for programmable fragment processing
US7456838B1 (en) 2001-06-08 2008-11-25 Nvidia Corporation System and method for converting a vertex program to a binary format capable of being executed by a hardware graphics pipeline

Families Citing this family (634)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8253729B1 (en) * 1983-05-09 2012-08-28 Geshwind David M Trimming depth buffer during 2D to 3D conversion
US6590996B1 (en) * 2000-02-14 2003-07-08 Digimarc Corporation Color adaptive watermarking
US7375727B1 (en) * 1998-07-22 2008-05-20 Nvidia Corporation System, method and computer program product for geometrically transforming geometric objects
US6480205B1 (en) 1998-07-22 2002-11-12 Nvidia Corporation Method and apparatus for occlusion culling in graphics systems
US6476807B1 (en) * 1998-08-20 2002-11-05 Apple Computer, Inc. Method and apparatus for performing conservative hidden surface removal in a graphics processor with deferred shading
US6771264B1 (en) * 1998-08-20 2004-08-03 Apple Computer, Inc. Method and apparatus for performing tangent space lighting and bump mapping in a deferred shading graphics processor
US6978045B1 (en) * 1998-10-02 2005-12-20 Minolta Co., Ltd. Image-processing apparatus
GB2343601B (en) * 1998-11-06 2002-11-27 Videologic Ltd Shading and texturing 3-dimensional computer generated images
US6509905B2 (en) * 1998-11-12 2003-01-21 Hewlett-Packard Company Method and apparatus for performing a perspective projection in a graphics device of a computer graphics display system
JP3258286B2 (en) * 1998-12-15 2002-02-18 インターナショナル・ビジネス・マシーンズ・コーポレーション Drawing method and drawing apparatus for displaying image data of a plurality of objects in which translucent and opaque objects are mixed on a computer display screen
US7224364B1 (en) * 1999-02-03 2007-05-29 Ati International Srl Optimal initial rasterization starting point
US6466223B1 (en) * 1999-03-24 2002-10-15 Microsoft Corporation Method and apparatus for texture memory management
US6791569B1 (en) * 1999-07-01 2004-09-14 Microsoft Corporation Antialiasing method using barycentric coordinates applied to lines
US6628836B1 (en) * 1999-10-05 2003-09-30 Hewlett-Packard Development Company, L.P. Sort middle, screen space, graphics geometry compression through redundancy elimination
JP3950926B2 (en) * 1999-11-30 2007-08-01 エーユー オプトロニクス コーポレイション Image display method, host device, image display device, and display interface
US6848029B2 (en) 2000-01-03 2005-01-25 Dirk Coldewey Method and apparatus for prefetching recursive data structures
US7058636B2 (en) * 2000-01-03 2006-06-06 Dirk Coldewey Method for prefetching recursive data structure traversals
US6731297B1 (en) * 2000-01-11 2004-05-04 Intel Corporation Multiple texture compositing
US7483042B1 (en) * 2000-01-13 2009-01-27 Ati International, Srl Video graphics module capable of blending multiple image layers
US6995761B1 (en) * 2000-01-14 2006-02-07 California Institute Of Technology Compression of 3D surfaces using progressive geometry
GB2363045B (en) * 2000-01-28 2004-06-02 Namco Ltd Game system and image creating method
US20020009293A1 (en) * 2000-02-03 2002-01-24 Aldrich Kipp A. HDTV video server
JP3349490B2 (en) * 2000-02-14 2002-11-25 インターナショナル・ビジネス・マシーンズ・コーポレーション Image display method, image display system, host device, image display device, and display interface
US7159041B2 (en) * 2000-03-07 2007-01-02 Microsoft Corporation Method and system for defining and controlling algorithmic elements in a graphics display system
US7098925B1 (en) * 2000-03-10 2006-08-29 Intel Corporation Shading of images using texture
US6819321B1 (en) * 2000-03-31 2004-11-16 Intel Corporation Method and apparatus for processing 2D operations in a tiled graphics architecture
US7038811B1 (en) * 2000-03-31 2006-05-02 Canon Kabushiki Kaisha Standardized device characterization
JP2003529860A (en) * 2000-03-31 2003-10-07 インテル・コーポレーション Tile graphics architecture
US7119813B1 (en) * 2000-06-02 2006-10-10 Nintendo Co., Ltd. Variable bit field encoding
US7032031B2 (en) * 2000-06-23 2006-04-18 Cloudshield Technologies, Inc. Edge adapter apparatus and method
US7405734B2 (en) * 2000-07-18 2008-07-29 Silicon Graphics, Inc. Method and system for presenting three-dimensional computer graphics images using multiple graphics processing units
US6963347B1 (en) * 2000-08-04 2005-11-08 Ati International, Srl Vertex data processing with multiple threads of execution
US6999100B1 (en) 2000-08-23 2006-02-14 Nintendo Co., Ltd. Method and apparatus for anti-aliasing in a graphics system
US6980218B1 (en) * 2000-08-23 2005-12-27 Nintendo Co., Ltd. Method and apparatus for efficient generation of texture coordinate displacements for implementing emboss-style bump mapping in a graphics rendering system
US6825851B1 (en) 2000-08-23 2004-11-30 Nintendo Co., Ltd. Method and apparatus for environment-mapped bump-mapping in a graphics system
US7061502B1 (en) * 2000-08-23 2006-06-13 Nintendo Co., Ltd. Method and apparatus for providing logical combination of N alpha operations within a graphics system
US7002591B1 (en) * 2000-08-23 2006-02-21 Nintendo Co., Ltd. Method and apparatus for interleaved processing of direct and indirect texture coordinates in a graphics system
US8692844B1 (en) * 2000-09-28 2014-04-08 Nvidia Corporation Method and system for efficient antialiased rendering
US6828980B1 (en) * 2000-10-02 2004-12-07 Nvidia Corporation System, method and computer program product for z-texture mapping
US6914618B2 (en) * 2000-11-02 2005-07-05 Sun Microsystems, Inc. Methods and systems for producing A 3-D rotational image from A 2-D image
US7079133B2 (en) * 2000-11-16 2006-07-18 S3 Graphics Co., Ltd. Superscalar 3D graphics engine
JP3705739B2 (en) * 2000-12-11 2005-10-12 株式会社ナムコ Information storage medium and game device
US6975320B1 (en) 2000-12-12 2005-12-13 Micron Technology, Inc. Method and apparatus for level-of-detail computations
US6664961B2 (en) * 2000-12-20 2003-12-16 Rutgers, The State University Of Nj Resample and composite engine for real-time volume rendering
US20030063095A1 (en) * 2000-12-29 2003-04-03 Sun Microsystems, Inc. Statistic logic for collecting a histogram of pixel exponent values
JP2002252770A (en) * 2001-02-22 2002-09-06 Matsushita Graphic Communication Systems Inc Classification method for image information, image coding method, and image coder
US6791559B2 (en) * 2001-02-28 2004-09-14 3Dlabs Inc., Ltd Parameter circular buffers
US6828975B2 (en) * 2001-03-01 2004-12-07 Microsoft Corporation Method and system for managing graphics objects in a graphics display system
FR2822274B1 (en) * 2001-03-13 2003-11-21 Stephane Clement Francoi Rehel METHOD FOR DISPLAYING AND HANDLING AN OBJECT IN THREE DIMENSIONS AND CORRESPONDING APPLICATIONS
EP1258837A1 (en) * 2001-05-14 2002-11-20 Thomson Licensing S.A. Method to generate mutual photometric effects
US6859209B2 (en) * 2001-05-18 2005-02-22 Sun Microsystems, Inc. Graphics data accumulation for improved multi-layer texture performance
GB2378108B (en) 2001-07-24 2005-08-17 Imagination Tech Ltd Three dimensional graphics system
US6778189B1 (en) * 2001-08-24 2004-08-17 Nvidia Corporation Two-sided stencil testing system and method
US6734853B2 (en) * 2001-08-28 2004-05-11 Intel Corporation Method of using view frustrum culling for scaleable collision detection
US7145577B2 (en) * 2001-08-31 2006-12-05 Micron Technology, Inc. System and method for multi-sampling primitives to reduce aliasing
US6924820B2 (en) * 2001-09-25 2005-08-02 Sun Microsystems, Inc. Over-evaluating samples during rasterization for improved datapath utilization
WO2003032253A2 (en) * 2001-10-10 2003-04-17 Sony Computer Entertainment America Inc. System and method for environment mapping
US6999076B2 (en) * 2001-10-29 2006-02-14 Ati Technologies, Inc. System, method, and apparatus for early culling
KR100450836B1 (en) * 2001-12-11 2004-10-01 삼성전자주식회사 Apparatus for generating 3-dimensional image from 2-dimensional image
US7426534B2 (en) * 2001-12-19 2008-09-16 International Business Machines Corporation Method and system for caching message fragments using an expansion attribute in a fragment link tag
US6816161B2 (en) * 2002-01-30 2004-11-09 Sun Microsystems, Inc. Vertex assembly buffer and primitive launch buffer
CN1625757A (en) * 2002-02-01 2005-06-08 皇家飞利浦电子股份有限公司 Stepless 3d texture mapping in computer graphics
US6774895B1 (en) 2002-02-01 2004-08-10 Nvidia Corporation System and method for depth clamping in a hardware graphics pipeline
US7310103B2 (en) * 2002-03-05 2007-12-18 Sun Microsystems, Inc. Pipelined 2D viewport clip circuit
US7535913B2 (en) * 2002-03-06 2009-05-19 Nvidia Corporation Gigabit ethernet adapter supporting the iSCSI and IPSEC protocols
US7159212B2 (en) * 2002-03-08 2007-01-02 Electronic Arts Inc. Systems and methods for implementing shader-driven compilation of rendering assets
US6975322B2 (en) * 2002-03-12 2005-12-13 Sun Microsystems, Inc. Dynamically adjusting a number of rendering passes in a graphics system
US7015909B1 (en) * 2002-03-19 2006-03-21 Aechelon Technology, Inc. Efficient use of user-defined shaders to implement graphics operations
US8284844B2 (en) 2002-04-01 2012-10-09 Broadcom Corporation Video decoding system supporting multiple standards
US7376743B1 (en) * 2002-04-02 2008-05-20 Cisco Technology, Inc. Method and apparatus for load balancing in a virtual private network
US7009608B2 (en) * 2002-06-06 2006-03-07 Nvidia Corporation System and method of using multiple representations per object in computer graphics
US6771271B2 (en) * 2002-06-13 2004-08-03 Analog Devices, Inc. Apparatus and method of processing image data
AUPS300502A0 (en) * 2002-06-17 2002-07-11 Canon Kabushiki Kaisha Generating one or more linear blends
US6812927B1 (en) * 2002-06-18 2004-11-02 Nvidia Corporation System and method for avoiding depth clears using a stencil buffer
KR20030097507A (en) * 2002-06-21 2003-12-31 삼성전자주식회사 Color calibrator for flat panel display and method thereof
US6977658B2 (en) * 2002-06-27 2005-12-20 Broadcom Corporation System for and method of performing an opacity calculation in a 3D graphics system
US6954215B2 (en) * 2002-06-28 2005-10-11 Microsoft Corporation System and method for employing non-alpha channel image data in an alpha-channel-aware environment
JP3845045B2 (en) * 2002-07-23 2006-11-15 株式会社リコー Image processing apparatus, image processing method, image forming apparatus, printing apparatus, and host PC
FR2842977A1 (en) * 2002-07-24 2004-01-30 Total Immersion METHOD AND SYSTEM FOR ENABLING A USER TO MIX REAL-TIME SYNTHESIS IMAGES WITH VIDEO IMAGES
US7002599B2 (en) * 2002-07-26 2006-02-21 Sun Microsystems, Inc. Method and apparatus for hardware acceleration of clipping and graphical fill in display systems
US6857108B2 (en) * 2002-07-31 2005-02-15 Lsi Logic Corporation Interactive representation of structural dependencies in semiconductor design flows
US7257519B2 (en) * 2002-08-02 2007-08-14 Evans & Sutherland Computer Corporation System and method for weighted correction of an eyepoint position
US7176917B1 (en) 2002-08-09 2007-02-13 Avid Technology, Inc. Visual programming interface for a three-dimensional animation system for defining real time shaders using a real-time rendering engine application programming interface
US7508398B1 (en) 2002-08-27 2009-03-24 Nvidia Corporation Transparent antialiased memory access
US20040088682A1 (en) * 2002-11-05 2004-05-06 Thompson Ryan C. Method, program product, and apparatus for cache entry tracking, collision detection, and address reasignment in processor testcases
US7242400B2 (en) * 2002-11-13 2007-07-10 Ati Technologies Ulc Compression and decompression of data using plane equations
US7656416B2 (en) * 2002-11-27 2010-02-02 Ati Technologies, Inc. Apparatus for generating anti-aliased and stippled 3d lines, points and surfaces using multi-dimensional procedural texture coordinates
US7633506B1 (en) * 2002-11-27 2009-12-15 Ati Technologies Ulc Parallel pipeline graphics system
JPWO2004055697A1 (en) * 2002-12-13 2006-04-20 富士通株式会社 Processing method, processing apparatus, and computer program
US7928997B2 (en) * 2003-02-06 2011-04-19 Nvidia Corporation Digital image compositing using a programmable graphics processor
US8749561B1 (en) * 2003-03-14 2014-06-10 Nvidia Corporation Method and system for coordinated data execution using a primary graphics processor and a secondary graphics processor
EP1620806A2 (en) * 2003-04-03 2006-02-01 Koninklijke Philips Electronics N.V. Runtime configurable virtual video pipeline
US7148888B2 (en) * 2003-04-04 2006-12-12 Via Technologies, Inc. Head/data request in 3D graphics
US7259765B2 (en) 2003-04-04 2007-08-21 S3 Graphics Co., Ltd. Head/data scheduling in 3D graphics
US7714858B2 (en) * 2003-04-18 2010-05-11 Hewlett-Packard Development Company, L.P. Distributed rendering of interactive soft shadows
JP3966832B2 (en) 2003-04-28 2007-08-29 株式会社東芝 Drawing processing apparatus and drawing processing method
US7218331B2 (en) * 2003-05-13 2007-05-15 Via Technologies, Inc. Bounding box in 3D graphics
US20050017969A1 (en) * 2003-05-27 2005-01-27 Pradeep Sen Computer graphics rendering using boundary information
US7681112B1 (en) 2003-05-30 2010-03-16 Adobe Systems Incorporated Embedded reuse meta information
US7852405B1 (en) * 2003-06-27 2010-12-14 Zoran Corporation Method and apparatus for high definition capture
US8275910B1 (en) 2003-07-02 2012-09-25 Apple Inc. Source packet bridge
US7164420B2 (en) * 2003-07-24 2007-01-16 Autodesk, Inc. Ray tracing hierarchy
WO2005013066A2 (en) * 2003-07-25 2005-02-10 New York University Logic arrangement, data structure, system and method for miltilinear representation of multimodal data ensembles for synthesis, rotation and compression
US7139005B2 (en) * 2003-09-13 2006-11-21 Microsoft Corporation Optimized fixed-point mathematical library and graphics functions for a software-implemented graphics rendering system and method using a normalized homogenous coordinate system
US8788996B2 (en) 2003-09-15 2014-07-22 Nvidia Corporation System and method for configuring semiconductor functional circuits
US8732644B1 (en) 2003-09-15 2014-05-20 Nvidia Corporation Micro electro mechanical switch system and method for testing and configuring semiconductor functional circuits
US8775997B2 (en) * 2003-09-15 2014-07-08 Nvidia Corporation System and method for testing and configuring semiconductor functional circuits
CN100483463C (en) * 2003-09-17 2009-04-29 皇家飞利浦电子股份有限公司 System and method for rendering 3-D images on a 3-d image display screen
US7593010B2 (en) * 2003-09-18 2009-09-22 Microsoft Corporation Software-implemented transform and lighting module and pipeline for graphics rendering on embedded platforms using a fixed-point normalized homogenous coordinate system
JP2005100176A (en) * 2003-09-25 2005-04-14 Sony Corp Image processor and its method
JP4183082B2 (en) * 2003-09-26 2008-11-19 シャープ株式会社 3D image drawing apparatus and 3D image drawing method
KR100546383B1 (en) * 2003-09-29 2006-01-26 삼성전자주식회사 3D graphics rendering engine for processing an invisible fragment and method thereof
US8133115B2 (en) 2003-10-22 2012-03-13 Sony Computer Entertainment America Llc System and method for recording and displaying a graphical path in a video game
US8860737B2 (en) * 2003-10-29 2014-10-14 Nvidia Corporation Programmable graphics processor for multithreaded execution of programs
US7139003B1 (en) * 2003-12-15 2006-11-21 Nvidia Corporation Methods of processing graphics data including reading and writing buffers
US7836276B2 (en) * 2005-12-02 2010-11-16 Nvidia Corporation System and method for processing thread groups in a SIMD architecture
US8174531B1 (en) 2003-10-29 2012-05-08 Nvidia Corporation Programmable graphics processor for multithreaded execution of programs
US8035646B2 (en) * 2003-11-14 2011-10-11 Microsoft Corporation Systems and methods for downloading algorithmic elements to a coprocessor and corresponding techniques
KR20050047741A (en) * 2003-11-18 2005-05-23 삼성전자주식회사 Image processing device and method thereof
US7015914B1 (en) * 2003-12-10 2006-03-21 Nvidia Corporation Multiple data buffers for processing graphics data
US7053904B1 (en) * 2003-12-15 2006-05-30 Nvidia Corporation Position conflict detection and avoidance in a programmable graphics processor
US7053893B1 (en) * 2003-12-15 2006-05-30 Nvidia Corporation Position conflict detection and avoidance in a programmable graphics processor using tile coverage data
US7102645B2 (en) * 2003-12-15 2006-09-05 Seiko Epson Corporation Graphics display controller providing enhanced read/write efficiency for interfacing with a RAM-integrated graphics display device
US7420568B1 (en) * 2003-12-17 2008-09-02 Nvidia Corporation System and method for packing data in different formats in a tiled graphics memory
US8711161B1 (en) 2003-12-18 2014-04-29 Nvidia Corporation Functional component compensation reconfiguration system and method
US7221368B1 (en) * 2003-12-18 2007-05-22 Nvidia Corporation Stippled lines using direct distance evaluation
US7450120B1 (en) * 2003-12-19 2008-11-11 Nvidia Corporation Apparatus, system, and method for Z-culling
US8390619B1 (en) * 2003-12-22 2013-03-05 Nvidia Corporation Occlusion prediction graphics processing system and method
US8854364B1 (en) * 2003-12-22 2014-10-07 Nvidia Corporation Tight depth range occlusion prediction system and method
US8269769B1 (en) 2003-12-22 2012-09-18 Nvidia Corporation Occlusion prediction compression system and method
US7995056B1 (en) 2003-12-22 2011-08-09 Nvidia Corporation Culling data selection system and method
US7433364B2 (en) * 2003-12-24 2008-10-07 Intel Corporation Method for optimizing queuing performance
US9098943B1 (en) 2003-12-31 2015-08-04 Ziilabs Inc., Ltd. Multiple simultaneous bin sizes
US8643659B1 (en) 2003-12-31 2014-02-04 3Dlabs Inc., Ltd. Shader with global and instruction caches
US7281122B2 (en) * 2004-01-14 2007-10-09 Ati Technologies Inc. Method and apparatus for nested control flow of instructions using context information and instructions having extra bits
US20050195186A1 (en) * 2004-03-02 2005-09-08 Ati Technologies Inc. Method and apparatus for object based visibility culling
FI117655B (en) * 2004-03-25 2006-12-29 Cadfaster Oy A method for processing a computer-aided polygon model, a device and a computer program
US7609902B2 (en) * 2004-04-13 2009-10-27 Microsoft Corporation Implementation of discrete cosine transformation and its inverse on programmable graphics processor
US7231632B2 (en) 2004-04-16 2007-06-12 Apple Computer, Inc. System for reducing the number of programs necessary to render an image
US7636489B2 (en) * 2004-04-16 2009-12-22 Apple Inc. Blur computation algorithm
US8134561B2 (en) 2004-04-16 2012-03-13 Apple Inc. System for optimizing graphics operations
US7847800B2 (en) * 2004-04-16 2010-12-07 Apple Inc. System for emulating graphics operations
US7248265B2 (en) * 2004-04-16 2007-07-24 Apple Inc. System and method for processing graphics operations with graphics processing unit
US8704837B2 (en) 2004-04-16 2014-04-22 Apple Inc. High-level program interface for graphics operations
KR100601952B1 (en) * 2004-04-20 2006-07-14 삼성전자주식회사 Apparatus and method for reconstitution of three-dimensional graphic data
US7190366B2 (en) * 2004-05-14 2007-03-13 Nvidia Corporation Method and system for a general instruction raster stage that generates programmable pixel packets
US8411105B1 (en) 2004-05-14 2013-04-02 Nvidia Corporation Method and system for computing pixel parameters
US8687010B1 (en) 2004-05-14 2014-04-01 Nvidia Corporation Arbitrary size texture palettes for use in graphics systems
US8736620B2 (en) * 2004-05-14 2014-05-27 Nvidia Corporation Kill bit graphics processing system and method
US8432394B1 (en) 2004-05-14 2013-04-30 Nvidia Corporation Method and system for implementing clamped z value interpolation in a raster stage of a graphics pipeline
US8743142B1 (en) 2004-05-14 2014-06-03 Nvidia Corporation Unified data fetch graphics processing system and method
US8860722B2 (en) * 2004-05-14 2014-10-14 Nvidia Corporation Early Z scoreboard tracking system and method
ATE534114T1 (en) 2004-05-14 2011-12-15 Nvidia Corp PROGRAMMABLE PROCESSOR WITH LOW POWER CONSUMPTION
US8416242B1 (en) 2004-05-14 2013-04-09 Nvidia Corporation Method and system for interpolating level-of-detail in graphics processors
US8711155B2 (en) * 2004-05-14 2014-04-29 Nvidia Corporation Early kill removal graphics processing system and method
US8736628B1 (en) 2004-05-14 2014-05-27 Nvidia Corporation Single thread graphics processing system and method
US20060007234A1 (en) * 2004-05-14 2006-01-12 Hutchins Edward A Coincident graphics pixel scoreboard tracking system and method
US7389006B2 (en) * 2004-05-14 2008-06-17 Nvidia Corporation Auto software configurable register address space for low power programmable processor
US7091982B2 (en) * 2004-05-14 2006-08-15 Nvidia Corporation Low power programmable processor
US7079156B1 (en) * 2004-05-14 2006-07-18 Nvidia Corporation Method and system for implementing multiple high precision and low precision interpolators for a graphics pipeline
JP4451717B2 (en) 2004-05-31 2010-04-14 株式会社ソニー・コンピュータエンタテインメント Information processing apparatus and information processing method
US20050275733A1 (en) * 2004-06-10 2005-12-15 Philip Chao Method and apparatus of rendering a video image by polynomial evaluation
US7382377B1 (en) * 2004-06-17 2008-06-03 Nvidia Corporation Render to texture cull
ES2368929T3 (en) 2004-06-23 2011-11-23 Quin Media Arts And Sciences Inc. FORMATION OF SCULPTURAL IMAGES WITH OPTICAL MOSAICS.
US7397964B2 (en) * 2004-06-24 2008-07-08 Apple Inc. Gaussian blur approximation suitable for GPU
US8068103B2 (en) * 2004-06-24 2011-11-29 Apple Inc. User-interface design
US8130237B2 (en) * 2004-06-24 2012-03-06 Apple Inc. Resolution independent user interface design
US7761800B2 (en) 2004-06-25 2010-07-20 Apple Inc. Unified interest layer for user interface
US7652678B2 (en) * 2004-06-25 2010-01-26 Apple Inc. Partial display updates in a windowing system using a programmable graphics processing unit
US8566732B2 (en) 2004-06-25 2013-10-22 Apple Inc. Synchronization of widgets and dashboards
US8239749B2 (en) 2004-06-25 2012-08-07 Apple Inc. Procedurally expressing graphic objects for web pages
US20050285866A1 (en) * 2004-06-25 2005-12-29 Apple Computer, Inc. Display-wide visual effects for a windowing system using a programmable graphics processing unit
US8453065B2 (en) 2004-06-25 2013-05-28 Apple Inc. Preview and installation of user interface elements in a display environment
US8302020B2 (en) 2004-06-25 2012-10-30 Apple Inc. Widget authoring and editing environment
US7546543B2 (en) 2004-06-25 2009-06-09 Apple Inc. Widget authoring and editing environment
US7490295B2 (en) 2004-06-25 2009-02-10 Apple Inc. Layer for accessing user interface elements
US7755629B2 (en) * 2004-06-30 2010-07-13 Canon Kabushiki Kaisha Method of rendering graphic objects
US7518608B2 (en) * 2004-07-30 2009-04-14 Sony Corporation Z-depth matting of particles in image rendering
US7256796B1 (en) * 2004-08-03 2007-08-14 Nvidia Corporation Per-fragment control for writing an output buffer
US7400325B1 (en) * 2004-08-06 2008-07-15 Nvidia Corporation Culling before setup in viewport and culling unit
US20060033736A1 (en) * 2004-08-10 2006-02-16 Wang Andy W Enhanced Color and Lighting Model for Computer Graphics Productions
CN101084493A (en) * 2004-08-31 2007-12-05 硅奥普迪思公司 Method and apparatus for reading and writing pixel-aligned subframes in a frame buffer
US7218291B2 (en) * 2004-09-13 2007-05-15 Nvidia Corporation Increased scalability in the fragment shading pipeline
US8723231B1 (en) 2004-09-15 2014-05-13 Nvidia Corporation Semiconductor die micro electro-mechanical switch management system and method
US7286139B2 (en) * 2004-09-17 2007-10-23 Via Technologies, Inc. Partial guardband clipping
US20060061577A1 (en) * 2004-09-22 2006-03-23 Vijay Subramaniam Efficient interface and assembler for a graphics processor
US8711156B1 (en) 2004-09-30 2014-04-29 Nvidia Corporation Method and system for remapping processing elements in a pipeline of a graphics processing unit
US20060071933A1 (en) 2004-10-06 2006-04-06 Sony Computer Entertainment Inc. Application binary interface for multi-pass shaders
US20060082577A1 (en) * 2004-10-20 2006-04-20 Ugs Corp. System, method, and computer program product for dynamic shader generation
US7385604B1 (en) * 2004-11-04 2008-06-10 Nvidia Corporation Fragment scattering
JP4692956B2 (en) * 2004-11-22 2011-06-01 株式会社ソニー・コンピュータエンタテインメント Drawing processing apparatus and drawing processing method
US7227551B2 (en) * 2004-12-23 2007-06-05 Apple Inc. Manipulating text and graphic appearance
US8140975B2 (en) 2005-01-07 2012-03-20 Apple Inc. Slide show navigation
US7209139B1 (en) * 2005-01-07 2007-04-24 Electronic Arts Efficient rendering of similar objects in a three-dimensional graphics engine
JP4812073B2 (en) * 2005-01-31 2011-11-09 キヤノン株式会社 Image capturing apparatus, image capturing method, program, and recording medium
KR100612890B1 (en) * 2005-02-17 2006-08-14 삼성전자주식회사 Multi-effect expression method and apparatus in 3-dimension graphic image
US7242169B2 (en) * 2005-03-01 2007-07-10 Apple Inc. Method and apparatus for voltage compensation for parasitic impedance
US8089486B2 (en) * 2005-03-21 2012-01-03 Qualcomm Incorporated Tiled prefetched and cached depth buffer
CA2597436C (en) * 2005-03-24 2011-09-20 Lg Electronics Inc. Method of executing scanning in broadband wireless access system
JP2006293553A (en) * 2005-04-07 2006-10-26 Aisin Aw Co Ltd Rotation processor for font data and map display system
US7479965B1 (en) 2005-04-12 2009-01-20 Nvidia Corporation Optimized alpha blend for anti-aliased render
US9363481B2 (en) * 2005-04-22 2016-06-07 Microsoft Technology Licensing, Llc Protected media pipeline
US7499051B1 (en) 2005-04-29 2009-03-03 Adobe Systems Incorporated GPU assisted 3D compositing
US7463261B1 (en) * 2005-04-29 2008-12-09 Adobe Systems Incorporated Three-dimensional image compositing on a GPU utilizing multiple transformations
US7802028B2 (en) * 2005-05-02 2010-09-21 Broadcom Corporation Total dynamic sharing of a transaction queue
US7349066B2 (en) * 2005-05-05 2008-03-25 Asml Masktools B.V. Apparatus, method and computer program product for performing a model based optical proximity correction factoring neighbor influence
US8427496B1 (en) 2005-05-13 2013-04-23 Nvidia Corporation Method and system for implementing compression across a graphics bus interconnect
US7847755B1 (en) * 2005-05-23 2010-12-07 Glance Networks Method and apparatus for the identification and selective encoding of changed host display information
US7894528B2 (en) * 2005-05-25 2011-02-22 Yissum Research Development Company Of The Hebrew University Of Jerusalem Fast and robust motion computations using direct methods
US8543931B2 (en) 2005-06-07 2013-09-24 Apple Inc. Preview including theme based installation of user interface elements in a display environment
US7636126B2 (en) 2005-06-22 2009-12-22 Sony Computer Entertainment Inc. Delay matching in audio/video systems
US9298311B2 (en) * 2005-06-23 2016-03-29 Apple Inc. Trackpad sensitivity compensation
US7432937B2 (en) * 2005-06-30 2008-10-07 Intel Corporation System and method for concave polygon rasterization
US7496416B2 (en) 2005-08-01 2009-02-24 Luxology, Llc Input/output curve editor
US20070035553A1 (en) * 2005-08-12 2007-02-15 Microsoft Corporation General framework for aligning textures
US7436412B2 (en) * 2005-08-24 2008-10-14 Qualcomm Incorporated Graphics engine with efficient interpolation
US7551177B2 (en) * 2005-08-31 2009-06-23 Ati Technologies, Inc. Methods and apparatus for retrieving and combining samples of graphics information
US8014615B2 (en) * 2005-09-02 2011-09-06 Adobe Systems Incorporated System and method for decompressing video data and alpha channel data using a single stream
US8189908B2 (en) * 2005-09-02 2012-05-29 Adobe Systems, Inc. System and method for compressing video data and alpha channel data using a single stream
US7433191B2 (en) * 2005-09-30 2008-10-07 Apple Inc. Thermal contact arrangement
US7441230B2 (en) 2005-10-07 2008-10-21 Lucasfilm Entertainment Company Ltd. Method of utilizing product proxies with a dependency graph
US8144149B2 (en) * 2005-10-14 2012-03-27 Via Technologies, Inc. System and method for dynamically load balancing multiple shader stages in a shared pool of processing units
US8266232B2 (en) * 2005-10-15 2012-09-11 International Business Machines Corporation Hardware processing of commands within virtual client computing environment
US7954064B2 (en) 2005-10-27 2011-05-31 Apple Inc. Multiple dashboards
US7752556B2 (en) 2005-10-27 2010-07-06 Apple Inc. Workflow widgets
US9104294B2 (en) 2005-10-27 2015-08-11 Apple Inc. Linked widgets
US8543824B2 (en) 2005-10-27 2013-09-24 Apple Inc. Safe distribution and use of content
US7743336B2 (en) 2005-10-27 2010-06-22 Apple Inc. Widget security
US7414624B2 (en) * 2005-10-28 2008-08-19 Intel Corporation Apparatus and method for a frustum culling algorithm suitable for hardware implementation
US20070097139A1 (en) * 2005-11-02 2007-05-03 Chao-Chin Chen Method and apparatus of primitive filter in graphic process applications
GB0524804D0 (en) 2005-12-05 2006-01-11 Falanx Microsystems As Method of and apparatus for processing graphics
US7934255B1 (en) * 2005-11-08 2011-04-26 Nvidia Corporation Apparatus, system, and method for offloading packet classification
US8294731B2 (en) * 2005-11-15 2012-10-23 Advanced Micro Devices, Inc. Buffer management in vector graphics hardware
US7707514B2 (en) 2005-11-18 2010-04-27 Apple Inc. Management of user interface elements in a display environment
US8624909B2 (en) * 2005-11-21 2014-01-07 Vixs Systems Inc. Image processing system and method thereof
US7598711B2 (en) * 2005-11-23 2009-10-06 Apple Inc. Power source switchover apparatus and method
KR100957760B1 (en) * 2005-11-30 2010-05-13 후지쯔 마이크로일렉트로닉스 가부시키가이샤 Three-dimensional graphic apparatus, three-dimensional graphic method, and computer readable recording medium having three-dimensional program
US8803872B2 (en) * 2005-12-01 2014-08-12 Intel Corporation Computer graphics processor and method for rendering a three-dimensional image on a display screen
US7439988B1 (en) 2005-12-05 2008-10-21 Nvidia Corporation Apparatus, system, and method for clipping graphics primitives with respect to a clipping plane
US7616218B1 (en) 2005-12-05 2009-11-10 Nvidia Corporation Apparatus, system, and method for clipping graphics primitives
US7434032B1 (en) 2005-12-13 2008-10-07 Nvidia Corporation Tracking register usage during multithreaded processing using a scoreboard having separate memory regions and storing sequential register size indicators
US7423642B2 (en) * 2005-12-14 2008-09-09 Winbond Electronics Corporation Efficient video frame capturing
US7593018B1 (en) * 2005-12-14 2009-09-22 Nvidia Corp. Method and apparatus for providing explicit weights for texture filtering
US8701091B1 (en) 2005-12-15 2014-04-15 Nvidia Corporation Method and system for providing a generic console interface for a graphics application
US9123173B2 (en) * 2005-12-15 2015-09-01 Nvidia Corporation Method for rasterizing non-rectangular tile groups in a raster stage of a graphics pipeline
US8698811B1 (en) 2005-12-15 2014-04-15 Nvidia Corporation Nested boustrophedonic patterns for rasterization
US7791617B2 (en) * 2005-12-19 2010-09-07 Nvidia Corporation Method and system for rendering polygons having abutting edges
US7714877B1 (en) 2005-12-19 2010-05-11 Nvidia Corporation Apparatus, system, and method for determining clipping distances
US9117309B1 (en) 2005-12-19 2015-08-25 Nvidia Corporation Method and system for rendering polygons with a bounding box in a graphics processor unit
US8390645B1 (en) 2005-12-19 2013-03-05 Nvidia Corporation Method and system for rendering connecting antialiased line segments
US7420572B1 (en) * 2005-12-19 2008-09-02 Nvidia Corporation Apparatus, system, and method for clipping graphics primitives with accelerated context switching
US8300059B2 (en) * 2006-02-03 2012-10-30 Ati Technologies Ulc Method and apparatus for selecting a mip map level based on a min-axis value for texture mapping
JP4734137B2 (en) * 2006-02-23 2011-07-27 株式会社バンダイナムコゲームス Program, information storage medium, and image generation system
JP4734138B2 (en) * 2006-02-23 2011-07-27 株式会社バンダイナムコゲームス Program, information storage medium, and image generation system
JP4782583B2 (en) * 2006-02-23 2011-09-28 株式会社バンダイナムコゲームス Program, information storage medium, and image generation system
US8006236B1 (en) * 2006-02-24 2011-08-23 Nvidia Corporation System and method for compiling high-level primitive programs into primitive program micro-code
US7825933B1 (en) * 2006-02-24 2010-11-02 Nvidia Corporation Managing primitive program vertex attributes as per-attribute arrays
US8171461B1 (en) 2006-02-24 2012-05-01 Nvidia Coporation Primitive program compilation for flat attributes with provoking vertex independence
US7891012B1 (en) 2006-03-01 2011-02-15 Nvidia Corporation Method and computer-usable medium for determining the authorization status of software
US8452981B1 (en) * 2006-03-01 2013-05-28 Nvidia Corporation Method for author verification and software authorization
TWI319166B (en) * 2006-03-06 2010-01-01 Via Tech Inc Method and related apparatus for graphic processing
JP2007287084A (en) * 2006-04-20 2007-11-01 Fuji Xerox Co Ltd Image processor and program
JP2007287085A (en) * 2006-04-20 2007-11-01 Fuji Xerox Co Ltd Program and device for processing images
JP5085642B2 (en) * 2006-04-20 2012-11-28 テレフオンアクチーボラゲット エル エム エリクソン(パブル) Method for compressing an image block, method for processing a compressed representation of an image block, block compressor and block decompressor
US8766995B2 (en) * 2006-04-26 2014-07-01 Qualcomm Incorporated Graphics system with configurable caches
WO2007130933A2 (en) * 2006-05-01 2007-11-15 Jeffrey W Bezanson Apparatuses, methods and systems for vector operations and storage in matrix models
US7965859B2 (en) 2006-05-04 2011-06-21 Sony Computer Entertainment Inc. Lighting control of a user environment via a display device
US7880746B2 (en) 2006-05-04 2011-02-01 Sony Computer Entertainment Inc. Bandwidth management through lighting control of a user environment via a display device
SG137754A1 (en) * 2006-05-12 2007-12-28 Nvidia Corp Antialiasing using multiple display heads of a graphics processor
US20070268289A1 (en) * 2006-05-16 2007-11-22 Chun Yu Graphics system with dynamic reposition of depth engine
US7395180B2 (en) * 2006-05-17 2008-07-01 Lockheed Martin Corporation Efficient translation of data from a two-dimensional array to a wedge
US8884972B2 (en) 2006-05-25 2014-11-11 Qualcomm Incorporated Graphics processor with arithmetic and elementary function units
US8869147B2 (en) * 2006-05-31 2014-10-21 Qualcomm Incorporated Multi-threaded processor with deferred thread output control
WO2007142326A1 (en) * 2006-06-09 2007-12-13 Aisin Aw Co., Ltd. Data update system, terminal device, server device, and data update method
US8644643B2 (en) 2006-06-14 2014-02-04 Qualcomm Incorporated Convolution filtering in a graphics processor
US20070291031A1 (en) * 2006-06-15 2007-12-20 Right Hemisphere Limited Three dimensional geometric data correction
US7940262B2 (en) * 2006-06-15 2011-05-10 Right Hemisphere Limited Unification and part hiding in three dimensional geometric data
US8766996B2 (en) * 2006-06-21 2014-07-01 Qualcomm Incorporated Unified virtual addressed register file
US8928676B2 (en) * 2006-06-23 2015-01-06 Nvidia Corporation Method for parallel fine rasterization in a raster stage of a graphics pipeline
JP2008009696A (en) * 2006-06-29 2008-01-17 Fuji Xerox Co Ltd Image processor and program
JP4795138B2 (en) * 2006-06-29 2011-10-19 富士ゼロックス株式会社 Image processing apparatus and program
US8477134B1 (en) 2006-06-30 2013-07-02 Nvidia Corporation Conservative triage of polygon status using low precision edge evaluation and high precision edge evaluation
US8284204B2 (en) * 2006-06-30 2012-10-09 Nokia Corporation Apparatus, method and a computer program product for providing a unified graphics pipeline for stereoscopic rendering
US8560495B1 (en) * 2006-07-07 2013-10-15 Sybase, Inc. System and method for synchronizing message processing in a continuous processing system
JP4979287B2 (en) * 2006-07-14 2012-07-18 富士ゼロックス株式会社 Image processing apparatus and program
US8633927B2 (en) * 2006-07-25 2014-01-21 Nvidia Corporation Re-render acceleration of frame with lighting change
US8085264B1 (en) 2006-07-26 2011-12-27 Nvidia Corporation Tile output using multiple queue output buffering in a raster stage
US9070213B2 (en) * 2006-07-26 2015-06-30 Nvidia Corporation Tile based precision rasterization in a graphics pipeline
US8963932B1 (en) 2006-08-01 2015-02-24 Nvidia Corporation Method and apparatus for visualizing component workloads in a unified shader GPU architecture
US7778800B2 (en) * 2006-08-01 2010-08-17 Nvidia Corporation Method and system for calculating performance parameters for a processor
US8607151B2 (en) * 2006-08-01 2013-12-10 Nvidia Corporation Method and system for debugging a graphics pipeline subunit
US8436870B1 (en) 2006-08-01 2013-05-07 Nvidia Corporation User interface and method for graphical processing analysis
US8436864B2 (en) * 2006-08-01 2013-05-07 Nvidia Corporation Method and user interface for enhanced graphical operation organization
US7952588B2 (en) * 2006-08-03 2011-05-31 Qualcomm Incorporated Graphics processing unit with extended vertex cache
US8869027B2 (en) 2006-08-04 2014-10-21 Apple Inc. Management and generation of dashboards
US8493388B2 (en) * 2006-08-09 2013-07-23 Siemens Medical Solutions Usa, Inc. Modular volume rendering using visual programming
KR20080014402A (en) * 2006-08-11 2008-02-14 삼성전자주식회사 Method and apparatus for processing computer graphics data
US7852347B1 (en) * 2006-08-24 2010-12-14 Nvidia Corporation Texture map pixel pairing optimization
KR100745768B1 (en) * 2006-08-29 2007-08-02 삼성전자주식회사 Method for calculate lod value for reducing power consumption and 3 dimension rendering system using the same
US7905610B1 (en) * 2006-08-29 2011-03-15 Nvidia Corporation Graphics processor system and associated method for projecting an image onto a three-dimensional object
US8237739B2 (en) 2006-09-12 2012-08-07 Qualcomm Incorporated Method and device for performing user-defined clipping in object space
JP4995827B2 (en) * 2006-09-13 2012-08-08 パナソニック株式会社 Image processing apparatus, integrated circuit for image processing, image processing system, input assembler apparatus, integrated circuit for input assembly
JP4079378B2 (en) 2006-09-21 2008-04-23 株式会社コナミデジタルエンタテインメント Image processing apparatus, image processing apparatus control method, and program
US8537168B1 (en) 2006-11-02 2013-09-17 Nvidia Corporation Method and system for deferred coverage mask generation in a raster stage
US8427487B1 (en) 2006-11-02 2013-04-23 Nvidia Corporation Multiple tile output using interface compression in a raster stage
US8237738B1 (en) 2006-11-02 2012-08-07 Nvidia Corporation Smooth rasterization of polygonal graphics primitives
US8482567B1 (en) 2006-11-03 2013-07-09 Nvidia Corporation Line rasterization techniques
US8232991B1 (en) 2006-11-03 2012-07-31 Nvidia Corporation Z-test result reconciliation with multiple partitions
US7701459B1 (en) * 2006-11-03 2010-04-20 Nvidia Corporation Primitive oriented assembly for parallel vertex/geometry processing
US8300050B2 (en) 2006-11-28 2012-10-30 Adobe Systems Incorporated Temporary low resolution rendering of 3D objects
US8059124B2 (en) 2006-11-28 2011-11-15 Adobe Systems Incorporated Temporary non-tiled rendering of 3D objects
US9965886B2 (en) 2006-12-04 2018-05-08 Arm Norway As Method of and apparatus for processing graphics
GB0710795D0 (en) * 2007-06-05 2007-07-18 Arm Norway As Method of and apparatus for processing graphics
WO2008073455A1 (en) * 2006-12-11 2008-06-19 Koplar Interactive Systems International, L.L.C. Spatial data encoding and decoding
EP2104930A2 (en) 2006-12-12 2009-09-30 Evans & Sutherland Computer Corporation System and method for aligning rgb light in a single modulator projector
US8736627B2 (en) * 2006-12-19 2014-05-27 Via Technologies, Inc. Systems and methods for providing a shared buffer in a multiple FIFO environment
US7580035B2 (en) * 2006-12-28 2009-08-25 Intel Corporation Real-time collision detection using clipping
CN102160087B (en) * 2007-01-05 2013-09-18 兰德马克绘图国际公司,哈里伯顿公司 Systems and methods for visualizing multiple volumetric data sets in real time
US7982733B2 (en) * 2007-01-05 2011-07-19 Qualcomm Incorporated Rendering 3D video images on a stereo-enabled display
ITMI20070038A1 (en) * 2007-01-12 2008-07-13 St Microelectronics Srl RENDERING DEVICE FOR GRAPHICS WITH THREE DIMENSIONS WITH SORT-MIDDLE TYPE ARCHITECTURE.
US7746355B1 (en) * 2007-01-24 2010-06-29 Vivante Corporation Method for distributed clipping outside of view volume
WO2008091198A1 (en) * 2007-01-24 2008-07-31 Swiftfoot Graphics Ab Method, display adapter and computer program product for improved graphics performance by using a replaceable culling program
US8549500B2 (en) * 2007-02-14 2013-10-01 The Mathworks, Inc. Saving and loading graphical processing unit (GPU) arrays providing high computational capabilities in a computing environment
WO2008103775A2 (en) 2007-02-20 2008-08-28 Pixologic, Inc. System and method for interactive masking and modifying of 3d objects
US7473258B2 (en) * 2007-03-08 2009-01-06 Cardica, Inc. Surgical stapler
US8471862B2 (en) * 2007-03-09 2013-06-25 Ati Technologies Ulc Offset tiles in vector graphics
US7694193B2 (en) * 2007-03-13 2010-04-06 Hewlett-Packard Development Company, L.P. Systems and methods for implementing a stride value for accessing memory
JP4446201B2 (en) * 2007-03-30 2010-04-07 アイシン・エィ・ダブリュ株式会社 Image recognition apparatus and image recognition method
US8155826B2 (en) * 2007-03-30 2012-04-10 Aisin Aw Co., Ltd. Vehicle behavior learning apparatuses, methods, and programs
WO2008123822A1 (en) * 2007-04-04 2008-10-16 Telefonaktiebolaget Lm Ericsson (Publ) Vector-based image processing
US10605610B2 (en) * 2007-04-09 2020-03-31 Ian Cummings Apparatus and methods for reducing data transmission in wireless client-server navigation systems
JP4588736B2 (en) * 2007-04-12 2010-12-01 富士フイルム株式会社 Image processing method, apparatus, and program
CA2686121C (en) * 2007-04-16 2016-06-21 Sunfish Studio, Llc Single-pass and order-independent transparency in computer graphics using constant memory
GB2448717B (en) * 2007-04-25 2012-09-19 David Hostettler Wain Method and apparatus for the efficient animation of textures based on images and graphical components
US8203560B2 (en) * 2007-04-27 2012-06-19 Sony Corporation Method for predictively splitting procedurally generated particle data into screen-space boxes
US20080273113A1 (en) * 2007-05-02 2008-11-06 Windbond Electronics Corporation Integrated graphics and KVM system
US7876677B2 (en) * 2007-05-22 2011-01-25 Apple Inc. Transmission control protocol queue sorting
FR2917211A1 (en) * 2007-06-08 2008-12-12 St Microelectronics Sa METHOD AND DEVICE FOR GENERATING GRAPHICS
US8558832B1 (en) * 2007-06-19 2013-10-15 Nvida Corporation System, method, and computer program product for generating a plurality of two-dimensional images and depth maps for a scene at a point in time
KR101378372B1 (en) * 2007-07-12 2014-03-27 삼성전자주식회사 Digital image processing apparatus, method for controlling the same, and recording medium storing program to implement the method
US8954871B2 (en) 2007-07-18 2015-02-10 Apple Inc. User-centric widgets and dashboards
US7805579B2 (en) * 2007-07-31 2010-09-28 International Business Machines Corporation Methods and arrangements for multi-buffering data
US7925100B2 (en) * 2007-07-31 2011-04-12 Microsoft Corporation Tiled packaging of vector image data
US8667415B2 (en) 2007-08-06 2014-03-04 Apple Inc. Web widgets
US8441497B1 (en) 2007-08-07 2013-05-14 Nvidia Corporation Interpolation of vertex attributes in a graphics processor
US8296738B1 (en) 2007-08-13 2012-10-23 Nvidia Corporation Methods and systems for in-place shader debugging and performance tuning
US8599208B2 (en) * 2007-08-15 2013-12-03 Nvidia Corporation Shared readable and writeable global values in a graphics processor unit pipeline
US8736624B1 (en) 2007-08-15 2014-05-27 Nvidia Corporation Conditional execution flag in graphics applications
US8314803B2 (en) * 2007-08-15 2012-11-20 Nvidia Corporation Buffering deserialized pixel data in a graphics processor unit pipeline
US8521800B1 (en) 2007-08-15 2013-08-27 Nvidia Corporation Interconnected arithmetic logic units
US8775777B2 (en) * 2007-08-15 2014-07-08 Nvidia Corporation Techniques for sourcing immediate values from a VLIW
US9035957B1 (en) 2007-08-15 2015-05-19 Nvidia Corporation Pipeline debug statistics system and method
US20090046105A1 (en) * 2007-08-15 2009-02-19 Bergland Tyson J Conditional execute bit in a graphics processor unit pipeline
US9183607B1 (en) 2007-08-15 2015-11-10 Nvidia Corporation Scoreboard cache coherence in a graphics pipeline
US8249391B2 (en) * 2007-08-24 2012-08-21 Ancestry.com Operations, Inc. User interface method for skew correction
US8156467B2 (en) 2007-08-27 2012-04-10 Adobe Systems Incorporated Reusing components in a running application
KR100933366B1 (en) * 2007-09-13 2009-12-22 한국전자통신연구원 Router device with black box function and network system including the device
JP4501983B2 (en) * 2007-09-28 2010-07-14 アイシン・エィ・ダブリュ株式会社 Parking support system, parking support method, parking support program
US8176466B2 (en) 2007-10-01 2012-05-08 Adobe Systems Incorporated System and method for generating an application fragment
US8724483B2 (en) 2007-10-22 2014-05-13 Nvidia Corporation Loopback configuration for bi-directional interfaces
KR101407639B1 (en) * 2007-10-22 2014-06-16 삼성전자주식회사 Apparatus and method for rendering 3D Graphic object
US8638341B2 (en) * 2007-10-23 2014-01-28 Qualcomm Incorporated Antialiasing of two-dimensional vector images
US8760450B2 (en) * 2007-10-30 2014-06-24 Advanced Micro Devices, Inc. Real-time mesh simplification using the graphics processing unit
US7765500B2 (en) * 2007-11-08 2010-07-27 Nvidia Corporation Automated generation of theoretical performance analysis based upon workload and design configuration
US8063903B2 (en) * 2007-11-09 2011-11-22 Nvidia Corporation Edge evaluation techniques for graphics hardware
US8035641B1 (en) 2007-11-28 2011-10-11 Adobe Systems Incorporated Fast depth of field simulation
US9153211B1 (en) * 2007-12-03 2015-10-06 Nvidia Corporation Method and system for tracking accesses to virtual addresses in graphics contexts
US8040349B1 (en) 2007-12-04 2011-10-18 Nvidia Corporation System and method for structuring an A-buffer
US8026912B1 (en) * 2007-12-04 2011-09-27 Nvidia Corporation System and method for structuring an A-buffer
US7940280B2 (en) * 2007-12-06 2011-05-10 Seiko Epson Corporation System and method for color format conversion in a graphics environment
US8179394B1 (en) 2007-12-13 2012-05-15 Nvidia Corporation Cull streams for fine-grained rendering predication
US8102393B1 (en) 2007-12-13 2012-01-24 Nvidia Corporation Cull streams for fine-grained rendering predication
US9489767B1 (en) * 2007-12-13 2016-11-08 Nvidia Corporation Cull streams for fine-grained rendering predication
US8878849B2 (en) * 2007-12-14 2014-11-04 Nvidia Corporation Horizon split ambient occlusion
US8780123B2 (en) 2007-12-17 2014-07-15 Nvidia Corporation Interrupt handling techniques in the rasterizer of a GPU
US9064333B2 (en) 2007-12-17 2015-06-23 Nvidia Corporation Interrupt handling techniques in the rasterizer of a GPU
CN101216944B (en) * 2008-01-07 2011-08-03 北大方正集团有限公司 A method and device for morphing shading in the process of typeset
US20090184972A1 (en) * 2008-01-18 2009-07-23 Qualcomm Incorporated Multi-buffer support for off-screen surfaces in a graphics processing system
US20090189896A1 (en) * 2008-01-25 2009-07-30 Via Technologies, Inc. Graphics Processor having Unified Shader Unit
US9214007B2 (en) * 2008-01-25 2015-12-15 Via Technologies, Inc. Graphics processor having unified cache system
US20110149340A1 (en) * 2008-01-30 2011-06-23 Ramot At Tel-Aviv University Ltd. Method, system and computer program product for manipulating a graphic entity
GB0801812D0 (en) * 2008-01-31 2008-03-05 Arm Noway As Methods of and apparatus for processing computer graphics
US9619304B2 (en) 2008-02-05 2017-04-11 Adobe Systems Incorporated Automatic connections between application components
US8098251B2 (en) * 2008-02-22 2012-01-17 Qualcomm Incorporated System and method for instruction latency reduction in graphics processing
KR100866573B1 (en) * 2008-02-22 2008-11-03 인하대학교 산학협력단 A point-based rendering method using visibility map
KR100914171B1 (en) 2008-02-28 2009-08-28 한국전자통신연구원 Apparatus and method for depth based image rendering on mobile broadcasting
US7675513B2 (en) * 2008-03-14 2010-03-09 Evans & Sutherland Computer Corp. System and method for displaying stereo images
GB2458488C (en) 2008-03-19 2018-09-12 Imagination Tech Ltd Untransformed display lists in a tile based rendering system
US7984317B2 (en) 2008-03-24 2011-07-19 Apple Inc. Hardware-based power management of functional blocks
US8125494B2 (en) * 2008-04-03 2012-02-28 American Panel Corporation Method for mapping optical properties for a display device
US8448002B2 (en) * 2008-04-10 2013-05-21 Nvidia Corporation Clock-gated series-coupled data processing modules
US8923385B2 (en) 2008-05-01 2014-12-30 Nvidia Corporation Rewind-enabled hardware encoder
US8681861B2 (en) 2008-05-01 2014-03-25 Nvidia Corporation Multistandard hardware video encoder
US8358317B2 (en) 2008-05-23 2013-01-22 Evans & Sutherland Computer Corporation System and method for displaying a planar image on a curved surface
CN102047315B (en) * 2008-05-30 2015-09-09 先进微装置公司 The computing system of easily extensible and integration
GB0810205D0 (en) * 2008-06-04 2008-07-09 Advanced Risc Mach Ltd Graphics processing systems
US8702248B1 (en) 2008-06-11 2014-04-22 Evans & Sutherland Computer Corporation Projection method for reducing interpixel gaps on a viewing surface
US8656293B1 (en) 2008-07-29 2014-02-18 Adobe Systems Incorporated Configuring mobile devices
US8427497B1 (en) * 2008-08-01 2013-04-23 Marvell International Ltd. Methods and apparatuses for processing cached image data
US8553041B1 (en) 2008-09-10 2013-10-08 Nvidia Corporation System and method for structuring an A-buffer to support multi-sample anti-aliasing
US8654135B1 (en) * 2008-09-10 2014-02-18 Nvidia Corporation A-Buffer compression for different compression formats
US8130223B1 (en) 2008-09-10 2012-03-06 Nvidia Corporation System and method for structuring an A-buffer to support multi-sample anti-aliasing
US8370759B2 (en) 2008-09-29 2013-02-05 Ancestry.com Operations Inc Visualizing, creating and editing blending modes methods and systems
US9336624B2 (en) * 2008-10-07 2016-05-10 Mitsubishi Electric Research Laboratories, Inc. Method and system for rendering 3D distance fields
KR101496340B1 (en) 2008-10-31 2015-03-04 삼성전자주식회사 Processor and method for controling memory
US8077378B1 (en) 2008-11-12 2011-12-13 Evans & Sutherland Computer Corporation Calibration system and method for light modulation device
WO2010062790A1 (en) * 2008-11-25 2010-06-03 Sony Computer Entertainment America Inc. Computer graphics method for aggregating light sources per-vertex and interpolating color and direction as one entity
US8355022B2 (en) * 2008-11-25 2013-01-15 Sony Computer Entertainment America Llc Method and apparatus for aggregating light sources per-vertex in computer graphics
US20100128038A1 (en) * 2008-11-25 2010-05-27 Sony Computer Entertainment America Inc. Method and apparatus for interpolating color and direction as one entity in computer graphics
CA2745421A1 (en) * 2008-12-01 2010-06-10 Life Image Inc. Medical imaging viewer
KR101511273B1 (en) * 2008-12-29 2015-04-10 삼성전자주식회사 System and method for 3d graphic rendering based on multi-core processor
GB0900700D0 (en) 2009-01-15 2009-03-04 Advanced Risc Mach Ltd Methods of and apparatus for processing graphics
CN102388620B (en) * 2009-02-01 2014-10-29 Lg电子株式会社 Broadcast receiver and 3d video data processing method
US8384740B1 (en) * 2009-02-24 2013-02-26 A9.Com, Inc. Method and system for virtually placing a tangible item on an appendage
US8854379B2 (en) * 2009-02-25 2014-10-07 Empire Technology Development Llc Routing across multicore networks using real world or modeled data
US8095560B2 (en) * 2009-02-26 2012-01-10 Yahoo! Inc. Edge attribute aggregation in a directed graph
US20100241638A1 (en) * 2009-03-18 2010-09-23 O'sullivan Patrick Joseph Sorting contacts
US8330767B2 (en) * 2009-03-24 2012-12-11 Advanced Micro Devices, Inc. Method and apparatus for angular invariant texture level of detail generation
CN101859330B (en) * 2009-04-09 2012-11-21 辉达公司 Method for verifying integrated circuit effectiveness models
KR100927128B1 (en) * 2009-04-30 2009-11-18 주식회사 넥서스칩스 Device and method of processing 3-dimension graphic using tile dirty table
JP5304443B2 (en) * 2009-05-28 2013-10-02 富士通セミコンダクター株式会社 Drawing data processing method, drawing system, and drawing data creation program
US8294714B1 (en) * 2009-06-26 2012-10-23 Nvidia Corporation Accelerated rendering with temporally interleaved details
KR101649098B1 (en) * 2009-06-30 2016-08-19 삼성전자주식회사 Apparatus and method for rendering using sensor in portable terminal
US7973705B2 (en) * 2009-07-17 2011-07-05 Garmin Switzerland Gmbh Marine bump map display
US9142057B2 (en) * 2009-09-03 2015-09-22 Advanced Micro Devices, Inc. Processing unit with a plurality of shader engines
US9300969B2 (en) 2009-09-09 2016-03-29 Apple Inc. Video storage
GB2473682B (en) * 2009-09-14 2011-11-16 Sony Comp Entertainment Europe A method of determining the state of a tile based deferred re ndering processor and apparatus thereof
US20110063309A1 (en) * 2009-09-16 2011-03-17 Nvidia Corporation User interface for co-processing techniques on heterogeneous graphics processing units
US8692829B2 (en) * 2009-10-05 2014-04-08 Nvidia Corporation Calculation of plane equations after determination of Z-buffer visibility
US9058672B2 (en) * 2009-10-06 2015-06-16 Nvidia Corporation Using a pixel offset for evaluating a plane equation
US9438861B2 (en) * 2009-10-06 2016-09-06 Microsoft Technology Licensing, Llc Integrating continuous and sparse streaming data
JP6035148B2 (en) * 2009-12-08 2016-11-30 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. Ablation treatment plan and device
CN102087752B (en) * 2009-12-08 2013-11-20 鸿富锦精密工业(深圳)有限公司 Illumination environment simulation system and method thereof
US9530189B2 (en) 2009-12-31 2016-12-27 Nvidia Corporation Alternate reduction ratios and threshold mechanisms for framebuffer compression
TWI482998B (en) * 2010-01-11 2015-05-01 Hon Hai Prec Ind Co Ltd Illumination environment simulation system and method
JP5571977B2 (en) * 2010-03-01 2014-08-13 キヤノン株式会社 Image processing device
US9331869B2 (en) 2010-03-04 2016-05-03 Nvidia Corporation Input/output request packet handling techniques by a device specific kernel mode driver
US9058685B2 (en) * 2010-03-11 2015-06-16 Broadcom Corporation Method and system for controlling a 3D processor using a control list in memory
US8320622B2 (en) * 2010-03-29 2012-11-27 Sharp Laboratories Of America, Inc. Color gradient object tracking
US10786736B2 (en) 2010-05-11 2020-09-29 Sony Interactive Entertainment LLC Placement of user information in a game space
US20110285736A1 (en) 2010-05-21 2011-11-24 Kilgard Mark J Decomposing cubic bèzier segments for tessellation-free stencil filling
KR101016075B1 (en) * 2010-06-04 2011-02-17 김시용 Wiper blade
US8593466B2 (en) * 2010-06-08 2013-11-26 Intel Corporation Tile rendering for image processing
US9053562B1 (en) 2010-06-24 2015-06-09 Gregory S. Rabin Two dimensional to three dimensional moving image converter
US10109103B2 (en) 2010-06-30 2018-10-23 Barry L. Jenkins Method of determining occluded ingress and egress routes using nav-cell to nav-cell visibility pre-computation
US8493404B2 (en) 2010-08-24 2013-07-23 Qualcomm Incorporated Pixel rendering on display
KR101064178B1 (en) * 2010-08-24 2011-09-14 한국과학기술원 System and method for managing buffer cache
EP3543958B1 (en) * 2010-09-13 2020-11-11 Barry Lynn Jenkins System and method of delivering and controlling streaming interactive media comprising predetermined packets of geometric, texture, lighting and other data which are rendered on a receiving device
KR101719485B1 (en) 2010-09-20 2017-03-27 삼성전자주식회사 Apparatus and method for early fragment discarding in graphic processing unit
US8811699B2 (en) * 2010-09-22 2014-08-19 Siemens Aktiengesellschaft Detection of landmarks and key-frames in cardiac perfusion MRI using a joint spatial-temporal context model
US9171350B2 (en) 2010-10-28 2015-10-27 Nvidia Corporation Adaptive resolution DGPU rendering to provide constant framerate with free IGPU scale up
US9245047B2 (en) 2010-12-10 2016-01-26 Wyse Technology L.L.C. Methods and systems for facilitating a remote desktop session utilizing a remote desktop client common interface
US9395885B1 (en) 2010-12-10 2016-07-19 Wyse Technology L.L.C. Methods and systems for a remote desktop session utilizing HTTP header
US9535560B1 (en) 2010-12-10 2017-01-03 Wyse Technology L.L.C. Methods and systems for facilitating a remote desktop session for a web browser and a remote desktop server
US9430036B1 (en) * 2010-12-10 2016-08-30 Wyse Technology L.L.C. Methods and systems for facilitating accessing and controlling a remote desktop of a remote machine in real time by a windows web browser utilizing HTTP
US8949726B2 (en) 2010-12-10 2015-02-03 Wyse Technology L.L.C. Methods and systems for conducting a remote desktop session via HTML that supports a 2D canvas and dynamic drawing
US9244912B1 (en) 2010-12-10 2016-01-26 Wyse Technology L.L.C. Methods and systems for facilitating a remote desktop redrawing session utilizing HTML
KR20120065589A (en) * 2010-12-13 2012-06-21 삼성전자주식회사 Apparatus and method for tile binning for low power
US9477597B2 (en) 2011-03-25 2016-10-25 Nvidia Corporation Techniques for different memory depths on different partitions
US8422770B2 (en) * 2011-03-30 2013-04-16 Mckesson Financial Holdings Method, apparatus and computer program product for displaying normalized medical images
US8701057B2 (en) 2011-04-11 2014-04-15 Nvidia Corporation Design, layout, and manufacturing techniques for multivariant integrated circuits
CN102739998B (en) * 2011-05-11 2017-03-01 新奥特(北京)视频技术有限公司 A kind of implementation method of space transformation in three-dimensional space
GB2491156B (en) 2011-05-25 2019-08-07 Advanced Risc Mach Ltd Processing pipeline control
AU2011202508B2 (en) 2011-05-27 2013-05-16 Canon Kabushiki Kaisha Method, apparatus and system for rendering an object on a page
US9311433B2 (en) * 2011-05-27 2016-04-12 Airbus Operations S.L. Systems and methods for improving the execution of computational algorithms
US9342817B2 (en) 2011-07-07 2016-05-17 Sony Interactive Entertainment LLC Auto-creating groups for sharing photos
US9652560B1 (en) 2011-07-18 2017-05-16 Apple Inc. Non-blocking memory management unit
US9529712B2 (en) 2011-07-26 2016-12-27 Nvidia Corporation Techniques for balancing accesses to memory having different memory types
US9342322B2 (en) 2011-09-12 2016-05-17 Microsoft Technology Licensing, Llc System and method for layering using tile-based renderers
US9641826B1 (en) 2011-10-06 2017-05-02 Evans & Sutherland Computer Corporation System and method for displaying distant 3-D stereo on a dome surface
US20130106887A1 (en) * 2011-10-31 2013-05-02 Christopher Tremblay Texture generation using a transformation matrix
CN103108197A (en) 2011-11-14 2013-05-15 辉达公司 Priority level compression method and priority level compression system for three-dimensional (3D) video wireless display
US9633458B2 (en) * 2012-01-23 2017-04-25 Nvidia Corporation Method and system for reducing a polygon bounding box
US9829715B2 (en) 2012-01-23 2017-11-28 Nvidia Corporation Eyewear device for transmitting signal and communication method thereof
US9087409B2 (en) 2012-03-01 2015-07-21 Qualcomm Incorporated Techniques for reducing memory access bandwidth in a graphics processing system based on destination alpha values
US20130235154A1 (en) * 2012-03-09 2013-09-12 Guy Salton-Morgenstern Method and apparatus to minimize computations in real time photo realistic rendering
US8959494B2 (en) * 2012-03-20 2015-02-17 Massively Parallel Technologies Inc. Parallelism from functional decomposition
US9411595B2 (en) 2012-05-31 2016-08-09 Nvidia Corporation Multi-threaded transactional memory coherence
US9148699B2 (en) * 2012-06-01 2015-09-29 Texas Instruments Incorporated Optimized algorithm for construction of composite video from a set of discrete video sources
US9251555B2 (en) 2012-06-08 2016-02-02 2236008 Ontario, Inc. Tiled viewport composition
JP2014006674A (en) * 2012-06-22 2014-01-16 Canon Inc Image processing device, control method of the same and program
US20140010479A1 (en) * 2012-07-09 2014-01-09 Samsung Electro-Mechanics Co., Ltd. Bilinear interpolation circuit for image and method thereof
US9105250B2 (en) * 2012-08-03 2015-08-11 Nvidia Corporation Coverage compaction
US9323315B2 (en) 2012-08-15 2016-04-26 Nvidia Corporation Method and system for automatic clock-gating of a clock grid at a clock source
US8786889B2 (en) * 2012-08-29 2014-07-22 Eastman Kodak Company Method for computing scale for tag insertion
US8928929B2 (en) * 2012-08-29 2015-01-06 Eastman Kodak Company System for generating tag layouts
US9578224B2 (en) 2012-09-10 2017-02-21 Nvidia Corporation System and method for enhanced monoimaging
US8850371B2 (en) 2012-09-14 2014-09-30 Nvidia Corporation Enhanced clock gating in retimed modules
US9002125B2 (en) 2012-10-15 2015-04-07 Nvidia Corporation Z-plane compression with z-plane predictors
US8941676B2 (en) 2012-10-26 2015-01-27 Nvidia Corporation On-chip anti-alias resolve in a cache tiling architecture
US9317948B2 (en) 2012-11-16 2016-04-19 Arm Limited Method of and apparatus for processing graphics
GB201223089D0 (en) 2012-12-20 2013-02-06 Imagination Tech Ltd Hidden culling in tile based computer generated graphics
US9824009B2 (en) 2012-12-21 2017-11-21 Nvidia Corporation Information coherency maintenance systems and methods
US9082212B2 (en) * 2012-12-21 2015-07-14 Nvidia Corporation Programmable blending via multiple pixel shader dispatches
US9251554B2 (en) * 2012-12-26 2016-02-02 Analog Devices, Inc. Block-based signal processing
US10102142B2 (en) 2012-12-26 2018-10-16 Nvidia Corporation Virtual address based memory reordering
US9607407B2 (en) 2012-12-31 2017-03-28 Nvidia Corporation Variable-width differential memory compression
US9591309B2 (en) 2012-12-31 2017-03-07 Nvidia Corporation Progressive lossy memory compression
US9317251B2 (en) 2012-12-31 2016-04-19 Nvidia Corporation Efficient correction of normalizer shift amount errors in fused multiply add operations
DE102013201377A1 (en) * 2013-01-29 2014-07-31 Bayerische Motoren Werke Aktiengesellschaft Method and apparatus for processing 3d image data
US20140225902A1 (en) * 2013-02-11 2014-08-14 Nvidia Corporation Image pyramid processor and method of multi-resolution image processing
KR101529942B1 (en) * 2013-02-18 2015-06-18 서경대학교 산학협력단 Parallel processing rasterizer and parallel processing method for rasterizing
US9992021B1 (en) 2013-03-14 2018-06-05 GoTenna, Inc. System and method for private and point-to-point communication between computing devices
GB2511817A (en) 2013-03-14 2014-09-17 Imagination Tech Ltd Rendering in computer graphics systems
WO2014152800A1 (en) 2013-03-14 2014-09-25 Massively Parallel Technologies, Inc. Project planning and debugging from functional decomposition
US10169906B2 (en) 2013-03-29 2019-01-01 Advanced Micro Devices, Inc. Hybrid render with deferred primitive batch binning
GB2506706B (en) 2013-04-02 2014-09-03 Imagination Tech Ltd Tile-based graphics
US10008029B2 (en) 2013-05-31 2018-06-26 Nvidia Corporation Updating depth related graphics data
US9710894B2 (en) 2013-06-04 2017-07-18 Nvidia Corporation System and method for enhanced multi-sample anti-aliasing
US10204391B2 (en) 2013-06-04 2019-02-12 Arm Limited Method of and apparatus for processing graphics
KR20140142863A (en) * 2013-06-05 2014-12-15 한국전자통신연구원 Apparatus and method for providing graphic editors
KR101451966B1 (en) * 2013-06-17 2014-10-22 (주)가비아 System and method for providing mobile movie rendering
US9418400B2 (en) 2013-06-18 2016-08-16 Nvidia Corporation Method and system for rendering simulated depth-of-field visual effect
US9177413B2 (en) * 2013-06-26 2015-11-03 Nvidia Corporation Unique primitive identifier generation
US9607574B2 (en) 2013-08-09 2017-03-28 Apple Inc. Video data compression format
US9569385B2 (en) 2013-09-09 2017-02-14 Nvidia Corporation Memory transaction ordering
US9230362B2 (en) 2013-09-11 2016-01-05 Nvidia Corporation System, method, and computer program product for using compression with programmable sample locations
US9230363B2 (en) 2013-09-11 2016-01-05 Nvidia Corporation System, method, and computer program product for using compression with programmable sample locations
US9437040B2 (en) * 2013-11-15 2016-09-06 Nvidia Corporation System, method, and computer program product for implementing anti-aliasing operations using a programmable sample pattern table
US10935788B2 (en) 2014-01-24 2021-03-02 Nvidia Corporation Hybrid virtual 3D rendering approach to stereovision
US9276610B2 (en) * 2014-01-27 2016-03-01 Tensorcom, Inc. Method and apparatus of a fully-pipelined layered LDPC decoder
US20150228106A1 (en) * 2014-02-13 2015-08-13 Vixs Systems Inc. Low latency video texture mapping via tight integration of codec engine with 3d graphics engine
US9710957B2 (en) * 2014-04-05 2017-07-18 Sony Interactive Entertainment America Llc Graphics processing enhancement by tracking object and/or primitive identifiers
CN105100862B (en) * 2014-04-18 2018-04-24 阿里巴巴集团控股有限公司 The display processing method and its system of Grid Mobile
GB2526598B (en) 2014-05-29 2018-11-28 Imagination Tech Ltd Allocation of primitives to primitive blocks
US9547918B2 (en) * 2014-05-30 2017-01-17 Intel Corporation Techniques for deferred decoupled shading
GB2524120B (en) * 2014-06-17 2016-03-02 Imagination Tech Ltd Assigning primitives to tiles in a graphics processing system
GB2524121B (en) * 2014-06-17 2016-03-02 Imagination Tech Ltd Assigning primitives to tiles in a graphics processing system
US9307249B2 (en) * 2014-06-20 2016-04-05 Freescale Semiconductor, Inc. Processing device and method of compressing images
US9721376B2 (en) * 2014-06-27 2017-08-01 Samsung Electronics Co., Ltd. Elimination of minimal use threads via quad merging
CN104217461B (en) * 2014-07-10 2017-05-10 无锡梵天信息技术股份有限公司 A parallax mapping method based on a depth map to simulate a real-time bump effect
US9832388B2 (en) 2014-08-04 2017-11-28 Nvidia Corporation Deinterleaving interleaved high dynamic range image by using YUV interpolation
US9569862B2 (en) * 2014-08-15 2017-02-14 Qualcomm Incorporated Bandwidth reduction using texture lookup by adaptive shading
US9665370B2 (en) * 2014-08-19 2017-05-30 Qualcomm Incorporated Skipping of data storage
US10019834B2 (en) 2014-09-26 2018-07-10 Microsoft Technology Licensing, Llc Real-time rendering of volumetric models with occlusive and emissive particles
KR102281180B1 (en) 2014-11-21 2021-07-23 삼성전자주식회사 Image processing apparatus and method
US9720769B2 (en) * 2014-12-03 2017-08-01 Sandisk Technologies Llc Storage parameters for a data storage device
US10249079B2 (en) * 2014-12-11 2019-04-02 Intel Corporation Relaxed sorting in a position-only pipeline
US10026204B2 (en) 2015-01-27 2018-07-17 Splunk Inc. Efficient point-in-polygon indexing technique for processing queries over geographic data sets
US9836874B2 (en) 2015-01-27 2017-12-05 Splunk Inc. Efficient polygon-clipping technique to reduce data transfer requirements for a viewport
US9607414B2 (en) 2015-01-27 2017-03-28 Splunk Inc. Three-dimensional point-in-polygon operation to facilitate displaying three-dimensional structures
US9916326B2 (en) 2015-01-27 2018-03-13 Splunk, Inc. Efficient point-in-polygon indexing technique for facilitating geofencing operations
US9530237B2 (en) * 2015-04-02 2016-12-27 Apple Inc. Interpolation circuitry and techniques for graphics processing
US10255651B2 (en) 2015-04-15 2019-04-09 Channel One Holdings Inc. Methods and systems for generating shaders to emulate a fixed-function graphics pipeline
US9922449B2 (en) 2015-06-01 2018-03-20 Intel Corporation Apparatus and method for dynamic polygon or primitive sorting for improved culling
US9959665B2 (en) 2015-07-21 2018-05-01 Qualcomm Incorporated Zero pixel culling for graphics processing
KR20170034727A (en) 2015-09-21 2017-03-29 삼성전자주식회사 Shadow information storing method and apparatus, 3d rendering method and apparatus
US10269154B2 (en) * 2015-12-21 2019-04-23 Intel Corporation Rasterization based on partial spans
KR102521654B1 (en) * 2016-01-25 2023-04-13 삼성전자주식회사 Computing system and method for performing graphics pipeline of tile-based rendering thereof
US9818051B2 (en) * 2016-01-29 2017-11-14 Ricoh Company, Ltd. Rotation and clipping mechanism
US9906981B2 (en) 2016-02-25 2018-02-27 Nvidia Corporation Method and system for dynamic regulation and control of Wi-Fi scans
CN107180441B (en) * 2016-03-10 2019-04-09 腾讯科技(深圳)有限公司 The method and apparatus for generating eye image
US11847040B2 (en) 2016-03-16 2023-12-19 Asg Technologies Group, Inc. Systems and methods for detecting data alteration from source to target
US10332290B2 (en) * 2016-03-21 2019-06-25 Adobe Inc. Fast, coverage-optimized, resolution-independent and anti-aliased graphics processing
KR101821124B1 (en) 2016-04-05 2018-01-23 한화테크윈 주식회사 Method and apparatus for playing media stream on web-browser
US10412130B2 (en) 2016-04-04 2019-09-10 Hanwha Techwin Co., Ltd. Method and apparatus for playing media stream on web browser
US9798672B1 (en) 2016-04-14 2017-10-24 Macom Connectivity Solutions, Llc Data managment for cache memory
EP3249612B1 (en) * 2016-04-29 2023-02-08 Imagination Technologies Limited Generation of a control stream for a tile
GB2553744B (en) 2016-04-29 2018-09-05 Advanced Risc Mach Ltd Graphics processing systems
JP7100624B2 (en) * 2016-08-29 2022-07-13 アドバンスト・マイクロ・ディバイシズ・インコーポレイテッド Hybrid rendering with binning and sorting of preferred primitive batches
US10756785B2 (en) * 2016-09-29 2020-08-25 Nokia Technologies Oy Flexible reference signal design
US10417134B2 (en) * 2016-11-10 2019-09-17 Oracle International Corporation Cache memory architecture and policies for accelerating graph algorithms
US10282889B2 (en) * 2016-11-29 2019-05-07 Samsung Electronics Co., Ltd. Vertex attribute compression and decompression in hardware
KR20180070314A (en) 2016-12-16 2018-06-26 삼성전자주식회사 Graphics processing apparatus and method for processing graphics pipeline thereof
KR102637736B1 (en) * 2017-01-04 2024-02-19 삼성전자주식회사 Graphics processing method and system
CA3058421A1 (en) 2017-03-30 2018-10-04 Magic Leap, Inc. Centralized rendering
US10977858B2 (en) 2017-03-30 2021-04-13 Magic Leap, Inc. Centralized rendering
US10157493B2 (en) * 2017-04-01 2018-12-18 Intel Corporation Adaptive multisampling based on vertex attributes
GB2562041B (en) * 2017-04-28 2020-11-25 Imagination Tech Ltd Multi-output decoder for texture decompression
US10521877B2 (en) 2017-05-23 2019-12-31 Samsung Electronics Co., Ltd Apparatus and method for speculative buffer reservations with cancellation mechanism
US10510181B2 (en) * 2017-06-27 2019-12-17 Samsung Electronics Co., Ltd. System and method for cache management using a cache status table
US10969740B2 (en) 2017-06-27 2021-04-06 Nvidia Corporation System and method for near-eye light field rendering for wide field of view interactive three-dimensional computer graphics
CN107463398B (en) 2017-07-21 2018-08-17 腾讯科技(深圳)有限公司 Game rendering intent, device, storage device and terminal
GB2569844B (en) 2017-10-20 2021-01-06 Graphcore Ltd Sending data off-chip
GB2569271B (en) 2017-10-20 2020-05-13 Graphcore Ltd Synchronization with a host processor
GB2569775B (en) 2017-10-20 2020-02-26 Graphcore Ltd Synchronization in a multi-tile, multi-chip processing arrangement
US10600142B2 (en) * 2017-11-02 2020-03-24 Advanced Micro Devices, Inc. Compression and decompression of indices in a graphics pipeline
US11057500B2 (en) 2017-11-20 2021-07-06 Asg Technologies Group, Inc. Publication of applications using server-side virtual screen change capture
US10699374B2 (en) 2017-12-05 2020-06-30 Microsoft Technology Licensing, Llc Lens contribution-based virtual reality display rendering
GB2569546B (en) * 2017-12-19 2020-10-14 Sony Interactive Entertainment Inc Determining pixel values using reference images
US10877740B2 (en) 2017-12-29 2020-12-29 Asg Technologies Group, Inc. Dynamically deploying a component in an application
US10812611B2 (en) 2017-12-29 2020-10-20 Asg Technologies Group, Inc. Platform-independent application publishing to a personalized front-end interface by encapsulating published content into a container
US11611633B2 (en) 2017-12-29 2023-03-21 Asg Technologies Group, Inc. Systems and methods for platform-independent application publishing to a front-end interface
GB2572617B (en) * 2018-04-05 2021-06-16 Imagination Tech Ltd Blending hardware
US10672182B2 (en) * 2018-04-19 2020-06-02 Microsoft Technology Licensing, Llc Compact visibility state for GPUs compatible with hardware instancing
JP7119081B2 (en) 2018-05-24 2022-08-16 株式会社Preferred Networks Projection data generation device, three-dimensional model, projection data generation method, neural network generation method and program
GB2575294B8 (en) 2018-07-04 2022-07-20 Graphcore Ltd Host Proxy On Gateway
US10861230B2 (en) * 2018-08-01 2020-12-08 Nvidia Corporation System-generated stable barycentric coordinates and direct plane equation access
CN113170114B (en) * 2018-09-13 2024-03-29 弗劳恩霍夫应用研究促进协会 Affine linear weighted intra prediction
US11138747B1 (en) * 2018-11-02 2021-10-05 Facebook Technologies, Llc Interpolation optimizations for a display engine for post-rendering processing
GB2579412B (en) 2018-11-30 2020-12-23 Graphcore Ltd Gateway pull model
US10909659B2 (en) 2018-12-12 2021-02-02 Apical Limited Super-resolution image processing using a machine learning system
US11715262B2 (en) * 2018-12-17 2023-08-01 Advanced Micro Devices, Inc. Optimizing primitive shaders
KR102216749B1 (en) * 2019-03-05 2021-02-17 네이버웹툰 유한회사 Method, apparatus and computer program for coloring of a target image
US10866280B2 (en) 2019-04-01 2020-12-15 Texas Instruments Incorporated Scan chain self-testing of lockstep cores on reset
US11640649B2 (en) * 2019-06-19 2023-05-02 Samsung Electronics Co., Ltd. Methods and apparatus for efficient range calculation
US11488349B2 (en) 2019-06-28 2022-11-01 Ati Technologies Ulc Method and apparatus for alpha blending images from different color formats
US11762634B2 (en) 2019-06-28 2023-09-19 Asg Technologies Group, Inc. Systems and methods for seamlessly integrating multiple products by using a common visual modeler
US10981059B2 (en) * 2019-07-03 2021-04-20 Sony Interactive Entertainment LLC Asset aware computing architecture for graphics processing
EP4004766A4 (en) 2019-07-30 2023-09-20 Falkonry, Inc. Fluid and resolution-friendly view of large volumes of time series data
US11055067B2 (en) 2019-10-18 2021-07-06 Asg Technologies Group, Inc. Unified digital automation platform
US11269660B2 (en) 2019-10-18 2022-03-08 Asg Technologies Group, Inc. Methods and systems for integrated development environment editor support with a single code base
US11886397B2 (en) 2019-10-18 2024-01-30 Asg Technologies Group, Inc. Multi-faceted trust system
US11941137B2 (en) 2019-10-18 2024-03-26 Asg Technologies Group, Inc. Use of multi-faceted trust scores for decision making, action triggering, and data analysis and interpretation
US11755760B2 (en) 2019-10-18 2023-09-12 Asg Technologies Group, Inc. Systems and methods for secure policies-based information governance
US11210821B2 (en) * 2019-11-27 2021-12-28 Arm Limited Graphics processing systems
US11170555B2 (en) 2019-11-27 2021-11-09 Arm Limited Graphics processing systems
US11216993B2 (en) * 2019-11-27 2022-01-04 Arm Limited Graphics processing systems
US11210847B2 (en) 2019-11-27 2021-12-28 Arm Limited Graphics processing systems
US11508110B2 (en) 2020-02-03 2022-11-22 Sony Interactive Entertainment Inc. System and method for efficient multi-GPU rendering of geometry by performing geometry analysis before rendering
US11514549B2 (en) * 2020-02-03 2022-11-29 Sony Interactive Entertainment Inc. System and method for efficient multi-GPU rendering of geometry by generating information in one rendering phase for use in another rendering phase
US11113858B2 (en) * 2020-02-04 2021-09-07 Inventive Software, LLC System and method for deep compositing of images in web browsers
US11321259B2 (en) * 2020-02-14 2022-05-03 Sony Interactive Entertainment Inc. Network architecture providing high speed storage access through a PCI express fabric between a compute node and a storage server
US11132831B1 (en) 2020-03-02 2021-09-28 Qualcomm Incorporated Methods and apparatus for efficient multi-view rasterization
US11243882B2 (en) * 2020-04-15 2022-02-08 International Business Machines Corporation In-array linked list identifier pool scheme
US11250627B2 (en) * 2020-06-29 2022-02-15 Intel Corporation Tile sequencing mechanism
WO2022040574A1 (en) 2020-08-21 2022-02-24 Beam, Inc. Integrating overlaid digital content into displayed data via graphics processing circuitry
WO2022081476A1 (en) 2020-10-13 2022-04-21 ASG Technologies Group, Inc. dba ASG Technologies Geolocation-based policy rules
CN116670723A (en) * 2020-10-22 2023-08-29 彩滋公司 System and method for high quality rendering of composite views of customized products
US11232628B1 (en) * 2020-11-10 2022-01-25 Weta Digital Limited Method for processing image data to provide for soft shadow effects using shadow depth information
US11481933B1 (en) 2021-04-08 2022-10-25 Mobeus Industries, Inc. Determining a change in position of displayed digital content in subsequent frames via graphics processing circuitry
US11586835B2 (en) 2021-04-30 2023-02-21 Mobeus Industries, Inc. Integrating overlaid textual digital content into displayed data via graphics processing circuitry using a frame buffer
US11601276B2 (en) * 2021-04-30 2023-03-07 Mobeus Industries, Inc. Integrating and detecting visual data security token in displayed data via graphics processing circuitry using a frame buffer
US11682101B2 (en) 2021-04-30 2023-06-20 Mobeus Industries, Inc. Overlaying displayed digital content transmitted over a communication network via graphics processing circuitry using a frame buffer
US11477020B1 (en) 2021-04-30 2022-10-18 Mobeus Industries, Inc. Generating a secure random number by determining a change in parameters of digital content in subsequent frames via graphics processing circuitry
US11483156B1 (en) 2021-04-30 2022-10-25 Mobeus Industries, Inc. Integrating digital content into displayed data on an application layer via processing circuitry of a server
US11475610B1 (en) 2021-04-30 2022-10-18 Mobeus Industries, Inc. Controlling interactivity of digital content overlaid onto displayed data via graphics processing circuitry using a frame buffer
CN113256485B (en) * 2021-05-21 2024-01-30 百果园技术(新加坡)有限公司 Image stretching method, device, electronic equipment and storage medium
US20220410002A1 (en) * 2021-06-29 2022-12-29 Bidstack Group PLC Mesh processing for viewability testing
US11562153B1 (en) 2021-07-16 2023-01-24 Mobeus Industries, Inc. Systems and methods for recognizability of objects in a multi-layer display
US20230334736A1 (en) * 2022-04-15 2023-10-19 Meta Platforms Technologies, Llc Rasterization Optimization for Analytic Anti-Aliasing
US11882295B2 (en) 2022-04-15 2024-01-23 Meta Platforms Technologies, Llc Low-power high throughput hardware decoder with random block access
US20230334728A1 (en) * 2022-04-15 2023-10-19 Meta Platforms Technologies, Llc Destination Update for Blending Modes in a Graphics Pipeline
CN114529705B (en) * 2022-04-22 2022-07-19 山东捷瑞数字科技股份有限公司 Interface layout processing method of three-dimensional engine editor

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4970636A (en) * 1989-01-23 1990-11-13 Honeywell Inc. Memory interface controller
US5083287A (en) * 1988-07-14 1992-01-21 Daikin Industries, Inc. Method and apparatus for applying a shadowing operation to figures to be drawn for displaying on crt-display
US5402532A (en) * 1991-03-12 1995-03-28 International Business Machines Corporation Direct display of CSG expression by use of depth buffers
US5828378A (en) * 1995-06-01 1998-10-27 Ricoh Company, Ltd. Three dimensional graphics processing apparatus processing ordinary and special objects

Family Cites Families (128)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2353185A1 (en) 1976-04-09 1977-12-23 Thomson Csf RAPID CORRELATOR DEVICE, AND SYSTEM FOR PROCESSING THE SIGNALS OF A RECEIVER INCLUDING SUCH A DEVICE
FR2481489A1 (en) 1980-04-25 1981-10-30 Thomson Csf BIDIMENSIONAL CORRELATOR DEVICE
US4484346A (en) 1980-08-15 1984-11-20 Sternberg Stanley R Neighborhood transformation logic circuitry for an image analyzer system
US4559618A (en) 1982-09-13 1985-12-17 Data General Corp. Content-addressable memory module with associative clear
US4783829A (en) 1983-02-23 1988-11-08 Hitachi, Ltd. Pattern recognition apparatus
US4581760A (en) 1983-04-27 1986-04-08 Fingermatrix, Inc. Fingerprint verification method
US4670858A (en) 1983-06-07 1987-06-02 Tektronix, Inc. High storage capacity associative memory
US4594673A (en) 1983-06-28 1986-06-10 Gti Corporation Hidden surface processor
US4532606A (en) 1983-07-14 1985-07-30 Burroughs Corporation Content addressable memory cell with shift capability
US4564952A (en) 1983-12-08 1986-01-14 At&T Bell Laboratories Compensation of filter symbol interference by adaptive estimation of received symbol sequences
US4694404A (en) 1984-01-12 1987-09-15 Key Bank N.A. High-speed image generation of complex solid objects using octree encoding
US4794559A (en) 1984-07-05 1988-12-27 American Telephone And Telegraph Company, At&T Bell Laboratories Content addressable semiconductor memory arrays
US4622653A (en) 1984-10-29 1986-11-11 Texas Instruments Incorporated Block associative memory
US4669054A (en) 1985-05-03 1987-05-26 General Dynamics, Pomona Division Device and method for optically correlating a pair of images
SE445154B (en) 1985-07-08 1986-06-02 Ibm Svenska Ab METHOD OF REMOVING HIDDEN LINES
US4695973A (en) 1985-10-22 1987-09-22 The United States Of America As Represented By The Secretary Of The Air Force Real-time programmable optical correlator
US4758982A (en) 1986-01-08 1988-07-19 Advanced Micro Devices, Inc. Quasi content addressable memory
US4890242A (en) 1986-06-05 1989-12-26 Xox Corporation Solid-modeling system using topology directed subdivision for determination of surface intersections
US5067162A (en) 1986-06-30 1991-11-19 Identix Incorporated Method and apparatus for verifying identity using image correlation
US4998286A (en) 1987-02-13 1991-03-05 Olympus Optical Co., Ltd. Correlation operational apparatus for multi-dimensional images
US4825391A (en) 1987-07-20 1989-04-25 General Electric Company Depth buffer priority processing for real time computer image generating systems
US5146592A (en) 1987-09-14 1992-09-08 Visual Information Technologies, Inc. High speed image processing computer with overlapping windows-div
US5129060A (en) 1987-09-14 1992-07-07 Visual Information Technologies, Inc. High speed image processing computer
US4841467A (en) 1987-10-05 1989-06-20 General Electric Company Architecture to implement floating point multiply/accumulate operations
GB2215623B (en) 1987-10-23 1991-07-31 Rotation Limited Apparatus for playing a game for one or more players and to games played with the apparatus
US4945500A (en) 1987-11-04 1990-07-31 Schlumberger Technologies, Inc. Triangle processor for 3-D graphics display system
US4888712A (en) 1987-11-04 1989-12-19 Schlumberger Systems, Inc. Guardband clipping method and apparatus for 3-D graphics display system
FR2625345A1 (en) 1987-12-24 1989-06-30 Thomson Cgr THREE-DIMENSIONAL VIEWING METHOD OF NUMERICALLY ENCODED OBJECTS IN TREE FORM AND DEVICE FOR IMPLEMENTING THE SAME
DE68918724T2 (en) 1988-02-17 1995-05-24 Nippon Denso Co Fingerprint verification process using multiple correlation decision levels and successive decision levels.
US4888583A (en) 1988-03-14 1989-12-19 Ligocki Terry J Method and apparatus for rendering an image from data arranged in a constructive solid geometry format
US5133052A (en) 1988-08-04 1992-07-21 Xerox Corporation Interactive graphical search and replace utility for computer-resident synthetic graphic image editors
US4996666A (en) 1988-08-12 1991-02-26 Duluk Jr Jerome F Content-addressable memory system capable of fully parallel magnitude comparisons
GB8828342D0 (en) 1988-12-05 1989-01-05 Rediffusion Simulation Ltd Image generator
FR2646046B1 (en) 1989-04-18 1995-08-25 France Etat METHOD AND DEVICE FOR COMPRESSING IMAGE DATA BY MATHEMATICAL TRANSFORMATION WITH REDUCED COST OF IMPLEMENTATION, IN PARTICULAR FOR TRANSMISSION AT REDUCED THROUGHPUT OF IMAGE SEQUENCES
JPH0776991B2 (en) 1989-10-24 1995-08-16 インターナショナル・ビジネス・マシーンズ・コーポレーション NURBS data conversion method and apparatus
US5245700A (en) 1989-11-21 1993-09-14 International Business Machines Corporation Adjustment of z-buffer values for lines on the surface of a polygon
JPH03166601A (en) 1989-11-27 1991-07-18 Hitachi Ltd Symbolizing device and process controller and control supporting device using the symbolizing device
US5129051A (en) 1990-03-16 1992-07-07 Hewlett-Packard Company Decomposition of arbitrary polygons into trapezoids
US5123085A (en) 1990-03-19 1992-06-16 Sun Microsystems, Inc. Method and apparatus for rendering anti-aliased polygons
US5128888A (en) 1990-04-02 1992-07-07 Advanced Micro Devices, Inc. Arithmetic unit having multiple accumulators
GB9009127D0 (en) 1990-04-24 1990-06-20 Rediffusion Simulation Ltd Image generator
US5369734A (en) 1990-05-18 1994-11-29 Kabushiki Kaisha Toshiba Method for processing and displaying hidden-line graphic images
DE69122557T2 (en) 1990-06-29 1997-04-24 Philips Electronics Nv Imaging
JPH0475183A (en) 1990-07-17 1992-03-10 Mitsubishi Electric Corp Correlativity detector for image
US5054090A (en) 1990-07-20 1991-10-01 Knight Arnold W Fingerprint correlation system with parallel FIFO processor
US5050220A (en) 1990-07-24 1991-09-17 The United States Of America As Represented By The Secretary Of The Navy Optical fingerprint correlator
JPH07120435B2 (en) 1990-12-06 1995-12-20 インターナショナル・ビジネス・マシーンズ・コーポレイション Method and system for initializing and updating high-speed Z buffer
FR2670923A1 (en) 1990-12-21 1992-06-26 Philips Lab Electronique CORRELATION DEVICE.
US5289567A (en) 1991-04-01 1994-02-22 Digital Equipment Corporation Computer apparatus and method for finite element identification in interactive modeling
US5293467A (en) 1991-04-03 1994-03-08 Buchner Gregory C Method for resolving priority between a calligraphically-displayed point feature and both raster-displayed faces and other calligraphically-displayed point features in a CIG system
US5315537A (en) 1991-04-08 1994-05-24 Blacker Teddy D Automated quadrilateral surface discretization method and apparatus usable to generate mesh in a finite element analysis system
US5263136A (en) 1991-04-30 1993-11-16 Optigraphics Corporation System for managing tiled images using multiple resolutions
US5347619A (en) 1991-04-30 1994-09-13 International Business Machines Corporation Nonconvex polygon identifier
US5299139A (en) 1991-06-21 1994-03-29 Cadence Design Systems, Inc. Short locator method
US5493644A (en) 1991-07-11 1996-02-20 Hewlett-Packard Company Polygon span interpolator with main memory Z buffer
US5295235A (en) 1992-02-14 1994-03-15 Steve Newman Polygon engine for updating computer graphic display employing compressed bit map data
US5319743A (en) 1992-04-02 1994-06-07 Digital Equipment Corporation Intelligent and compact bucketing method for region queries in two-dimensional space
WO1993023816A1 (en) 1992-05-18 1993-11-25 Silicon Engines Inc. System and method for cross correlation with application to video motion vector estimation
US5669010A (en) 1992-05-18 1997-09-16 Silicon Engines Cascaded two-stage computational SIMD engine having multi-port memory and multiple arithmetic units
US5621866A (en) 1992-07-24 1997-04-15 Fujitsu Limited Image processing apparatus having improved frame buffer with Z buffer and SAM port
US5455900A (en) 1992-10-20 1995-10-03 Ricoh Company, Ltd. Image processing apparatus
US5388206A (en) 1992-11-13 1995-02-07 The University Of North Carolina Architecture and apparatus for image generation
TW241196B (en) 1993-01-15 1995-02-21 Du Pont
JP3240447B2 (en) 1993-02-19 2001-12-17 株式会社リコー Image processing device
US5574835A (en) 1993-04-06 1996-11-12 Silicon Engines, Inc. Bounding box and projections detection of hidden polygons in three-dimensional spatial databases
US5509110A (en) 1993-04-26 1996-04-16 Loral Aerospace Corporation Method for tree-structured hierarchical occlusion in image generators
US6167143A (en) * 1993-05-03 2000-12-26 U.S. Philips Corporation Monitoring system
US5684939A (en) 1993-07-09 1997-11-04 Silicon Graphics, Inc. Antialiased imaging with improved pixel supersampling
US5579455A (en) 1993-07-30 1996-11-26 Apple Computer, Inc. Rendering of 3D scenes on a display using hierarchical z-buffer visibility
GB9316214D0 (en) * 1993-08-05 1993-09-22 Philips Electronics Uk Ltd Image processing
JPH07182537A (en) 1993-12-21 1995-07-21 Toshiba Corp Device and method for plotting graphic
US5699497A (en) 1994-02-17 1997-12-16 Evans & Sutherland Computer Corporation Rendering global macro texture, for producing a dynamic image, as on computer generated terrain, seen from a moving viewpoint
US5778245A (en) 1994-03-01 1998-07-07 Intel Corporation Method and apparatus for dynamic allocation of multiple buffers in a processor
US5623628A (en) 1994-03-02 1997-04-22 Intel Corporation Computer system and method for maintaining memory consistency in a pipelined, non-blocking caching bus request queue
US5546194A (en) * 1994-03-23 1996-08-13 Videofaxx, Inc. Method and apparatus for converting a video image format to a group III fax format
US5596686A (en) * 1994-04-21 1997-01-21 Silicon Engines, Inc. Method and apparatus for simultaneous parallel query graphics rendering Z-coordinate buffer
US5544306A (en) 1994-05-03 1996-08-06 Sun Microsystems, Inc. Flexible dram access in a frame buffer memory and system
JPH0855239A (en) * 1994-07-21 1996-02-27 Internatl Business Mach Corp <Ibm> Method and apparatus for judgment of visibility of graphicalobject
US5572634A (en) 1994-10-26 1996-11-05 Silicon Engines, Inc. Method and apparatus for spatial simulation acceleration
JPH08127167A (en) * 1994-11-01 1996-05-21 Arutetsuku Kk Apparatus and method for detecting end of rolled sheet
US5594854A (en) 1995-03-24 1997-01-14 3Dlabs Inc. Ltd. Graphics subsystem with coarse subpixel correction
US5798770A (en) 1995-03-24 1998-08-25 3Dlabs Inc. Ltd. Graphics rendering system with reconfigurable pipeline sequence
US5710876A (en) 1995-05-25 1998-01-20 Silicon Graphics, Inc. Computer graphics system for rendering images using full spectral illumination data
WO1997005576A1 (en) * 1995-07-26 1997-02-13 Silicon Engines, Inc. Method and apparatus for span and subspan sorting rendering system
US5841447A (en) 1995-08-02 1998-11-24 Evans & Sutherland Computer Corporation System and method for improving pixel update performance
AU6766096A (en) * 1995-08-04 1997-03-05 Microsoft Corporation Method and system for rendering graphical objects to image chunks and combining image layers into a display image
US5977977A (en) 1995-08-04 1999-11-02 Microsoft Corporation Method and system for multi-pass rendering
US5864342A (en) 1995-08-04 1999-01-26 Microsoft Corporation Method and system for rendering graphical objects to image chunks
US5949428A (en) 1995-08-04 1999-09-07 Microsoft Corporation Method and apparatus for resolving pixel data in a graphics rendering system
US5990904A (en) 1995-08-04 1999-11-23 Microsoft Corporation Method and system for merging pixel fragments in a graphics rendering system
US5767859A (en) 1995-09-28 1998-06-16 Hewlett-Packard Company Method and apparatus for clipping non-planar polygons
US5854631A (en) 1995-11-22 1998-12-29 Silicon Graphics, Inc. System and method for merging pixel fragments based on depth range values
JP2882465B2 (en) * 1995-12-25 1999-04-12 日本電気株式会社 Image generation method and apparatus
US5574836A (en) 1996-01-22 1996-11-12 Broemmelsiek; Raymond M. Interactive display apparatus and method with viewer position compensation
US5850225A (en) 1996-01-24 1998-12-15 Evans & Sutherland Computer Corp. Image mapping system and process using panel shear transforms
US6046746A (en) * 1996-07-01 2000-04-04 Sun Microsystems, Inc. Method and apparatus implementing high resolution rendition of Z-buffered primitives
US5751291A (en) * 1996-07-26 1998-05-12 Hewlett-Packard Company System and method for accelerated occlusion culling
US5767589A (en) 1996-09-03 1998-06-16 Maximum Products Inc. Lighting control circuit for vehicle brake light/tail light/indicator light assembly
US5860158A (en) 1996-11-15 1999-01-12 Samsung Electronics Company, Ltd. Cache control unit with a cache request transaction-oriented protocol
US6167486A (en) 1996-11-18 2000-12-26 Nec Electronics, Inc. Parallel access virtual channel memory system with cacheable channels
US5936629A (en) 1996-11-20 1999-08-10 International Business Machines Corporation Accelerated single source 3D lighting mechanism
US6111582A (en) * 1996-12-20 2000-08-29 Jenkins; Barry L. System and method of image generation and encoding using primitive reprojection
US6697063B1 (en) * 1997-01-03 2004-02-24 Nvidia U.S. Investment Company Rendering pipeline
US5852451A (en) 1997-01-09 1998-12-22 S3 Incorporation Pixel reordering for improved texture mapping
US5949424A (en) * 1997-02-28 1999-09-07 Silicon Graphics, Inc. Method, system, and computer program product for bump mapping in tangent space
US5880736A (en) 1997-02-28 1999-03-09 Silicon Graphics, Inc. Method system and computer program product for shading
US6259452B1 (en) 1997-04-14 2001-07-10 Massachusetts Institute Of Technology Image drawing system and method with real-time occlusion culling
US6084591A (en) * 1997-04-29 2000-07-04 Ati Technologies, Inc. Method and apparatus for deferred video rendering
US5889997A (en) 1997-05-30 1999-03-30 Hewlett-Packard Company Assembler system and method for a geometry accelerator
US5920326A (en) 1997-05-30 1999-07-06 Hewlett Packard Company Caching and coherency control of multiple geometry accelerators in a computer graphics system
US6002412A (en) 1997-05-30 1999-12-14 Hewlett-Packard Co. Increased performance of graphics memory using page sorting fifos
US6118452A (en) 1997-08-05 2000-09-12 Hewlett-Packard Company Fragment visibility pretest system and methodology for improved performance of a graphics system
US6002410A (en) 1997-08-25 1999-12-14 Chromatic Research, Inc. Reconfigurable texture cache
US6128000A (en) 1997-10-15 2000-10-03 Compaq Computer Corporation Full-scene antialiasing using improved supersampling techniques
US6204859B1 (en) 1997-10-15 2001-03-20 Digital Equipment Corporation Method and apparatus for compositing colors of images with memory constraints for storing pixel data
JPH11161819A (en) * 1997-11-27 1999-06-18 Sega Enterp Ltd Image processor, its method and recording medium recording image processing program
US6201540B1 (en) * 1998-01-07 2001-03-13 Microsoft Corporation Graphical interface components for in-dash automotive accessories
US6259460B1 (en) 1998-03-26 2001-07-10 Silicon Graphics, Inc. Method for efficient handling of texture cache misses by recirculation
US6246415B1 (en) 1998-04-30 2001-06-12 Silicon Graphics, Inc. Method and apparatus for culling polygons
US6243744B1 (en) * 1998-05-26 2001-06-05 Compaq Computer Corporation Computer network cluster generation indicator
US6650327B1 (en) * 1998-06-16 2003-11-18 Silicon Graphics, Inc. Display system having floating point rasterization and floating point framebuffering
US6216004B1 (en) * 1998-06-23 2001-04-10 Qualcomm Incorporated Cellular communication system with common channel soft handoff and associated method
US6263493B1 (en) * 1998-07-08 2001-07-17 International Business Machines Corporation Method and system for controlling the generation of program statements
US6771264B1 (en) * 1998-08-20 2004-08-03 Apple Computer, Inc. Method and apparatus for performing tangent space lighting and bump mapping in a deferred shading graphics processor
US6577317B1 (en) * 1998-08-20 2003-06-10 Apple Computer, Inc. Apparatus and method for geometry operations in a 3D-graphics pipeline
US6476807B1 (en) * 1998-08-20 2002-11-05 Apple Computer, Inc. Method and apparatus for performing conservative hidden surface removal in a graphics processor with deferred shading
US6275235B1 (en) * 1998-12-21 2001-08-14 Silicon Graphics, Inc. High precision texture wrapping method and device
US6228730B1 (en) * 1999-04-28 2001-05-08 United Microelectronics Corp. Method of fabricating field effect transistor

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5083287A (en) * 1988-07-14 1992-01-21 Daikin Industries, Inc. Method and apparatus for applying a shadowing operation to figures to be drawn for displaying on crt-display
US4970636A (en) * 1989-01-23 1990-11-13 Honeywell Inc. Memory interface controller
US5402532A (en) * 1991-03-12 1995-03-28 International Business Machines Corporation Direct display of CSG expression by use of depth buffers
US5828378A (en) * 1995-06-01 1998-10-27 Ricoh Company, Ltd. Three dimensional graphics processing apparatus processing ordinary and special objects

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7170513B1 (en) 1998-07-22 2007-01-30 Nvidia Corporation System and method for display list occlusion branching
US7023437B1 (en) 1998-07-22 2006-04-04 Nvidia Corporation System and method for accelerating graphics processing using a post-geometry data stream during multiple-pass rendering
US7002588B1 (en) 1999-12-06 2006-02-21 Nvidia Corporation System, method and computer program product for branching during programmable vertex processing
US7209140B1 (en) 1999-12-06 2007-04-24 Nvidia Corporation System, method and article of manufacture for a programmable vertex processing model with instruction set
US6844880B1 (en) 1999-12-06 2005-01-18 Nvidia Corporation System, method and computer program product for an improved programmable vertex processing model with instruction set
US6870540B1 (en) * 1999-12-06 2005-03-22 Nvidia Corporation System, method and computer program product for a programmable pixel processing model with instruction set
US6664963B1 (en) 2000-05-31 2003-12-16 Nvidia Corporation System, method and computer program product for programmable shading using pixel shaders
US6690372B2 (en) 2000-05-31 2004-02-10 Nvidia Corporation System, method and article of manufacture for shadow mapping
US6734861B1 (en) 2000-05-31 2004-05-11 Nvidia Corporation System, method and article of manufacture for an interlock module in a computer graphics processing pipeline
US6532013B1 (en) 2000-05-31 2003-03-11 Nvidia Corporation System, method and article of manufacture for pixel shaders for programmable shading
US7068272B1 (en) 2000-05-31 2006-06-27 Nvidia Corporation System, method and article of manufacture for Z-value and stencil culling prior to rendering in a computer graphics processing pipeline
US6778181B1 (en) 2000-12-07 2004-08-17 Nvidia Corporation Graphics processing system having a virtual texturing array
US7006101B1 (en) 2001-06-08 2006-02-28 Nvidia Corporation Graphics API with branching capabilities
US6982718B2 (en) 2001-06-08 2006-01-03 Nvidia Corporation System, method and computer program product for programmable fragment processing in a graphics pipeline
US7162716B2 (en) 2001-06-08 2007-01-09 Nvidia Corporation Software emulator for optimizing application-programmable vertex processing
US6697064B1 (en) 2001-06-08 2004-02-24 Nvidia Corporation System, method and computer program product for matrix tracking during vertex processing in a graphics pipeline
US7286133B2 (en) 2001-06-08 2007-10-23 Nvidia Corporation System, method and computer program product for programmable fragment processing
US7456838B1 (en) 2001-06-08 2008-11-25 Nvidia Corporation System and method for converting a vertex program to a binary format capable of being executed by a hardware graphics pipeline
US6704025B1 (en) 2001-08-31 2004-03-09 Nvidia Corporation System and method for dual-depth shadow-mapping
EP1450308A1 (en) * 2001-11-27 2004-08-25 Sony Computer Entertainment Inc. Image processing apparatus and constituent parts thereof, rendering method
EP1450308A4 (en) * 2001-11-27 2010-06-09 Sony Computer Entertainment Inc Image processing apparatus and constituent parts thereof, rendering method
US7009615B1 (en) 2001-11-30 2006-03-07 Nvidia Corporation Floating point buffer system and method for use during programmable fragment processing in a graphics pipeline
US7009605B2 (en) 2002-03-20 2006-03-07 Nvidia Corporation System, method and computer program product for generating a shader program
US8106904B2 (en) 2002-03-20 2012-01-31 Nvidia Corporation Shader program generation system and method

Also Published As

Publication number Publication date
EP1138023A1 (en) 2001-10-04
WO2000011614B1 (en) 2000-07-27
AU5690499A (en) 2000-03-14
US6597363B1 (en) 2003-07-22
US7167181B2 (en) 2007-01-23
US6614444B1 (en) 2003-09-02
AU5687599A (en) 2000-03-14
JP2004272928A (en) 2004-09-30
KR20010085424A (en) 2001-09-07
WO2000019377A1 (en) 2000-04-06
WO2000011605A2 (en) 2000-03-02
KR20010085426A (en) 2001-09-07
AU5782599A (en) 2000-03-14
US6717576B1 (en) 2004-04-06
US6771264B1 (en) 2004-08-03
JP2003515798A (en) 2003-05-07
WO2000011614A2 (en) 2000-03-02
WO2000011605B1 (en) 2001-04-12
WO2000011613A3 (en) 2000-06-29
WO2000011605A9 (en) 2001-06-21
WO2000011613A9 (en) 2000-08-03
WO2000011604A3 (en) 2000-06-02
WO2000030040A1 (en) 2000-05-25
AU5687899A (en) 2000-04-17
WO2000011605A3 (en) 2000-06-22
KR100478767B1 (en) 2005-03-24
US20040130552A1 (en) 2004-07-08
EP1138023A4 (en) 2005-09-07
WO2000011604B1 (en) 2000-07-20
EP1105844A1 (en) 2001-06-13
JP2002526842A (en) 2002-08-20
KR100485241B1 (en) 2005-04-27
WO2000011604A2 (en) 2000-03-02
JP3657519B2 (en) 2005-06-08
WO2000011614A3 (en) 2000-06-15
WO2000030040A9 (en) 2001-10-18
AU5576599A (en) 2000-03-14
WO2000019377B1 (en) 2000-06-08
JP4516350B2 (en) 2010-08-04
JP3657518B2 (en) 2005-06-08
AU5779799A (en) 2000-06-05

Similar Documents

Publication Publication Date Title
US6664959B2 (en) Method and apparatus for culling in a graphics processor with deferred shading
WO2000011613A2 (en) Performing hidden surface removal in a graphics processor with deferred shading
US5870097A (en) Method and system for improving shadowing in a graphics rendering system
US5977977A (en) Method and system for multi-pass rendering
US6326964B1 (en) Method for sorting 3D object geometry among image chunks for rendering in a layered graphics rendering system
US5867166A (en) Method and system for generating images using Gsprites
US5949428A (en) Method and apparatus for resolving pixel data in a graphics rendering system
US5990904A (en) Method and system for merging pixel fragments in a graphics rendering system
US6577317B1 (en) Apparatus and method for geometry operations in a 3D-graphics pipeline
EP0850462B1 (en) Method and system for rendering graphical objects to image chunks and combining image layers into a display image
US7030887B2 (en) Methods and systems for transparent depth sorting
WO1997005576A9 (en) Method and apparatus for span and subspan sorting rendering system
WO1997005576A1 (en) Method and apparatus for span and subspan sorting rendering system
EP1434172A2 (en) Method and system for generating a display image using Gsprites.
US5926183A (en) Efficient rendering utilizing user defined rooms and windows

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW SD SL SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
AK Designated states

Kind code of ref document: A3

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A3

Designated state(s): GH GM KE LS MW SD SL SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

AK Designated states

Kind code of ref document: C2

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: C2

Designated state(s): GH GM KE LS MW SD SL SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

COP Corrected version of pamphlet

Free format text: PAGES 1-55, DESCRIPTION, REPLACED BY NEW PAGES 1-55; PAGES 56-61, CLAIMS, REPLACED BY NEW PAGES 56-61; PAGES 1/26-26/26, DRAWINGS, REPLACED BY NEW PAGES 1/26-26/26; DUE TO LATE TRANSMITTAL BY THE RECEIVING OFFICE (WITH AN UPDATED VERSION OF THE PAMPHLET FRONT PAGE)

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase