US 7937179 B2
In one embodiment, a method for predicting yield includes calculating a criticality factor (CF) for each of a plurality of defects detected in an inspection process step of a wafer, and determining a yield-loss contribution of the inspection process step to the final yield based on CFs of the plurality of defects and the yield model built for a relevant design. The yield-loss contribution of the inspection process step is then used to predict the final yield for the wafer.
1. A computer-implemented method comprising:
calculating, using a computer, a criticality factor (CF) for each of a plurality of defects detected in at least one inspection process step of a wafer, the wafer comprising a plurality of dies;
determining a non-defect yield component of a final yield for the wafer using a subset of the plurality of dies that have no visual defects;
determining a yield-loss contribution of the inspection process step to the final yield based on the CFs of the plurality of defects, a yield model built for a relevant design, and the non-defect yield component; and
predicting the final yield for the wafer using the yield-loss contribution of the inspection process step.
2. The computer-implemented method of
3. The computer-implemented method of
distributing defects among a number of inspection steps; and
for each step:
classifying the defects according to similar CF values into one or more CF bins;
creating a second subset of dies from the subset of the plurality of dies, wherein the subset is a set of dies having inspection results in all inspection steps, wherein the second subset is a set of dies having defects only in a specific inspection step;
binning together the dies in the second subset according to a similar CF value to create one or more sub-subsets;
for each sub-subset, calculating a defect-specific fail probability (FPdefects s,b); and
plotting the defect-specific fail probability in a functional form, FPdefects=f(CF), to create a correlation curve.
4. The computer-implemented method of
5. The computer-implemented method of
6. The computer-implemented method of
identifying an open area, wherein the open area is a total area of all locations inside a design layout area to cause an open failure, and identifying a short area, wherein the short area is a total area of all locations inside the design layout to cause a short failure; and
calculating a probability of the defect to cause an open failure and a probability of the defect to cause a short failure.
7. The computer-implemented method of
8. The computer-implemented method of
9. The computer-implemented method of
for each defect on an inspected layer, extracting the defect-specific fail probability value (FPdefects) from a correlation curve corresponding to the step; and
computing the yield-loss contribution to final yield as:
10. The computer-implemented method of
using the defect-specific fail probability value (FPdefects) from the correlation curve and a defective die count per CF bin to compute the final yield as:
11. The computer-implemented method of
using the defect-specific fail probability value (FPdefects) from the correlation curve and a defective die count per CF bin to compute the final yield as:
12. The computer-implemented method of
aggregating all of the defects in a die into a virtual defect; and
for each virtual defect, calculating an aggregated CF value as:
13. A computer-implemented system comprising:
processor coupled to the memory;
a model constructor executed from the memory by the processor to build a yield model for a design; and
a yield analyzer, coupled to the model constructor, to calculate a criticality factor (CF) for each of a plurality of defects detected in at least one inspection process step of a wafer, the wafer comprising a plurality of dies, to determine a non-defect yield component of a final yield for the wafer using a subset of the plurality of dies that have no visual defects, to determine a yield-loss contribution of the inspection process step to the final yield based on the built yield model and the criticality factors calculated for individual defects detected at different process steps, and the non-defect yield component, and to predict the final yield for the wafer using the yield-loss contribution.
14. The computer-implemented system of
a CF calculator, to calculate the CFs of the individual defects; and
a yield predictor, to calculate the yield-loss contribution of an inspection process step to a final yield using the CFs.
15. The computer-implemented system of
16. The computer-implemented system of
a yield model database, coupled to the model constructor, to store the yield model;
a design data database, coupled to the yield analyzer, to store design data, wherein the design data includes at least one of design layouts and routing information for the design layouts; and
a defect reporter, coupled to the yield analyzer, to provide geometric characteristics of individual defects detected during different process steps.
17. The computer-implemented system of
18. The computer-implemented system of
19. A non-transitory computer readable storage medium, comprising executable instructions which when executed on a processing system cause the processing system to perform a method comprising:
calculating a criticality factor (CF) for each of a plurality of defects detected in an at least one inspection process step of a wafer, the wafer comprising a plurality of dies;
determining a non-defect yield component of a final yield for the wafer using a subset of the plurality of dies, wherein the subset of dies has no visual defects;
determining a yield-loss contribution of the inspection process step to the final yield based on CFs of the plurality of defects, a yield model built for a relevant design, and the non-defect yield component; and
predicting the final yield for the wafer using the yield-loss contribution of the inspection process step.
The present application claims priority to U.S. Provisional Application Ser. No. 60/931,725, filed May 24, 2007, which is incorporated herein in its entirety.
The present invention relates to product yield prediction and analysis, and more specifically to dynamic inline yield analysis and prediction.
The fabrication of integrated circuits is an extremely complex process that may involve hundreds of individual operations. In view of the device and interconnect densities required in present day integrated circuits, it is imperative that the manufacturing processes be carried out with utmost precision and in a way that minimizes defects. For reliable operation, the electrical characteristics of the circuits must be kept within carefully controlled limits, which imply a high degree of control over the myriad of operations and fabrication processes. For example, in the photoresist and photomask operations, the presence of contaminants such as dust, minute scratches and other imperfections in the patterns on the photomasks can produce defective patterns on the semiconductor wafers, resulting in defective integrated circuits. Further, defects can be introduced in the circuits during the diffusion operations themselves.
The figure of merit of a semiconductor manufacturing facility is the sort yield obtained by electrically probing the completed devices. However, due to the multitude and complexity of process steps and their associated cost, it is desirable to detect problems early in the design phase in order to correct them, and to predict the yield in order to plan, during the manufacturing phase, wafer starts appropriately. Currently, designers use yield prediction software to decide which design layout alternative will produce a better yield, and thus be printed, and to decide how many wafers to put inline, i.e., adjust the number of wafer starts for production per product based on real inline data to meet the yielding die commitments.
During the design phase, existing software predicts yield based on the wafer design and fabrication defect data using a statistical critical area calculation. The problem with the existing yield prediction software lies in the hidden assumption that the defect distribution is random over the die and the likelihood of a defect to occur on different design elements is the same. With older technology nodes (90 nm and above), where most of the defects were actual random defects, these assumptions could hold true. However, with new technology nodes, where the number of systematic defects, i.e., defects due to non-random errors that are conditioned by the specifics of a design layout or the equipment, has increased significantly, the typical yield prediction software still distributes the defects randomly, even though the defects may occur in specific design elements.
During the manufacturing phase, one way to plan wafer starts appropriately is by utilizing inline inspection tools to detect process defects on the wafers in process. These tools are typically optical microscopes, but of late electron-beam inspection tools have been introduced for certain critical layers. The defects detected by inspection tools are referred to as ‘visual defects’. Not all visual defects will cause an electrical fail. Conversely, not all yield loss can be attributed to visual defects. It has been a goal in the industry to be able to predict the yield loss due to visual defects. The methodology most commonly used is the Kill Ratio method that empirically deduces the fail probabilities of different defect classes by overlaying inline defects with sorted yield data. It is performed on an initial training set, and then applied to wafers in process. This method either requires significant manual classification for the learning set or relies on inspection tool classification that typically has low accuracy and purity. In addition, when new defect classes arise, the learning phase has to be repeated.
The present invention is illustrated by way of example and not limitation in the figures of the accompanying drawings, in which like references indicate similar elements and in which:
Embodiments of the invention provide a mechanism to dynamically predict a defect limited yield using inline inspection defects. The mechanism uses a criticality factor (CF) that indicates whether the defect is likely to cause a killer failure of the device. The CF is based on the critical area for a given defect size on the die layout in the vicinity of the reported defect location. The defect size reported by the inspection tool and the location of the defect defined by the inspection tool may not be absolutely accurate, introducing an uncertainty in the yield outcome. In addition, physical properties of the defects (such as conductivity), which are not characterized by optical inspection, may also affect the yield outcome and introduce an uncertainty in the yield outcome. The mechanism addresses the above uncertainties by adding an empirical learning phase that enables correlation of the calculated CF to the observed die yield. The correlation is updated dynamically and automatically and is constantly improved. The mechanism does not require any kind of manual classification or reliance on imperfect automatic classification.
Some portions of the detailed description which follows are presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of steps leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussion, it is appreciated that throughout the description, discussions utilizing terms such as “processing”, “computing”, “calculating”, “determining”, “displaying” or the like, refer to the actions and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (e.g., electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
The present invention also relates to an apparatus for performing the operations herein. This apparatus may be specially constructed for the required purposes, or it may comprise a general purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions.
The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general purpose systems may be used with programs in accordance with the teachings herein, or it may prove convenient to construct a more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will appear from the description below. In addition, the present invention is not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the invention as described herein.
A machine-readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer). For example, a machine-readable medium includes a machine readable storage medium (e.g., read only memory (“ROM”), random access memory (“RAM”), magnetic disk storage media, optical storage media, flash memory devices, etc.), a machine readable transmission medium (electrical, optical, acoustical or other form of propagated signals (e.g., carrier waves, infrared signals, digital signals, etc.)), etc.
A wafer may be produced according to a design via a fabrication process controlled by a set of process parameters. The design data database 108 may store design data such as design layout, the routing information for the design layout, etc.
During the fabrication process, due to various factors, the wafers may not be formed to exactly match the design. Hence, to determine how the actual wafers vary from the design, one or more wafers undergo an inspection process at various steps during the fabrication process. The inspection process may be performed using any suitable inspection system, such as an UVision™ inspection system available from Applied Materials®. In some cases, the inspection process may be a separate process, and in other cases, it may be performed in conjunction (inline) with the fabrication process. As part of the inspection process, defects in the wafers are identified.
The defect reporter 106 provides geometric characteristics of the individual defects detected during different inspection process steps. The geometric characteristics may include, for example; the size of the defect, the volume of the defect, the location of the defect within the design on which the defect is found, etc. The defect reporter 106 may be part of a defect inspection system such as Dark Field, Bright Field or E-Beam inspection system.
The model constructor 102 is responsible for performing the learning phase during which it builds a yield model for a specific design. The yield model is built using a learning set that may include a set of dies for which there exist inspection results in all critical steps, as well as sort yield data. One embodiment of a method for building yield models will be discussed in more detail below. The yield models corresponding to different designs are stored in the yield model database 104.
The yield analyzer 110 predicts yield based on a Criticality Factor (CF) of individual defects detected at different process steps. The CF indicates whether a defect is likely to be a nuisance defect, or alternatively is likely to have a significant impact on yield. In particular, the CF represents a probability of a defect with certain geometric characteristics (e.g., size or volume) to kill a device. In one embodiment, the yield analyzer 110 includes a CF calculator 112 and a yield predictor 114. The CF calculator 112 calculates the CF of individual defects. The CF may be calculated based on the geometric characteristics of the defect and the layout of the design on which the defect is found. One embodiment of a method for calculating the CF will be discussed in more detail below in conjunction with
The yield predictor 114 correlates CFs calculated for defects in a specific process step to the observed die yield provided by a yield model associated with a relevant design. The yield predictor 114 also calculates the yield-loss contribution of the specific process step to the final yield on that wafer using the correlated CFs. The yield-loss contribution can then be used to predict a final defect limited yield (“final yield”). Based on this prediction, the number of wafer starts can be adjusted, and the adjusted number 116 can be reported to corresponding manufacturing processes in the FAB.
At block 202, data pertaining to defects detected in an inspection step is received. The defects are detected in a specific inspection process step on a wafer having a particular design. The data may include, for example, the size or volume of the defects, the layout of the design in the vicinity of the reported defect locations, etc.
At block 204, the defects are binned together according to Criticality Factor (CF) values. Defects may be categorized into “bins” by CFs. Where defects have similar CFs, they are added to the same defect bin. A CF is calculated for each of the defects using the above defect data. The algorithm determines the CF, i.e., the probability of the defect to cause a killer failure of the device based on geometric considerations such as a likelihood of the defect to bridge two lines (cause a short circuit failure) and a likelihood of the defect to open a line (cause an open circuit failure).
As noted above, one of the weaknesses with existing yield prediction software is the hidden assumption that the defect distribution is random over the die and the likelihood of a defect to occur on different design elements is the same. To overcome this weakness, both random and systematic defects is considered in predicting the yield.
Defect phenomena are factors that affect yield and are usually classified as random defects or systematic defects. For example, yield can be affected by random defects such as extra material particles or missing material particles. Yield could also be affected by systematic defects, such as under-etching an isolated via (connector pad), the collapse of a photo-resist pattern, or erosion during chemical-mechanical planarization (CMP), etc.
Random defects are attributed to contamination of a wafer, e.g., the extra material particles or missing material particles. Systematic defects are due to non-random errors and are conditioned by the specifics of a design layout or the equipment, e.g., defects attributed to under-etching a via or collapsing a photo-resist pattern. Thus, systematic defects are strongly dependent on the design interactions (the relationship between design features and process parameters). As a consequence, systematic defects will occur in a very specific location or locations. This suggests that the CF approach for predicting fail probability of a defect might not work well for systematic defects.
Thus, prior to calculating the CF, the defects are separated into random defects and systematic defects and the systematic defects are excluded from the CF calculation.
Typically, the defects are hard to separate. However, using Design Based Binning (DBB), defects in fabrications of a wafer may be categorized as either systematic or random defects. In a patent application entitled “Design-Based Method for Grouping Systematic Defects in Lithography Pattern Writing System,” publication number US20060269120A1, incorporated herein by reference, a method for grouping defects is described.
The location of the defects is determined on a wafer. By analyzing the layout in the neighborhood (the design layout in the area surrounding the defect) of the defect, and matching it to similar defected neighborhoods in different location across the die, defects may be categorized by common structures in which they occur. The inspection defects can be classified as random defects or systematic defects and the systematic defects may be excluded from the CF calculations.
Once the defects are separated, and the systematic defects excluded, the CF for each random defect detected in a process step is calculated. The CF of a defect is the probability to cause a killer failure based strictly on geometrical consideration: what is the likelihood of the defect to bridge (short) two lines, or to open a line. The CF values may be determined by using an exemplary CF calculation algorithm discussed in detail below in conjunction with
At block 206, a yield model is built for a specific design using a learning set to correlate CF to sort yield. Had the defect location been known to absolute accuracy, and had geometrical considerations alone been sufficient to determine whether a fail occurred or not, then the yield outcome would be completely determinate. However, as noted above, the location inaccuracy of inspection tools can introduce an uncertainty that necessitates probability estimation. Moreover, physical properties of the defects (such as conductivity), which are not characterized by optical inspection, may affect the yield outcome and can introduce an uncertainty in the yield outcome even if the location was known to absolute accuracy. Therefore, in one embodiment, an empirical “learning” phase to correlate the calculated CF to the observed die yield is introduced to account for such uncertainties. The learning phase addresses the above uncertainties by correlating CF to sort yield using yield models.
A yield model provides quantitative characterization of the defect phenomenon in terms of probability of failure of a design (or some part of the design) under certain conditions. It allows a fail probability value to be attributed to a defect according to its CF using a learning set, i.e., a set of dies (presumably over multiple wafers) for which there are inspection results in all critical steps.
The yield model described here is constructed strictly for random visual defects. In one embodiment, by separating out the systematic visual defects using the DBB methodology mentioned above, the systematic visual defects are treated separately as discussed below.
For simplicity, in one embodiment, when building a yield model, it is assumed that there is no more than one defect per die. In another embodiment, a generalization to multiple defects per die is explained below.
As illustrated in the following exemplary expression, the final yield may comprise two components: (1) a visual defect component and (2) a non-defect component, each having a yield value and a yield-loss value associated with them (Y and YL respectively):
The random defects are distributed among S inspection steps. In each step the random defects are classified, according to similar CF values, into CF bins, Bs, each with Nb defects. The classification may be represented in the following exemplary expression.
In the yield model, all defects within one bin have one fail probability value (and hence one yield-loss value) associated with them. Therefore Expression 2 can be rewritten in the following exemplary expression:
Inputting Expression 3 and Expression 4 into Expression 1, the final yield is represented by the following exemplary expression:
In order to predict the final yield (from Expression 5), the non-defect yield component (Ynon-defects) and defect-specific fail probability (FPdefects s,b), are evaluated over all detected inline defects.
The evaluation is performed by using a learning set, i.e., a set of dies for which there are inspection results in all critical steps, and sort yield data. One embodiment of constructing a learning set is discussed in more detail below. The learning set is divided into two subsets “A,” and “B.” Subset A consists of dies from the learning set having no visual defects (henceforth, “Clean Dies”). Subset B consists of dies from the learning set having defects (henceforth, “Dirty Dies”).
With respect to the non-defect yield component, (Ynon-defects), in Expression 1 (Yfinal=Ynon-defects·Ydefects) there is an implicit assumption that defect and non-defect yield defect phenomena arise completely independently of each other. Therefore, it may be assumed that the yield-loss for subset A, the clean dies, at any inspection step is due strictly to non-defect causes. As a result, the non-defect yield component is a constant value, c, represented in the exemplary expression below, in that it does not depend on any defect attributes.
In the case of a subset B, the dirty dies, the yield may be attributed to both defect and non-defect causes, as seen in the following exemplary expression:
The dies in subset B can then be binned according to the CF value of their defects (e.g., divide the CF range into 10 equal bins of 0.1). The binning of the dies in subset B by CF value thus creates sub-subsets of subset B.
For example, one of the sub-subsets, “C,” may consist of dies with defects only of CF bin b in step s (and no defects in any other step). Expression 7 would still apply to this new sub-subset. Moreover, for such a sub-subset, the yield-loss equals the fail probability since all of the dies in the sub-subset have only defects.
Expression 7, YDirtyDies=Ynon-defects·Ydefects=c·Ydefects=c·(1−YLdefects), can then be solved for the fail probability attributed strictly to the defects in this subset using the following exemplary expression:
For each of the remaining CF bin sub-subsets of subset B, the FPdefects s,b is calculated. The calculated results are plotted to a functional form FPdefects=f (CF).
The above operations are repeated for each of the critical inspection steps, thus building a yield model that allows a fail probability value to be attributed to a defect according to its CF using a learning set. The above correlations defined by the yield model can be used to predict yield for new incoming wafers throughout the process stack.
In regards to systematic defects, once a systematic defect is classified to a certain design-based bin, a fail probability value can be attributed to that bin by one of two methods:
At block 208, a yield-loss contribution of the process step to a final yield is determined based on the calculated CFs and the observed die yield provided by the yield model associated with a relevant design.
In one embodiment, the yield model is used to calculate the yield-loss contribution for a wafer that has performed up to inspection step s. The yield-loss contribution to the final yield for this wafer up to inspection step s can be calculated in the following exemplary expression:
In practice, the correlation curve for step s in
At block 210, the final yield is predicted using the yield-loss contribution. In predicting the final yield, Expression 5 is simplified to the following exemplary expression:
Wafers may be at different stages (i.e., steps) in the fabrication process. There are wafers that have reached the final critical inspection step, e.g., step S. There are also inline wafers, i.e., wafers that have not reached step S, but rather have completed only up to step M, where step M is a step before the final step S. Thus, predicting the final yield depends upon which step a wafer is located.
For a wafer that has gone through all the critical inspection steps, the final predicted yield may be calculated by inputting the appropriate FPdefects for the defects (as obtained from the correlation curves) as well as the appropriate defective die count per CF bin.
For an inline wafer, an average yield-loss contribution is used to represent the not-yet completed steps, e.g., the steps between M and the final step S. Thus, for a wafer located at step M (where M<S), the final yield may be predicted using the following exemplary expression:
Note that the yield-loss contribution of the process step to a final yield determined at block 208 can be used in two ways: (a) for trending/monitoring and to capture baseline yield drifts (and limit major yield excursions) due to defects early in the process, and (b) to obtain representative values of the yield loss contribution for each step (this can be an average value,
Note, in one embodiment, the “fundamental unit” of yield-loss is a single defect. In another embodiment, there is more than one defect per die, especially for large dies. Where there is more than one defect per die, all of the defects that fall into one die are aggregated into one virtual defect.
For each virtual defect an aggregated CF value is calculated using the following exemplary expression:
This calculation may be performed in both the learning phase and in the yield prediction phase.
At block 212, an adjusted number of wafer starts is determined based on the predicted yield.
At block 306, a short area is identified. The short area is the total area of all locations inside the design layout area, in which the defect would cause a shorting failure. At block 308, a probability of the defect to cause a shorting failure (short probability) is calculated. Short probability may be calculated as a ratio between the short area and the entire design layout area.
At block 310, a CF is calculated as a kill ratio of the defect considering both open probability and short probability. For example, the CF can be calculated using the following exemplary expression:
One embodiment of constructing a learning set will now be discussed. A learning set is a set of representative wafers having inspection data at all critical layers, and die sort data (i.e., pass, fail). The defect data should be of adders only, i.e., only the defects that originate for the first time in the specific inspected layer are included. Each die in this learning set is identified by a die ID. An exemplary learning set is illustrated in the following table:
Each column contains information for a different die. Each row, with the exception of the last row, contains information for a different inspection step. The table is populated by the calculated CF values for defects that fell in a certain die at a certain step. If more than one defect falls in a die at a certain step, the aggregated die CF may be used as outlined in Expression 12,
If no defects were detected on a certain die at a certain step, ND, (denoting No Defects) may be entered. Note the case where no defects were detected on a certain die at a certain step is distinguished from the case in which a defect occurred with CF=0, i.e., when the defect is smaller than the smallest critical dimension in the layout window. The last row of the table contains the sort data for each die.
The non-defect yield component from a subset of clean dies is then calculated, using Expression 6,
From the above table, correlation curves for each inspection step may be constructed as follows:
The learning set may be constantly updated with new data resulting in more refined CF correlation curves and non-defect yield components. The following are guidelines for a minimum learning set size:
The yield calculation is done as follows:
From Expression 1 Ydefects is now split into a random defect component and a systematic defect component as seen in the following exemplary expression:
Yran is calculated as outlined above for all random defects in all inspected layers. Ysys may be calculated using the following exemplary expression for all systematic defects in all inspected layers (this is equivalent to Expression 10 for random defects):
The exemplary computer system 500 includes a processing device (processor) 502, a main memory 504 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.), a static memory 506 (e.g., flash memory, static random access memory (SRAM), etc.), and a data storage device 518, which communicate with each other via a bus 530.
Processor 502 represents one or more general-purpose processing devices such as a microprocessor, central processing unit, or the like. More particularly, the processor 502 may be a complex instruction set computing (CISC) microprocessor, reduced instruction set computing (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or a processor implementing other instruction sets or processors implementing a combination of instruction sets. The processor 502 may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like. The processor 502 is configured to execute the processing logic 526 for performing the operations and steps discussed herein.
The computer system 500 may further include a network interface device 508. The computer system 500 also may include a video display unit 510 (e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)), an alphanumeric input device 512 (e.g., a keyboard), a cursor control device 514 (e.g., a mouse), and a signal generation device 516 (e.g., a speaker).
The data storage device 518 may include a machine-accessible storage medium 531 on which is stored one or more sets of instructions (e.g., software 522) embodying any one or more of the methodologies or functions described herein. The software 522 may also reside, completely or at least partially, within the main memory 504 and/or within the processor 502 during execution thereof by the computer system 500, the main memory 504 and the processor 502 also constituting machine-accessible storage media. The software 522 may further be transmitted or received over a network 520 via the network interface device 508.
The machine-accessible storage medium 531 may also be used to store data structure sets that define user identifying states and user preferences that define user profiles. Data structure sets and user profiles may also be stored in other sections of computer system 500, such as static memory 506.
While the machine-accessible storage medium 531 is shown in an exemplary embodiment to be a single medium, the term “machine-accessible storage medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “machine-accessible storage medium” shall also be taken to include any medium that is capable of storing, encoding or carrying a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methodologies of the present invention. The term “machine-accessible storage medium” shall accordingly be taken to include, but not be limited to, solid-state memories, and optical and magnetic media.
It is to be understood that the above description is intended to be illustrative, and not restrictive. Many other embodiments will be apparent to those of skill in the art upon reading and understanding the above description. The scope of the invention should, therefore, be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled.