|Publication number||US8040727 B1|
|Application number||US 09/143,233|
|Publication date||Oct 18, 2011|
|Priority date||Apr 13, 1989|
|Publication number||09143233, 143233, US 8040727 B1, US 8040727B1, US-B1-8040727, US8040727 B1, US8040727B1|
|Original Assignee||Sandisk Corporation|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (129), Non-Patent Citations (16), Referenced by (7), Classifications (61), Legal Events (3)|
|External Links: USPTO, USPTO Assignment, Espacenet|
This is a continuation of patent application Ser. No. 08/771,708, filed Dec. 20, 1996, now U.S. Pat. No. 5,991,517 which is a continuation of patent application Ser. No. 08/174,768, filed Dec. 29, 1993, now U.S. Pat. No. 5,602,987, which in turn is a continuation of patent application Ser. No. 07/963,838, filed Oct. 20, 1992, now U.S. Pat. No. 5,297,148, which in turn is a division of patent application Ser. No. 07/337,566, filed Apr. 13, 1989, now abandoned.
This invention relates generally to semiconductor electrically erasable programmable read only memories (EEprom), and specifically to a system of integrated circuit Flash EEprom chips.
Computer systems typically use magnetic disk drives for mass storage of data. However, disk drives are disadvantageous in that they are bulky and in their requirement for high precision moving mechanical parts. Consequently they are not rugged and are prone to reliability problems, as well as consuming significant amounts of power. Solid state memory devices such as DRAM's and SRAM's do not suffer from these disadvantages. However, they are much more expensive, and require constant power to maintain their memory (volatile). Consequently, they are typically used as temporary storage.
EEprom's and Flash EEprom's are also solid state memory devices. Moreover, they are nonvolatile, and retain their memory even after power is shut down. However, conventional Flash EEprom's have a limited lifetime in terms of the number of write (or program)/erase cycles they can endure. Typically the devices are rendered unreliable after 102 to 103 write/erase cycles. Traditionally, they are typically used in applications where semi-permanent storage of data or program is required but with a limited need for reprogramming.
Accordingly, it is an object of the present invention to provide a Flash EEprom memory system with enhanced performance and which remains reliable after enduring a large number of write/erase cycles.
It is another object of the present invention to provide an improved Flash EEprom system which can serve as non-volatile memory in a computer system.
It is another object of the present invention to provide an improved Flash EEprom system that can replace magnetic disk storage devices in computer systems.
It is another object of the present invention to provide a Flash EEprom system with improved erase operation.
It is another object of the present invention to provide a Flash EEprom system with improved error correction.
It is yet another object of the present invention to provide a Flash EEprom with improved write operation that minimizes stress to the Flash EEprom device.
It is still another object of the present invention to provide a Flash EEprom system with enhanced write operation.
These and additional objects are accomplished by improvements in the architecture of a system of EEprom chips, and the circuits and techniques therein.
According to one aspect of the present invention, an array of Flash EEprom cells on a chip is organized into sectors such that all cells within each sector are erasable at once. A Flash EEprom memory system comprises one or more Flash EEprom chips under the control of a controller. The invention allows any combination of sectors among the chips to be selected and then erased simultaneously. This is faster and more efficient than prior art schemes where all the sectors must be erased every time or only one sector at a time can be erased. The invention further allows any combination of sectors selected for erase to be deselected and prevented from further erasing during the erase operation. This feature is important for stopping those sectors that are first to be erased correctly to the “erased” state from over erasing, thereby preventing unnecessary stress to the Flash EEprom device. The invention also allows a global de-select of all sectors in the system so that no sectors are selected for erase. This global reset can quickly put the system back to its initial state ready for selecting the next combination of sectors for erase. Another feature of the invention is that the selection is independent of the chip select signal which enables a particular chip for read or write operation. Therefore it is possible to perform an erase operation on some of the Flash EEprom chips while read and write operations may be performed on other chips not involved in the erase operation.
According to another aspect of the invention, improved error correction circuits and techniques are used to correct for errors arising from defective Flash EEprom memory cells. One feature of the invention allows defect mapping at cell level in which a defective cell is replaced by a substitute cell from the same sector. The defect pointer which connects the address of the defective cell to that of the substitute cell is stored in a defect map. Every time the defective cell is accessed, its bad data is replaced by the good data from the substitute cell.
Another feature of the invention allows defect mapping at the sector level. When the number of defective cells in a sector exceeds a predetermined number, the sector containing the defective cells is replaced by a substitute sector.
An important feature of the invention allows defective cells or defective sectors to be remapped as soon as they are detected thereby enabling error correction codes to adequately rectify the relatively few errors that may crop up in the system.
According to yet another aspect of the present invention, a write cache is used to minimize the number of writes to the Flash EEprom memory. In this way the Flash EEprom memory will be subject to fewer stress inducing write/erase cycles, thereby retarding its aging. The most active data files are written to the cache memory instead of the Flash EEprom memory. Only when the activity levels have reduced to a predetermined level are the data files written from the cache memory to the Flash EEprom memory. Another advantage of the invention is the increase in write throughput by virtue of the faster cache memory.
According to yet another aspect of the present invention, one or more printed circuit cards are provided which contain controller and EEprom circuit chips for use in a computer system memory for long term, non-volatile storage, in place of a hard disk system, and which incorporate various of the other aspects of this invention alone and in combination.
Additional objects, features, and advantages of the present invention will be understood from the following description of its preferred embodiments, which description should be taken in conjunction with the accompanying drawings.
A computer system in which the various aspects of the present invention are incorporated is illustrated generally in
One aspect of the present invention is the substitution of a specific type of semiconductor memory system for the disk drive but without having to sacrifice non-volatility, ease of erasing and rewriting data into the memory, speed of access, low cost and reliability. This is accomplished by employing an array of electrically erasable programmable read only memories (EEprom's) integrated circuit chips. This type of memory has additional advantages of requiring less power to operate, and of being lighter in weight than a hard disk drive magnetic media memory, thereby being especially suited for battery operated portable computers.
The bulk storage memory 29 is constructed of a memory controller 31, connected to the computer system bus 23, and an array 33 of EEprom integrated circuit chips. Data and instructions are communicated from the controller 31 to the EEprom array 33 primarily over a serial data line 35. Similarly, data and status signals are communicated from the EEprom 33 to the controller 31 over serial data lines 37. Other control and status circuits between the controller 31 and the EEprom array 33 are not shown in
The EEprom array 33 includes a number of EEprom integrated circuit chips 43, 45, 47, etc. Each includes a respective chip select and enable line 49, 51 and 53 from interface circuits 40. The interface circuits 40 also act to interface between the serial data lines 35, 37 and a circuit 55. Memory location addresses and data being written into or read from the EEprom chips 43, 45, 47, etc. are communicated from a bus 55, through logic and register circuits 57 and thence by another bus 59 to each of the memory chips 43, 45, 47 etc.
The bulk storage memory 29 of
For large amounts of memory, that which is conveniently provided by a single array 33 may not be enough. In such a case, additional EEprom arrays can be connected to the serial data lines 35 and 37 of the controller chip 31, as indicated in
Erase of Memory Structures
In system designs that store data in files or blocks the data will need to be periodically updated with revised or new information. It may also be desirable to overwrite some no longer needed information, in order to accommodate additional information. In a Flash EEprom memory, the memory cells must first be erased before information is placed in them. That is, a write (or program) operation is always preceded by an erase operation.
In conventional Flash erase memory devices, the erase operation is done in one of several ways. For example, in some devices such as the Intel corporation's model 27F-256 CMOS Flash EEprom, the entire chip is erased at one time. If not all the information in the chip is to be erased, the information must first be temporarily saved, and is usually written into another memory (typically RAM). The information is then restored into the nonvolatile Flash erase memory by programming back into the device. This is very slow and requires extra memory as holding space.
In other devices such as Seeq Technology Incorporated's model 48512 Flash EEprom chip, the memory is divided into blocks (or sectors) that are each separately erasable, but only one at a time. By selecting the desired sector and going through the erase sequence the designated area is erased. While, the need for temporary memory is reduced, erase in various areas of the memory still requires a time consuming sequential approach.
In the present invention, the Flash EEprom memory is divided into sectors where all cells within each sector are erasable together. Each sector can be addressed separately and selected for erase. One important feature is the ability to select any combination of sectors for erase together. This will allow for a much faster system erase than by doing each one independently as in prior art.
For example, in order to select the sector 211 for erase, the controller sends the address of the sector 211 to the circuit 220. The address is decoded in line 235 and is used in combination with a set erase enable signal in bus 237 to set an output 239 of the register 221 to HIGH. This enables the sector 211 in a subsequent erase operation. Similarly, if the sector 213 is also desired to be erased, its associated register 223 may be set HIGH.
After all sectors intended for erase have been selected, the controller then issues to the circuit 220, as well as all other chips in the system a global erase command in line 251 along with the high voltage for erasing in line 209. The device will then erase all the sectors that have been selected (i.e. the sectors 211 and 213) at one time. In addition to erasing the desired sectors within a chip, the architecture of the present system permits selection of sectors across various chips for simultaneous erase.
FIGS. 4(1)-4(11) illustrate the algorithm used in conjunction with the circuit 220 of
Optimized erase implementations have been disclosed in two U.S. patent applications. They are U.S. patent applications, Ser. No. 204,175, filed Jun. 8, 1988, by Dr. Eliyahou Harari, now U.S. Pat. No. 5,095,344, and one entitled “Multi-State EEprom Read and Write Circuits and Techniques, Ser. No. 07/337,579, filed Apr. 13, 1989, now abandoned, by Sanjay Mehrotra and Dr. Eliyahou Harari. The disclosures of the two applications are hereby incorporate by reference. The Flash EEprom cells are erased by applying a pulse of erasing voltage followed by a read to verify if the cells are erased to the “erased” state. If not, further pulsing and verifying are repeated until the cells are verified to be erased. By erasing in this controlled manner, the cells are not subject to over-erasure which tends to age the EEprom device prematurely as well as make the cells harder to program.
As the group of selected sectors is going through the erase cycle, some sectors will reach the “erase” state earlier than others. Another important feature of the present invention is the ability to remove those sectors that have been verified to be erased from the group of selected sectors, thereby preventing them from over-erasing.
Returning to FIG. 4(4), after all sectors intended for erase have been tagged, the controller initiates an erase cycle to erase the group of tagged sectors. In FIG. 4(5), the controller shifts in a global command called Enable Erase into each Flash EEprom chip that is to perform an erase. This is followed in FIG. 4(5) by the controller raising of the erase voltage line (Ve) to a specified value for a specified duration. The controller will lower this voltage at the end of the erase duration time. In FIG. 4(6), the controller will then do a read verify sequence on the sectors selected for erase. In FIG. 4(7), if none of the sectors are verified, the sequences illustrated in FIGS. 4(5)-4(7) are repeated. In FIGS. 4(8) and 3(9), if one or more sectors are verified to be erased, they are taken out of the sequence. Referring also to
The ability to select which sectors to erase and which ones not to, as well as which ones to stop erasing is advantageous. It will allow sectors that have erased before the slower erased sectors to be removed from the erase sequence so no further stress on the device will occur. This will increase the reliability of the system. Additional advantage is that if a sector is bad or is not used for some reason, that sector can be skipped over with no erase occurring within that sector. For example, if a sector is defective and have shorts in it, it may consume much power. A significant system advantage is gained by the present invention which allows it to be skipped on erase cycles so that it may greatly reduce the power required to erase the chip.
Another consideration in having the ability to pick the sectors to be erased within a device is the power savings to the system. The flexibility in erase configuration of the present invention enables the adaptation of the erase needs to the power capability of the system. This can be done by configuring the systems to be erased differently by software on a fixed basis between different systems. It also will allow the controller to adaptively change the amount of erasing being done by monitoring the voltage level in a system, such as a laptop computer.
An additional performance capability of the system in the present invention is the ability to issue a reset command to a Flash EEprom chip which will clear all erase enable latches and will prevent any further erase cycles from occurring. This is illustrated in
An additional performance capability is to have the ability to do erase operations without regard to chip select. Once an erase is started in some of the memory chips, the controller in the system can access other memory chips and do read and write operations on them. In addition, the device(s) doing the erase can be selected and have an address loaded for the next command following the erase.
Physical defects in memory devices give rise to hard errors. Data becomes corrupted whenever it is stored in the defective cells. In conventional memory devices such as RAM's and Disks, any physical defects arising from the manufacturing process are corrected at the factory. In RAM's, spare redundant memory cells on chip may be patched on, in place of the defective cells. In the traditional disk drive, the medium is imperfect and susceptible to defects. To overcome this problem manufacturers have devised various methods of operating with these defects present, the most usual being defect mapping of sectors. In a normal disk system the media is divided into cylinders and sectors. The sector being the basic unit in which data is stored. When a system is partitioned into the various sectors the sectors containing the defects are identified and are marked as bad and not to be used by the system. This is done in several ways. A defect map table is stored on a particular portion of the disk to be used by the interfacing controller. In addition, the bad sectors are marked as bad by special ID and flag markers. When the defect is addressed, the data that would normally be stored there is placed in an alternative location. The requirement for alternative sectors makes the system assign spare sectors at some specific interval or location. This reduces the amount of memory capacity and is a performance issue in how the alternative sectors are located.
One important application of the present invention is to replace a conventional disk storage device with a system incorporating an array of Flash EEprom memory chips. The EEprom system is preferably set up to emulate a conventional disk, and may be regarded as a “solid-state disk”.
In a “disk” system made from such solid-state memory devices, low cost considerations necessitate efficient handling of defects. Another important feature of the invention enables the error correction scheme to conserve as much memory as possible. Essentially, it calls for the defective cells to be remapped cell by cell rather than by throwing away the whole sector (512 bytes typically) whenever a defect occurs in it. This scheme is especially suited to the Flash EEprom medium since the majority of errors will be bit errors rather than a long stream of adjacent defects as is typical in traditional disk medium.
In both cases of the prior art RAM and magnetic disk, once the device is shipped from the factory, there is little or no provision for replacing hard errors resulting from physical defects that appear later during normal operation. Error corrections then mainly rely on schemes using error correction codes (ECC).
The nature of the Flash EEprom device predicates a higher rate of cell failure especially with increasing program/erase cycling. The hard errors that accumulate with use would eventually overwhelm the ECC and render the device unusable. One important feature of the present invention is the ability for the system to correct for hard errors whenever they occur. Defective cells are detected by their failure to program or erase correctly. Also during read operation, defective cells are detected and located by the ECC. As soon as a defective cell is identified, the controller will apply defect mapping to replace the defective cell with a space cell located usually within the same sector. This dynamic correction of hard errors, in addition to conventional error correction schemes, significantly prolongs the life of the device.
Another feature of the present invention is an adaptive approach to error correction. Error correction code (ECC) is employed at all times to correct for soft errors as well as any hard errors that may arise. As soon as a hard error is detected, defect mapping is used to replace the defective cell with a spare cell in the same sector block. Only when the number of defective cells in a sector exceeds the defect mapping's capacity for that specific sector will the whole sector be replaced as in a, conventional disk system. This scheme minimized wastage without compromising reliability.
Whenever a defective cell is detected in the sector, a good cell in the alternative defects data area 407 is assigned to backup the data designated for the defective cell. Thus even if the defective cell stores the data incorrectly, an error-free copy is stored in the backup cell. The addresses of the defective cell and the backup cell are stored as defect pointers in the defect map 409.
It is to be understood that the partitioning between the user data portion 403 and the spare portion 405 need not be rigid. The relative size of the various partitioned areas may be logically reassigned. Also the grouping of the various areas is largely for the purpose of discussion and not necessarily physically so. For example, the alternative defects data area 407 has been schematically grouped under the spare portion 405 to express the point that the space it occupies is no longer available to the user.
In a read operation, the controller first reads the header, the defect map and the alternative defects data. It then reads the actual data. It keeps track of defective cells and the location of the substitute data by means of the defect map. Whenever a defective cell is encountered, the controller substitutes its bad data with the good data from the alternative defects.
After assuming control, the controller 31 first addresses the header of the sector and verifies that the memory is accessed at the address that the user had specified. This is achieved by the following sequence. The controller selects a memory chip (chip select) among the memory device 33 and shifts the address for the header area from the address generator 503 out to the selected memory chip in the memory device 33. The controller then switches the multiplexer 513 and shifts also the read command out to the memory device 33. Then the memory device reads the address sent it and begins sending serial data from the addressed sector back to the controller. A receiver 515 in the controller receives this data and puts it in parallel format. In one embodiment, once a byte (8 bits) is compiled, the controller compares the received data against the header data previously stored by the microprocessor in the holding register file 509. If the compare is correct, the proper location is verified and the sequence continues.
Next the controller 31 reads the defect pointers and loads these bad address locations into the holding register file 509. This is followed by the controller reading the alternative defects data that were written to replace the bad bits as they were written. The alternative bits are stored in an alternative defects data file 517 that will be accessed as the data bits are read.
Once the Header has been determined to be a match and the defect pointers and alternative bits have been loaded, the controller begins to shift out the address of the lowest address of the desired sector to be read. The data from the sector in the memory device 33 is then shifted into the controller chip 31. The receiver 515 converts the data to a parallel format and transfers each byte into a temporary holding FIFO 519 to be shipped out of the controller.
A pipeline architecture is employed to provide efficient throughput as the data is gated through the controller from the receiver 515 to the FIFO 519. As each data bit is received from memory the controller is comparing the address of the data being sent (stored in the address generator 507) against the defect pointer map (stored in the register file 509). If the address is determined to be a bad location, by a match at the output of the comparator 521, the bad bit from the memory received by the receiver 515 is replaced by the good bit for that location. The good bit is obtained from the alternative defects data file 517. This is done by switching the multiplexer 523 to receive the good bit from the alternative defects data file instead of the bad bit from the receiver 515, as the data is sent to the FIFO 519. Once the corrected data is in the FIFO it is ready to be sent to buffer memory or system memory (not shown). The data is sent from the controller's FIFO 519 to the system memory by the controller's DMA controller 507. This controller 507 then requests and gets access to the system bus and puts out an address and gates the data via an output interface 525 out to the system bus. This is done as each byte gets loaded into the FIFO 519. As the corrected data is loaded into the FIFO it will also be gated into an ECC hardware 527 where the data file will be acted on by the ECC.
Thus in the manner described, the data read from the memory device 33 is gated through the controller 31 to be sent to the system. This process continues until the last bit of addressed data has been transferred.
In spite of defect mapping of previously detected defective cells, new hard errors might occur since the last mapping. As the dynamic defect mapping constantly “puts away” new defective cells, the latest hard error that may arise between defect mapping would be adequately handled by the ECC. As the data is gated through the controller 31, the controller is gating the ECC bits into the ECC hardware 527 to determine if the stored value matched the just calculated remainder value. If it matches then the data transferred out to the system memory was good and the read operation was completed. However, if the ECC registers an error then a correction calculation on the data sent to system memory is performed and the corrected data re-transmitted. The method for calculating the error can be done in hardware or software by conventional methods. The ECC is also able to calculate and locate the defective cell causing the error. This may be used by the controller 31 to update the defect map associated with the sector in which the defective cell is detected. In this manner, hard errors are constantly removed from the Flash EEprom system.
Next, the controller begins to fetch the write data from system memory (not shown). It does this by getting access to the system bus, outputs the memory or bus address and does the read cycle. It pulls the data into a FIFO 601 through an input interface 603. The controller then shifts the starting sector address (lowest byte address) from the address generator 503 to the selected memory device 33. This is followed by data from the FIFO 601. These data are routed through multiplexers 605 and 513 and converted to serial format before being sent to the memory device 33. This sequence continues until all bytes for a write cycle have been loaded into the selected memory.
A pipeline architecture is employed to provide efficient throughput as the data is gated from the FIFO 601 to the selected memory 33. The data gated out of the FIFO 601 is sent to the ECC hardware 527 where a remainder value will be calculated within the ECC. In the next stage, as the data is being sent to the memory device through multiplexers 605 and 513, the comparator 521 is comparing its address from the address generator 503 to the defect pointer address values in the holding register file 509. When a match occurs, indicating that a defective location is about to be written, the controller saves this bit into the alternative defect data file 517. At the same time, all bad bits sent to memory will be sent as zeroes.
After the bytes for a write cycle have been loaded into the selected memory device, the controller issues a program command to the memory device and initiate a write cycle. Optimized implementations of write operation for Flash EEprom device have been disclosed in two previously cited U.S. patent applications, Ser. No. 204,175 now U.S. Pat. No. 5,095,344, and one entitled “Multi-State EEprom Read and Write Circuits and Techniques, Ser. No. 07/337,579, filed Apr. 13, 1989, now abandoned. Relevant portions of the disclosures are hereby incorporated by reference. Briefly, during the write cycle, the controller applies a pulse of programming (or writing) voltages. This is followed by a verify read to determine if all the bits have been programmed properly. If the bits did not verify, the controller repeats the program/verify cycle until all bits are correctly programmed.
If a bit fails to verify after prolonged program/verify cycling, the controller will designate that bit as defective and update the defect map accordingly. The updating is done dynamically, as soon as the defective cell is detected. Similar actions are taken in the case of failure in erase verify.
After all the bits have been programmed and verified, the controller loads the next data bits from the FIFO 601 and addresses the next location in the addressed sector. It then performs another program/verify sequence on the next set of bytes. The sequence continues until the end of the data for that sector. Once this has occurred, the controller addresses the shadow memory (header area) associated with the sector (see
In addition, the collection of bits that was flagged as defective and were saved in the alternative defects data file 516 is then written in memory at the alternative defects data locations (see
The present invention also has provision for defect mapping of the whole sector, but only after the number of defective cells in the sector has exceeded the cell defect mapping's capacity for that specific sector. A count is kept of the number of defective cells in each sector. When the number in a sector exceeds a predetermined value, the controller marks that sector as defective and maps it to another sector. The defect pointer for the linked sectors may be stored in a sector defect map. The sector defect map may be located in the original defective sector if its spare area is sufficiently defect-free. However, when the data area of the sector has accumulated a large number of defects, it is quite likely that the spare area WILL also be full of defects.
Thus, it is preferable in another embodiment to locate the sector map in another memory maintained by the controller. The memory may be located in the controller hardware or be part of the Flash EEprom memory. When the controller is given an address to access data, the controller compares this address against the sector defect-map. If a match occurs then access to the defective sector is denied and the substitute address present in the defect map is entered, and the corresponding substitute sector is accessed instead.
In yet another embodiment, the sector remapping is performed by the microprocessor. The microprocessor looks at the incoming address and compares it against the sector defect map. If a match occurs, it does not issue the command to the controller but instead substitute the alternative location as the new command.
Apart from the much higher speed of the solid-state disk, another advantage is the lack of mechanical parts. The long seek times, rotational latency inherent in disk drives are not present. In addition, the long synchronization times, sync mark detects and write gaps are not required. Thus the overhead needed for accessing the location where data is to be read or written is much less. All of these simplifications and lack of constraints result in a much faster system with much reduced overheads. In addition, the files can be arranged in memory in any address order desired, only requiring the controller to know how to get at the data as needed.
Another feature of the invention is that defect mapping is implemented without the need to interrupt the data stream transferred to or from the sector. The data in a block which may contain errors are transferred regardless, and is corrected afterwards. Preserving the sequential addressing will result in higher speed by itself. Further, it allows the implementation of an efficient pipeline architecture in the read and write data paths.
Write Cache System
Cache memory is generally used to speed up the performance of systems having slower access devices. For example in a computer system, access of data from disk storage is slow and the speed would be greatly improved if the data could be obtained from the much faster RAM. Typically a part of system RAM is used as a cache for temporarily holding the most recently accessed data from disk. The next time the data is needed, it may be obtained from the fast cache instead of the slow disk. The scheme works well in situations where the same data is repeatedly operated on. This is the case in most structures and programs since the computer tends to work within a small area of memory at a time in running a program. Another example of caching is the using of, faster SRAM cache to speed up access of data normally stored in cheaper but slower DRAM.
Most of the conventional cache designs are read caches for speeding up reads from memory. In some cases, write caches are used for speeding up writes to memory. However in the case of writes to system memory (e.g. disks), data is still being written to system memory directly every time they occur, while being written into cache at the same time. This is done because of concern for loss of updated data files in case of power loss. If the write data is only stored in the cache memory (volatile) a loss of power will result in the new updated files being lost from cache before having the old data updated in system memory (non-volatile). The system will then be operating on the old data when these files are used in further processing. The need to write to main memory every time defeats the caching mechanism for writes. Read caching does not have this concern since the data that could be lost from cache has a backup on disk.
In the present invention, a system of Flash EEprom is used to provide non-volatile memory in place of traditional system memories such as disk storage. However, Flash EEprom memory is subject to wearing out by excessive program/erase cycles. Even with the improved Flash EEprom memory device as disclosed in U.S. patent applications, Ser. No. 204,175, now U.S. Pat. No. 5,095,344, and Harari, Ser. No. 07/337,579, filed Apr. 13, 1989, now abandoned, and Techniques,” by Sanjay Mehrotra and Dr. Eliyahou Harari filed on the same day as the present application, the endurance limit is approximately 106 program/erase cycles. In a ten-year projected life time of the device, this translates to a limit of one program/erase cycle per 5 minutes. This may be marginal in normal computer usage.
To overcome this problem, a cache memory is used in a novel way to insulate the Flash EEprom memory device from enduring too many program/erase cycles. The primary function of the cache is to act on writes to the Flash EEprom memory and not on reads of the Flash EEprom memory, unlike the case with traditional caches. Instead of writing to the Flash EEprom memory every time the data is updated, the data may be operated on several times in the cache before being committed to the Flash EEprom memory. This reduces the number of writes to the Flash EEprom memory. Also, by writing mostly into the faster cache memory and reducing the number of writes to the slower Flash EEprom, an additional benefit is the increase in system write throughput.
A relatively small size cache memory is quite effective to implement the present invention. This helps to overcome the problem of data loss in the volatile cache memory during a power loss. In that event, it is relatively easy to have sufficient power reserve to maintain the cache memory long enough and have the data dumped into a non-volatile memory such as a specially reserved space in the Flash EEprom memory. In the event of a power down or and power loss to the system, the write cache system may be isolated from the system and a dedicated rechargeable power supply may be switch in only to power the cache system and the reserved space in the Flash EEprom memory.
In the present invention, the Flash EEprom memory array 33 is organized into sectors (typically 512 byte size) such that all memory cells within each sector are erasable together. Thus each sector may be considered to store a data file and a write operation on the memory array acts on one or more such files.
During read of a new sector in the Flash EEprom memory 33, the data file is read out and sent directly to the host through the controller. This file is not used to fill the cache memory 705 as is done in the traditional cache systems.
After the host system has processed the data within a file and wishes to write it back to the Flash EEprom memory 33, it accesses the cache system 701 with a write cycle request. The controller then intercepts this request and acts on the cycle.
In one embodiment of the invention, the data file is written to the cache memory 705. At the same time, two other pieces of information about the data file are written to a tag memory 709. The first is a file pointer which identifies the file present in the cache memory 705. The second is a time stamp that tells what time the file was last written into the cache memory. In this way, each time the host wishes to write to the Flash EEprom memory 33, the data file is actually first stored in the cache memory 705 along with pointers and time stamps in the tag memory 709.
In another embodiment of the invention, when a write from the host occurs, the controller first checks to see if that file already existed in the cache memory 705 or has been tagged in the tag memory 709. If it has not been tagged, the file is written to the Flash memory 33, while its identifier and time stamp are written to the tag memory 709. If the file already is present in the cache memory or has been tagged, it is updated in the cache memory and not written to the Flash memory. In this way only infrequently used data files are written into the Flash memory while frequently used data files are trapped in the cache memory.
In yet another embodiment of the invention, when a write from the host occurs, the controller first checks to see if that data file has been last written anywhere within a predetermined period of time (for example, 5 minutes). If it has not, the data file is written to the Flash memory 33, while its identifier and time stamp are written to the tag memory 709. If the data file has been last written within the predetermined period of time, it is written into the cache memory 705 and not written to the Flash memory. At the same time, its identifier and time stamp are written to the tag memory 709 as in the other embodiments. In this way also, only infrequently used data files are written into the Flash memory while frequently used data files are trapped in the cache memory.
In all embodiments, over time the cache memory 705 will start to fill up. When the controller has detected that some predetermined state of fullness has been reached, it begins to archive preferentially some files over others in the cache memory 705 by writing them to the Flash memory 33.
In either embodiments, over time the cache memory 705 will start to fill up. When the controller has detected that some predetermined state of fullness has been reached, it begins to archive preferentially some files over others in the cache memory 705 by writing them to the Flash memory 33. The file identifier tag bits for these files are then reset, indicating that these files may be written over. This makes room for new data files entering the cache memory.
The controller is responsible for first moving the least active files back into the Flash memory 33 to make room for new active files. To keep track of each file's activity level, the time stamp for each file is incremented by the controller at every time step unless reset by a new activity of the file. The timing is provided by timers 711. At every time step (count), the controller systematically accesses each data file in the cache memory and reads the last time stamp written for this data file. The controller then increments the time stamp by another time step (i.e. increments the count by one).
Two things can happen to a file's time stamp, depending on the activity of the file. One possibility is for the time stamp to be reset in the event of a new activity occurring. The other possibility is that no new activity occurs for the file and the time stamp continues to increment until the file is removed from the cache. In practice a maximum limit may be reached if the time stamp is allowed to increase indefinitely. For example, the system may allow the time stamp to increment to a maximum period of inactivity of 5 minutes. Thus, when a data file is written in the cache memory, the time stamp for the file is set at its initial value. Then the time stamp will start to age, incrementing at every time step unless reset to its initial value again by another write update. After say, 5 minutes of inactivity, the time stamp has incremented to a maximum terminal count.
In one embodiment of keeping count, a bit can be shifted one place in a shift register each time a count increment for a file occurs. If the file is updated (a new activity has occurred) the bit's location will be reset to the initial location the shift register. On the other hand, if the file remains inactive the bit will eventually be shifted to the terminal shift position. In another embodiment, a count value for each file is stored and incremented at each time step. After each increment, the count value is compared to a master counter, the difference being the time delay in question.
Thus, if a file is active its incremented time stamp is reset back to the initial value each time the data file is rewritten. In this manner, files that are constantly updated will have low time stamp identifiers and will be kept in cache until their activity decreases. After a period of inactivity has expired, they acquire the maximum time stamp identifiers. The inactive files are eventually archived to the Flash memory freeing space in the cache memory for new, more active files. Space is also freed up in the tag memory when these inactive files are moved to the Flash memory.
At any time when room must be made available for new data files coming into the cache memory, the controller removes some of the older files and archives them to the Flash memory 33. Scheduling is done by a memory timing/control circuit 713 in the controller. The decision to archive the file is based on several criteria. The controller looks at the frequency of writes occurring in the system and looks at how full the cache is. If there is still room in the cache, no archiving need to be done. If more room is needed, the files with the earliest time stamps are first removed and archived to the Flash memory.
Although the invention has been described with implementation in hardware in the controller, it is to be understood that other implementations are possible. For example, the cache system may be located elsewhere in the system, or be implemented by software using the existing microprocessor system. Such variations are within the scope of protection for the present invention.
The Profile of how often data is written back to the Flash memory is determined by several factors. It depends on the size of the cache memory and the frequency of writes occurring in the system. With a small cache memory system, only the highest frequency files will be cached. Less frequently accessed files will also be cached with increasing cache memory size. In the present invention, a relatively cheap and small amount of cache memory, preferably about 1 Mbyte, may be used to good advantage. By not constantly writing the most active files (the top 5%), the write frequency of the Flash EEprom may be reduced from the usual one every millisecond to one every 5 minutes. In this way the wear-out time for the memory can be extended almost indefinitely. This improvement is also accompanied by increased system performance during write.
Incorporating time tag into the write cache concept has the advantage that the size of the write cache buffer memory can be relatively small, since it is used only to store frequently written data files, with all other files written directly into the Flash EEprom memory. A second advantage is that the management of moving data files in and out of the write cache buffer can be automated since it does not require advanced knowledge of which data files are to be called next.
The sequence for a complete erase cycle of the new algorithm is as follows (see
1. Read S. This value can be stored in a register file. (This step can be omitted if S is not expected to approach the endurance limit during the operating lifetime of the device).
1a. Apply a first erase pulse with VERASE=V1+nΔV, n=0, pulse duration=t. This pulse (and the next few successive pulses) is insufficient to fully erase all memory cells, but it serves to reduce the charge Q on programmed cells at a relatively low erase field stress, i.e., it is equivalent to a “conditioning” pulse.
1b. Read a sparse pattern of cells in the array. A diagonal read pattern for example will read m+n cells (rather than m×n cells for a complete read) and will have at least one cell from each row and one cell from each column in the array. The number N of cells not fully erased to state “3” is counted and compared with X.
1c. If N is greater than x (array not adequately erased) a second erase pulse is applied of magnitude greater by ΔV than the magnitude of the first pulse, with the same pulse duration, t. Read diagonal cells, count N.
This cycling of erase pulse/read/increment erase pulse is continued until either N≦X or the number n of erase pulses exceed nmax. The first one of these two conditions to occur leads to a final erase pulse.
2a. The final erase pulse is applied to assure that the array is solidly and fully erased. The magnitude of VERASE can be the same as in the previous pulse or higher by another increment ΔV. The duration can be between 1t and 5t.
2b. 100% of the array is read. The number N of cells not fully erased is counted. If N is less than or equal to X, then the erase pulsing is completed at this point.
2c. If N is greater than X, then address locations of the N unerased bits are generated, possibly for substitution with redundant good bits at the system level. If N is significantly larger than X (for example, if N represents perhaps 5% of the total number of cells), then a flag may be raised, to indicate to the user that the array may have reached its endurance end of life.
2d. Erase pulsing is ended.
3a. S is incremented by one and the new S is stored for future reference. This step is optional. The new S can be stored either by writing it into the newly erased block or off chip in a separate register file.
3b. The erase cycle is ended. The complete cycle is expected to be completed with between 10 to 20 erase pulses and to last a total of approximately one second.
The new algorithm has the following advantages:
(a) No cell in the array experiences the peak electric field stress. By the time VERASE is incremented to a relatively high voltage any charge Q on the floating gates has already been removed in previous lower voltage erase pulses.
(b) The total erase time is significantly shorter than the fixed VERASE pulse of the prior art. Virgin devices see the minimum pulse duration necessary to erase. Devices which have undergone more than 1×104 cycles require only several more ΔV voltage increments to overcome dielectric trapped charge, which only adds several hundred milliseconds to their total erase time.
(c) The window closure on the erase side (see curve (b) in FIG. 11d of U.S. Pat. No. 5,095,344) is avoided indefinitely (until the device experiences failure by a catastrophic breakdown) because VERASE is simply incremented until the device is erased properly to state “3”. Thus, the new erase algorithm preserves the full memory window.
Flash EEprom Memory Array Implementations
The Flash EEprom cells of this invention can be implemented in dense memory arrays in several different array architectures. The first architecture, shown in
A second Flash EEprom memory array architecture which lends itself to better packing density than the array of
The array can be erased in a block, or in entire rows by decoding the erase voltage to the corresponding erase lines.
The various aspects of the present invention that have been described co-operate in a system of Flash EEprom memory array to make the Flash EEprom memory a viable alternative to conventional non-volatile mass storage devices.
While the embodiments of the various aspects of the present invention that have been described are the preferred implementation, those skilled in the art will understand that variations thereof may also be possible. Therefore, the invention is entitled to protection within the full scope of the appended claims.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US3778776||Jun 10, 1971||Dec 11, 1973||Nippon Electric Co||Electronic computer comprising a plurality of general purpose registers and having a dynamic relocation capability|
|US4037209||Nov 24, 1975||Jul 19, 1977||Fujitsu Ltd.||Data processing system for converting from logical addresses to physical addresses|
|US4051354||Jul 3, 1975||Sep 27, 1977||Texas Instruments Incorporated||Fault-tolerant cell addressable array|
|US4066880||Mar 30, 1976||Jan 3, 1978||Engineered Systems, Inc.||System for pretesting electronic memory locations and automatically identifying faulty memory sections|
|US4087795||Dec 15, 1976||May 2, 1978||Siemens Aktiengesellschaft||Memory field effect storage device|
|US4181980||May 15, 1978||Jan 1, 1980||Electronic Arrays, Inc.||Acquisition and storage of analog signals|
|US4188662||Apr 14, 1977||Feb 12, 1980||Fujitsu Limited||Address converter in a data processing apparatus|
|US4279024||Jun 12, 1979||Jul 14, 1981||Siemens Aktiengesellschaft||Word-by-word electrically reprogrammable nonvolatile memory|
|US4287570||Jun 1, 1979||Sep 1, 1981||Intel Corporation||Multiple bit read-only memory cell and its sense amplifier|
|US4357685||Jul 14, 1980||Nov 2, 1982||Sgs-Ates Componenti Elettronici S.P.A.||Method of programming an electrically alterable nonvolatile memory|
|US4365318||Sep 15, 1980||Dec 21, 1982||International Business Machines Corp.||Two speed recirculating memory system using partially good components|
|US4422161||Oct 8, 1981||Dec 20, 1983||Rca Corporation||Memory array with redundant elements|
|US4430727||Nov 10, 1981||Feb 7, 1984||International Business Machines Corp.||Storage element reconfiguration|
|US4448400||Mar 8, 1982||May 15, 1984||Eliyahou Harari||Highly scalable dynamic RAM cell with self-signal amplification|
|US4459661||Apr 21, 1982||Jul 10, 1984||Fujitsu Limited||Channel address control system for a virtual machine system|
|US4460982||May 20, 1982||Jul 17, 1984||Intel Corporation||Intelligent electrically programmable and electrically erasable ROM|
|US4466055||Mar 17, 1981||Aug 14, 1984||Tokyo Shibaura Denki Kabushiki Kaisha||Information processing system including a one-chip arithmetic control unit|
|US4475194 *||Mar 30, 1982||Oct 2, 1984||International Business Machines Corporation||Dynamic replacement of defective memory words|
|US4482952||Dec 8, 1981||Nov 13, 1984||Nippon Electric Co., Ltd.||Virtual addressing system using page field comparisons to selectively validate cache buffer data on read main memory data|
|US4493075||May 17, 1982||Jan 8, 1985||National Semiconductor Corporation||Self repairing bulk memory|
|US4525839||Oct 26, 1982||Jun 25, 1985||Hitachi, Ltd.||Method of controlling storage device|
|US4527251||Dec 17, 1982||Jul 2, 1985||Honeywell Information Systems Inc.||Remap method and apparatus for a memory system which uses partially good memory devices|
|US4528683||May 25, 1982||Jul 9, 1985||Vdo Adolf Schindling Ag||Circuit for storing a multi-digit decimal numerical value of the distance traversed by a vehicle|
|US4562532||May 24, 1982||Dec 31, 1985||Fujitsu Limited||Main storage configuration control system|
|US4608671||May 2, 1983||Aug 26, 1986||Hitachi, Ltd.||Buffer storage including a swapping circuit|
|US4612640||Feb 21, 1984||Sep 16, 1986||Seeq Technology, Inc.||Error checking and correction circuitry for use with an electrically-programmable and electrically-erasable memory array|
|US4616311||Apr 29, 1985||Oct 7, 1986||Tokyo Shibaura Denki Kabushiki Kaisha||Data processing system|
|US4617651||Feb 22, 1984||Oct 14, 1986||Seeq Technology, Inc.||Redundancy circuit for use in a semiconductor memory array|
|US4638457||May 28, 1982||Jan 20, 1987||Siemens Aktiengesellschaft||Method and apparatus for the non-volatile storage of the count of an electronic counting circuit|
|US4652897||Jul 15, 1985||Mar 24, 1987||Hitachi, Ltd.||Semiconductor memory device|
|US4663770||Feb 24, 1986||May 5, 1987||Hughes Microlectronics Limited||Non-volatile counting circuit and method which provides for extended counter life|
|US4667217||Apr 19, 1985||May 19, 1987||Ncr Corporation||Two bit vertically/horizontally integrated memory cell|
|US4682287||Aug 28, 1984||Jul 21, 1987||Nippondenso Co., Ltd.||Electronic odometer|
|US4718041||Jan 9, 1986||Jan 5, 1988||Texas Instruments Incorporated||EEPROM memory having extended life|
|US4733394||Apr 23, 1986||Mar 22, 1988||Deutsche Itt Industries Gmbh||Electrically programmable semiconductor memory showing redundance|
|US4757474||Jan 21, 1987||Jul 12, 1988||Fujitsu Limited||Semiconductor memory device having redundancy circuit portion|
|US4763305||Nov 27, 1985||Aug 9, 1988||Motorola, Inc.||Intelligent write in an EEPROM with data and erase check|
|US4774652||Feb 18, 1987||Sep 27, 1988||Apple Computer, Inc.||Memory mapping unit for decoding address signals|
|US4774700||Mar 7, 1986||Sep 27, 1988||Matsushita Electric Industrial Co., Ltd.||Information recording and reproducing apparatus with detection and management of defective sectors|
|US4800520||Oct 23, 1986||Jan 24, 1989||Kabushiki Kaisha Toshiba||Portable electronic device with garbage collection function|
|US4803707||Dec 21, 1987||Feb 7, 1989||Ncr Corporation||Nonvolatile electronic odometer with excess write cycle protection|
|US4819211||Jun 25, 1987||Apr 4, 1989||Hitachi, Ltd.||Microcomputer system for high speed address translation|
|US4841482||Feb 17, 1988||Jun 20, 1989||Intel Corporation||Leakage verification for flash EPROM|
|US4860228||Feb 24, 1987||Aug 22, 1989||Motorola, Inc.||Non-volatile memory incremental counting system|
|US4870621 *||Nov 27, 1987||Sep 26, 1989||Nec Corporation||Dual port memory device with improved serial access scheme|
|US4887234||Jan 13, 1989||Dec 12, 1989||Kabushiki Kaisha Toshiba||Portable electronic device with plural memory areas|
|US4888686||Dec 19, 1986||Dec 19, 1989||Vdo Adolf Schindling Ag||System for storing information with comparison of stored data values|
|US4896262||Feb 22, 1985||Jan 23, 1990||Kabushiki Kaisha Meidensha||Emulation device for converting magnetic disc memory mode signal from computer into semiconductor memory access mode signal for semiconductor memory|
|US4931997||Feb 23, 1988||Jun 5, 1990||Hitachi Ltd.||Semiconductor memory having storage buffer to save control data during bulk erase|
|US4942556||Jul 10, 1989||Jul 17, 1990||Hitachi, Ltd.||Semiconductor memory device|
|US4949240||Mar 7, 1988||Aug 14, 1990||Kabushiki Kaisha Toshiba||Data storage system having circuitry for dividing received data into sequential wards each stored in storage region identified by chain data|
|US4989181 *||May 30, 1989||Jan 29, 1991||Nec Corporation||Serial memory device provided with high-speed address control circuit|
|US4992984||Dec 28, 1989||Feb 12, 1991||International Business Machines Corporation||Memory module utilizing partially defective memory chips|
|US5016215||Mar 12, 1990||May 14, 1991||Texas Instruments Incorporated||High speed EPROM with reverse polarity voltages applied to source and drain regions during reading and writing|
|US5033990||Jul 23, 1990||Jul 23, 1991||Arthur Silverman||Pulley having spring loaded release mechanism|
|US5043940||Jul 17, 1989||Aug 27, 1991||Eliyahou Harari||Flash EEPROM memory systems having multistate storage cells|
|US5051994||Apr 28, 1989||Sep 24, 1991||International Business Machines Corporation||Computer memory module|
|US5053990||Feb 17, 1988||Oct 1, 1991||Intel Corporation||Program/erase selection for flash memory|
|US5070502||Jun 23, 1989||Dec 3, 1991||Digital Equipment Corporation||Defect tolerant set associative cache|
|US5095344 *||Jun 8, 1988||Mar 10, 1992||Eliyahou Harari||Highly compact eprom and flash eeprom devices|
|US5126973||Feb 14, 1990||Jun 30, 1992||Texas Instruments Incorporated||Redundancy scheme for eliminating defects in a memory device|
|US5134584||Jul 22, 1988||Jul 28, 1992||Vtc Incorporated||Reconfigurable memory|
|US5163021||Jul 22, 1991||Nov 10, 1992||Sundisk Corporation||Multi-state EEprom read and write circuits and techniques|
|US5172338||Apr 11, 1990||Dec 15, 1992||Sundisk Corporation||Multi-state EEprom read and write circuits and techniques|
|US5210716||Jul 23, 1990||May 11, 1993||Seiko Instruments Inc.||Semiconductor nonvolatile memory|
|US5214657||Jun 30, 1992||May 25, 1993||Micron Technology, Inc.||Method for fabricating wafer-scale integration wafers and method for utilizing defective wafer-scale integration wafers|
|US5222109||Dec 28, 1990||Jun 22, 1993||Ibm Corporation||Endurance management for solid state files|
|US5243570||Feb 1, 1993||Sep 7, 1993||Nec Corporation||Semiconductor memory device having redundant memory cell columns concurrently accessible together with regular memory cell arrays|
|US5268870||Aug 6, 1990||Dec 7, 1993||Eliyahou Harari||Flash EEPROM system and intelligent programming and erasing methods therefor|
|US5270979||Mar 15, 1991||Dec 14, 1993||Sundisk Corporation||Method for optimum erasing of EEPROM|
|US5278793||Feb 25, 1992||Jan 11, 1994||Yeh Tsuei Chi||Memory defect masking device|
|US5293560||Nov 3, 1992||Mar 8, 1994||Eliyahou Harari||Multi-state flash EEPROM system using incremental programing and erasing methods|
|US5297103||Jan 19, 1993||Mar 22, 1994||Fujitsu Limited||Electrically erasable and programmable semiconductor memory device|
|US5297148 *||Oct 20, 1992||Mar 22, 1994||Sundisk Corporation||Flash eeprom system|
|US5369615||Nov 8, 1993||Nov 29, 1994||Sundisk Corporation||Method for optimum erasing of EEPROM|
|US5377148||Nov 29, 1990||Dec 27, 1994||Case Western Reserve University||Apparatus and method to test random access memories for a plurality of possible types of faults|
|US5396468||Nov 8, 1993||Mar 7, 1995||Sundisk Corporation||Streamlined write operation for EEPROM system|
|US5402376||Aug 10, 1993||Mar 28, 1995||Hitachi, Ltd.||Semiconductor memory having redundancy circuit|
|US5414664||May 28, 1993||May 9, 1995||Macronix International Co., Ltd.||Flash EPROM with block erase flags for over-erase protection|
|US5465234||Nov 23, 1994||Nov 7, 1995||Nec Corporation||Semiconductor memory device having shifting circuit connected between data bus lines and data buffer circuits for changing connections therebetween|
|US5471431||Apr 22, 1993||Nov 28, 1995||Sgs-Thomson Microelectronics, Inc.||Structure to recover a portion of a partially functional embedded memory|
|US5535328||Feb 23, 1995||Jul 9, 1996||Sandisk Corporation||Non-volatile memory system card with flash erasable sectors of EEprom cells including a mechanism for substituting defective cells|
|US5546402 *||Jun 7, 1995||Aug 13, 1996||International Business Machines Corporation||Flash-erase-type nonvolatile semiconductor storage device|
|US5563832||Oct 28, 1994||Oct 8, 1996||Nec Corporation||Semiconductor memory device having interface units memorizing available memory cell sub-arrays|
|US5568439||Jun 6, 1995||Oct 22, 1996||Harari; Eliyahou||Flash EEPROM system which maintains individual memory block cycle counts|
|US5594689||Mar 11, 1996||Jan 14, 1997||Nec Corporation||Non-volatile semiconductor memory capable of erase- verifying memory cells in a test mode using a defective count circuit activated by a test mode signal|
|US5602987 *||Dec 29, 1993||Feb 11, 1997||Sandisk Corporation||Flash EEprom system|
|US5668763||Feb 26, 1996||Sep 16, 1997||Fujitsu Limited||Semiconductor memory for increasing the number of half good memories by selecting and using good memory blocks|
|US5689676 *||Nov 18, 1996||Nov 18, 1997||Sony Corporation||Sequential EEPROM writing apparatus which sequentially and repetitively replaces a head position pointer with a last position pointer|
|US5818754 *||Dec 27, 1996||Oct 6, 1998||Nec Corporation||Nonvolatile memory having data storing area and attribute data area for storing attribute data of data storing area|
|US5991517 *||Dec 20, 1996||Nov 23, 1999||Sandisk Corporation||Flash EEprom system with cell by cell programming verification|
|DE3200872A1||Jan 14, 1982||Jul 21, 1983||Sartorius Gmbh||Electronic balance|
|EP0251889A1||Jun 23, 1987||Jan 7, 1988||Sgs-Thomson Microelectronics S.A.||Method for programming data into an electrically programmable read-only-memory|
|EP0349775A2||Jun 6, 1989||Jan 10, 1990||Eliyahou Harari||Flash eeprom memory systems and methods of using them|
|FR2430065A1||Title not available|
|GB2215155A||Title not available|
|JP5189981B2||Title not available|
|JP54158141U||Title not available|
|JP58060493U||Title not available|
|JP58086777U||Title not available|
|JP58215794A||Title not available|
|JP58215795A||Title not available|
|JP59045695U||Title not available|
|JP59162695A||Title not available|
|JP60076097U||Title not available|
|JP60212900A||Title not available|
|JP61096598U||Title not available|
|JP62229598A||Title not available|
|JP62245600A||Title not available|
|JP62283496A||Title not available|
|JP62283497A||Title not available|
|JP63124298A||Title not available|
|JP63183700A||Title not available|
|JP63200398A||Title not available|
|JP63308797A||Title not available|
|JP363136259A||Title not available|
|JPS5860493A||Title not available|
|JPS5886777A||Title not available|
|JPS5945695A||Title not available|
|JPS6076097A||Title not available|
|JPS6196598A||Title not available|
|JPS54158141A||Title not available|
|JPS58215794A||Title not available|
|JPS58215795A||Title not available|
|JPS59144098A||Title not available|
|JPS59162695A||Title not available|
|JPS60212900A||Title not available|
|JPS61184795A||Title not available|
|JPS62229598A||Title not available|
|1||"Notification of Reasons for Refusal", Application No. H11(1999) 023842, Japan Patent Office, Feb. 22, 2005, 4 pages.|
|2||Alberts et al., "Multi-Bit Storage Fet Earom Cell", IBM Technical Disclosure Bulletin, vol. 24, No. 7A, Dec. 1981, pp. 3311-3314, 4193.|
|3||Berenga et al."E2-PROM TV Synthesizer", 1978 IEEE International Solid-State Circuits Conference: ISCCC 78, Feb. 17, 1978, pp. 196-197.|
|4||Bleiker et al., "A Four State EEPROM using Floating-Gate Memory Cells", IEEE, vol. SC-22, No. 3, Jounal of Solid-State Circuits, Jun. 1987, pp. 460-463.|
|5||Decision of Refusal, Japanese Patent Office, Japanese Divisional Patent Application No. 023842/1999 for SanDisk Corporation, Sep. 20, 2005, 5 pages.|
|6||EPO/Communication Pursuant to Article 96(2) mailed Oct. 1, 2004 in connection with EPO application No. 00 107 163.8 2210 received Jan. 18, 2005.|
|7||Furuyama et al, "An Experimental 2-Bit/Cell Storage Dram for Macro Cell or Memory-on-Logic Application", IEEE Custom Integrated Circuits Conference, May 1988, pp. 4.4.1-4.4.4.|
|8||Harold, "Production EPROM Loading", New Electronics, vol. 15 (1982) Feb. 1982, No. 3, London, Great Britain, pp. 46-50.|
|9||Listing of Pending Claims for Japanese Divisional Patent Application No. 023842/1999, 6 pages.|
|10||Lucerno et al."A 16 kbit Smart 5 V-only EEPROM with Redundancy", IEEE Journal of Solid-State Circuits, vol. SC-18, No. 5, pp. 539-543, Oct. 1983.|
|11||Office Action for U.S. Appl. No. 09/064,250 mailed Nov. 10, 2005, 15 pages.|
|12||Office Action for U.S. Appl. No. 10/414,718 mailed Jan. 5, 2005, 16 pages.|
|13||Stark Moshe, "Two Bits Per Cell ROM", Digest of Papers of Spring COMPCONS 81, Feb. 23-26, VLSI Laboratory, pp. 209-212.|
|14||Torelli et al, "An Improved Method for Programming a Word-Erasable EEPROM", Alta Frequenza, vol. 52, Nov.-Dec. 1983, No. 6, pp. 487-494.|
|15||Torelli, Guido "An LSI Technology Fully Compatible EAROM Cell", S-ATES Componenti Elettronici SpA, No. 6, vol. LI, 7 pp., 1982.|
|16||Yasuo Tarui, "Japanese Develop Nondestructive Analog Semiconductor Memory", Electronics Review, pp. 29-30, Jul. 11, 1974.|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US8886853 *||Sep 16, 2013||Nov 11, 2014||Commvault Systems, Inc.||Systems and methods for uniquely identifying removable media by its manufacturing defects wherein defects includes bad memory or redundant cells or both|
|US8924428||Dec 21, 2012||Dec 30, 2014||Commvault Systems, Inc.||Systems and methods of media management, such as management of media to and from a media storage library|
|US8996823||Dec 23, 2013||Mar 31, 2015||Commvault Systems, Inc.||Parallel access virtual tape library and drives|
|US9201917||Sep 19, 2014||Dec 1, 2015||Commvault Systems, Inc.||Systems and methods for performing storage operations in a computer network|
|US9244779||Sep 23, 2011||Jan 26, 2016||Commvault Systems, Inc.||Data recovery operations, such as recovery from modified network data management protocol data|
|US9251190||Mar 26, 2015||Feb 2, 2016||Commvault Systems, Inc.||System and method for sharing media in a computer network|
|US20140019647 *||Sep 16, 2013||Jan 16, 2014||Commvault Systems, Inc.||Systems and methods of media management, such as management of media to and from a media storage library, including removable media|
|U.S. Classification||365/185.09, 365/185.29, 365/185.11|
|International Classification||G11C16/06, G06F11/10, G06F12/12, G11C29/00, G06F12/02, G06F3/06, G11C16/16, G11C16/34, G11C29/52, G11C29/34, G06F12/08, G11C11/56, G11C29/26|
|Cooperative Classification||G11C16/16, G11C29/765, G11C2211/5634, G11C11/5621, G11C16/3454, G11C29/82, G11C7/1039, G11C16/3445, G06F3/0679, G11C29/34, G11C29/26, G11C29/52, G11C16/344, G06F11/1068, G11C2211/5643, G11C2211/5621, G11C11/5635, G06F3/0652, G11C11/5628, G11C16/3459, G06F2212/312, G06F12/123, G06F12/0246, G06F3/0616, G06F12/0804, G11C29/00, G11C16/3436|
|European Classification||G11C11/56D2E, G11C16/16, G11C29/26, G11C16/34V2, G11C16/34V, G06F12/02D2E2, G11C11/56D, G11C29/52, G11C16/34V4, G11C11/56D2, G11C29/765, G11C29/82, G11C7/10M5, G06F3/06A6L2F, G06F3/06A2R2, G06F3/06A4H6, G11C16/34V2C, G11C16/34V4C|
|May 29, 2015||REMI||Maintenance fee reminder mailed|
|Oct 18, 2015||LAPS||Lapse for failure to pay maintenance fees|
|Dec 8, 2015||FP||Expired due to failure to pay maintenance fee|
Effective date: 20151018