US20010001872A1 - Data caching with a partially compressed cache - Google Patents

Data caching with a partially compressed cache Download PDF

Info

Publication number
US20010001872A1
US20010001872A1 US09/090,028 US9002898A US2001001872A1 US 20010001872 A1 US20010001872 A1 US 20010001872A1 US 9002898 A US9002898 A US 9002898A US 2001001872 A1 US2001001872 A1 US 2001001872A1
Authority
US
United States
Prior art keywords
data
partition
cache
compressed
storage
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US09/090,028
Other versions
US6324621B2 (en
Inventor
Shanker Singh
Joe-Ming Cheng
Brent Cameron Beardsley
Dell Patrick Leabo
Forrest Lee Wade
Michael Thomas Benhase
Marc Ethan Goldfeder
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US09/090,028 priority Critical patent/US6324621B2/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SINGH, SHANKER, CHENG, JOE-MING, BEARDSLEY, BRENT CAMERON, GOLDFEDER, MARC ETHAN, WADE, FORREST LEE, BENHASE, MICHAEL THOMAS, LEABO, DELL PATRICK
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SINGH, SHANKER, CHENG, JOE-MING, BEARDSLEY, BRENT CAMERON, GOLDFEDER, MARC ETHAN, WADE, FORREST LEE, BENHASE, MICHAEL THOMAS, LEABO, DELL PATRICK
Publication of US20010001872A1 publication Critical patent/US20010001872A1/en
Application granted granted Critical
Publication of US6324621B2 publication Critical patent/US6324621B2/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0893Caches characterised by their organisation or structure
    • G06F12/0897Caches characterised by their organisation or structure with two or more cache hierarchy levels
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/12Replacement control
    • G06F12/121Replacement control using replacement algorithms
    • G06F12/126Replacement control using replacement algorithms with special data handling, e.g. priority of data or instructions, handling errors or pinning
    • G06F12/127Replacement control using replacement algorithms with special data handling, e.g. priority of data or instructions, handling errors or pinning using additional replacement algorithms
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0866Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches for peripheral storage systems, e.g. disk cache
    • G06F12/0871Allocation or management of cache space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/40Specific encoding of data in memory or cache
    • G06F2212/401Compressed data

Definitions

  • the present invention relates generally to storage caches, and more particularly to partially compressed storage caches.
  • one typical method of improving performance by increasing the hit ratio involves expanding the size of the cache.
  • an equivalent increase in the hit ratio percentage is not always achieved. For example, doubling a 4 GB (gigabytes) cache with a 75% hit ratio to 8 GB does not result in a doubling of the hit ratio. While a small percentage of improvement in the hit ratio occurs, the doubling in size results in considerable cost expenses.
  • a method aspect for caching storage data includes partitioning a storage cache to include a compressed data partition and an uncompressed data partition, and adjusting the compressed data partition and the uncompressed data partition for chosen performance characteristics, including overall cache size.
  • a data caching system aspect in a data processing system having a host system in communication with a storage system includes at least one storage device and at least one partially compressed cache.
  • the at least one partially compressed cache further includes an uncompressed partition and a compressed partition, where the compressed partition stores at least a victim data unit from the uncompressed partition.
  • FIG. 1 illustrates a block diagram of a data processing system arrangement that utilizes caching in accordance with the present invention.
  • FIGS. 2 and 3 present performance modeling results that illustrate response time changes and maximum IO values with varying levels of compression of the storage cache of FIG. 1.
  • FIG. 4 illustrates an alternate arrangement of the system of FIG. 1 that realizes significant bandwidth advantages through the use of partial cache compression.
  • FIG. 5( a ) illustrates a block diagram of typical multi-level caching.
  • FIG. 5( b ) illustrates a multi-level caching arrangement in accordance with a further aspect of the present invention.
  • the present invention relates to partially compressed data caching.
  • the following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements.
  • Various modifications to the preferred embodiment will be readily apparent to those skilled in the art and the generic principles herein may be applied to other embodiments.
  • the present invention is not intended to be limited to the embodiment shown but is to be accorded the widest scope consistent with the principles and features described herein.
  • FIG. 1 illustrates a block diagram of a data processing system arrangement that utilizes caching in accordance with the present invention.
  • a host system 10 e.g., a central processing unit (CPU), storage cache 12 , and storage (DASD, direct access storage device) 14 .
  • storage cache 12 is used to provide quicker access to storage data than storage 14 , e.g., on the order of microseconds per access, rather than milliseconds.
  • the storage cache 12 in accordance with the present invention optimizes cache organization and algorithms dynamically to achieve maximum performance without requiring additional storage cache. It should be appreciated that although the discussion focuses primarily on storage controller cache applications, the invention is equally applicable to CPU cache and other caching applications, as is well appreciated by those skilled in the art.
  • storage cache 12 is partitioned into at least two partitions, partition L 3 and partition L 4 .
  • one partition e.g., partition L 3
  • a second partition e.g., partition L 4
  • a compressor unit 16 and decompressor unit 18 are also included in the system of FIG. 1.
  • partitions L 3 and L 4 are shown as distinct partitions in the storage cache 12 .
  • the partitions L 3 and L 4 are part of the same physical memory, so that compressed and uncompressed data are intermingled in the storage cache 12 .
  • a “line” suitably refers to a line cache replacement unit (or a replaceable data unit) in a cache.
  • a cache line suitably refers to a storage device 14 record, track, or a fraction of a track.
  • a L 3 and L 4 cache directory 19 tracks the status of compressed and uncompressed locations/data blocks within the storage cache 12 by storing a listing of the uncompressed blocks and a listing of compressed blocks, with the formation and updating of the directory achieved according to methods well understood by those skilled in the art.
  • the listings could be provided as two separate directories, and a third directory (not shown) may also be included to store a listing of empty blocks, if desired.
  • partition L 3 stores data that is changing.
  • a least recently used (LRU) algorithm is employed to make room for the new data by removing the least recently used data, which takes advantage of the storage access temporal locality and determines which uncompressed line becomes the removed data/victim line, as is well-known to those skilled in the art.
  • the victim data unit i.e., victim line of a record, track, or fraction of a track, of partition L 3 is compressed through compressor unit 16 and written to the same location in the storage cache 12 .
  • the original uncompressed location now stores the compressed line and opens two empty compressed blocks in the storage cache 12 .
  • Directory 19 is suitably updated to indicate the change in status of the block from uncompressed (i.e., part of partition L 3 ) to compressed (i.e., part of partition L 4 ).
  • the uncompressed victim line is also preferably written/destaged to storage 14 . If L 3 and L 4 are provided as separate physical memory devices, and the L 4 physical distance and interface is different than L 3 , then compressed data to L 4 is physically moved to the empty slot in L 4 , and the directory of L 4 stored in L 3 or L 4 is appropriately updated.
  • a compression sniffer 20 is useful for trail compression for a data line which has been seen once.
  • a sniffer is a device and/or software routine that monitors data lines as they are transferred through the caching system, i.e., that monitors the ‘trail’ the data lines travel, to identify how the data lines are compressed.
  • the compression sniffer/trail compression unit 20 suitably gives actual compressed size.
  • alternative methods for estimating space allocation include: assigning enough space to hold the worst case expanded data (e.g., 1.125 times the line size), storing the compressed data, and reclaiming the left over space; allocating space based on a priori compression statistics, e.g., using 2 sigma the compression statistics (or other threshold) for allocation and keeping the overflow if compressed size is exceeded; or assigning the threshold (e.g., about 3:1) and merely dropping the compressed data if it does not fit (with the assumption that the data is destaged to DASD).
  • Partition L 4 suitably uses a LRU replacement policy and compression weighted replacement (CWR) to determine its victim lines, which takes into account that data that is directly interpretable by human and machine usually compresses well and that data that can be compressed moderately well has a higher chance of being reused than data that does not compress well.
  • the victim line of partition L 4 may suitably be discarded/removed via a bit discarder mechanism 24 , i.e., a ‘bit bucket’ or universal data sink, to make space in the storage cache 12 when new data is written to partition L 4 .
  • a bit discarder mechanism 24 i.e., a ‘bit bucket’ or universal data sink
  • partition L 4 When a read attempt results in a read miss in partition L 3 with a read hit in partition L 4 , partition L 4 suitably passes the uncompressed line (decompressed through decompressor unit 18 ) to partition L 3 and to the host system 10 .
  • the partition L 4 entry for that line suitably becomes invalidated.
  • a miss occurs in both partitions L 3 and L 4 , suitably the requested data is read back from storage 14 to the partition L 3 for transfer to host system 10 .
  • the source block size is 512 Bytes or 1024 Bytes and a cache line, i.e., 2K or 4K Bytes, track, or half track, consists of multiple blocks, with a read or modify write to a specific block in a cache line done through the partition L 3 .
  • a performance penalty may result if a read or modify write occurs to a specific block that is kept in the partition L 4 .
  • the primary impacts are the extra reads and writes between L 3 and L 4 which consumes additional data bandwidth.
  • the cache line contains only a single block.
  • a block size of 2KB allows a good compression region (i.e., about 3.0), since a block size of 512 B usually achieves less than 2.0 compression.
  • the compression ratio i.e., the ratio of uncompressed size/compressed size
  • the compression factor acts as a discrimination factor for cache replacement (the term discrimination factor being known commonly in Pattern Recognition (PR) research), with the objectives of the caching algorithm being a classification process that keeps the class of most useful data in the cache through some discrimination process (as less well-compressed data usually is used less often).
  • PR Pattern Recognition
  • partitions L 3 and L 4 are adjusted according to cache size, storage device size, compression/decompression hardware speed, and performance objectives, e.g., access time, response time, maximum rate of input and output required (MAX IO requirements), etc.
  • the storage cache 12 of the present invention realizes significant advantages over straight caching.
  • FIGS. 2 and 3 present performance modeling results that illustrate response time changes and maximum IO values as adjustments are made to the percentage of the storage cache 12 that is compressed.
  • the performance model of FIG. 2 suitably illustrates a storage cache 12 of size 4 GB
  • the performance model of FIG. 3 suitably illustrates a storage cache 12 of size 8 GB, both used with a storage 14 of 360 GB and a compression ratio of about 3:1.
  • the base response time (BASE RT) is 2.9 ms (milliseconds) with a maximum IO per second (MAX IO) of 9830.
  • MAX IO maximum IO per second
  • FIG. 2 ( c )) provides better response time and maximum 10 than an 8 GB storage cache with 100% uncompressed data (i.e., FIG. 3, ( a )).
  • the response time for the 4 GB arrangement is 2.6 ms with a maximum IO of 11190
  • the response time for the 8 GB arrangement is 2.7 ms with a maximum of 10770.
  • FIGS. 2 and 3 further reveal that the most beneficial partition percentages are dependent on cache size.
  • a 50% L 3 and 50% L 4 partitioning provides overall optimum performance.
  • a 75% L 3 and 25% L 4 partitioning provides overall optimum performance.
  • a system 50 interfaces with a storage subsystem 52 via a system adapter 54 .
  • the system adapter 54 suitably includes compressor logic 56 and decompressor logic 58 .
  • storage cache 60 and storage 62 e.g., DASD.
  • partition L 3 of storage cache 60 includes a plurality of ‘home areas’, where each ‘home area’ refers to an uncompressed data block, as described with reference to FIG. 1.
  • the system adapter 54 sends compressed data to these home areas at a preferred compression ratio, e.g., more than approximately 2:1.
  • a preferred compression ratio e.g., more than approximately 2:1.
  • the portion of each home area required to store data varies depending upon the amount of data being compressed, of course, but will never exceed the entire home area due to the compression.
  • an effective increase in bandwidth for data transfer results by a factor dependent upon the compression ratio. For example, with a compression ratio of about 3:1, data transfer can now effectively occur at three times the normal transfer rate, since three times as much data is able to be transferred, e.g., a rate of 100 MB/s (megabyte/second) now appears as 300 MB/s.
  • a similar increase in bandwidth may be realized for data transfers from partition L 3 to storage 62 , since the data being transferred from partition L 3 has already been compressed.
  • an LRU algorithm suitably determines which data from partition L 3 becomes a victim line for storage in partition L 4 with writing to storage 62 , and similarly, which lines are victims from partition L 4 .
  • the size of the blocks in partition L 4 are smaller in comparison to the data block storage area/home areas of partition L 3 .
  • a data sniffer 64 suitably determines whether a victim line from L 3 will fit in partition L 4 .
  • Compacting of data may occur in order to realize line transfer from L 3 to L 4 .
  • Data is suitably discarded from partition L 4 via a bit bucket/discarder 65 .
  • the line is returned to partition L 3 and to system adapter 54 .
  • the decompressor logic 58 suitably decompresses the data before being returned to system 50 . If a miss occurs in both partitions L 3 and L 4 , the data is returned from storage 62 to partition L 3 and directly to system adapter 54 for decompression and return to system 50 .
  • FIG. 5( a ) illustrates a block diagram of typical multi-level caching, such as in a RAMAC storage system from IBM Corporation, Armonk, N.Y.
  • An upper level cache 70 is shared among all of the lower level caches 72 , each of which is shared among all of the drawer level caches 74 , each of which is then shared among respective disk level caches 76 .
  • the LRU algorithm is employed at each of the levels.
  • the addition of new data at the upper level cache 70 normally results in replacement of the data in each of the next lower cache levels, as well. Unfortunately, this results in duplication of cache lines, i.e., cache pollution, among the caches.
  • a multi-level caching arrangement utilizes dynamic partitioning, as described with reference to FIG. 5( b ).
  • each level of the multi-level caching arrangement employs dynamic partitioning of uncompressed and compressed data partitions.
  • an upper, level one system cache 80 is partitioned into an uncompressed partition, IL 3 , and compressed partition, IL 4 .
  • a drawer level cache 82 comprises an uncompressed partition IIL 3 , and compressed partition IIL 4
  • a drive cache 84 associated with a DASD 86 includes a compressed partition IIIL 4 and an uncompressed partition IIIL 3 .
  • a compression sniffer 88 is further included and operates in a manner as previously described with reference to FIG. 1 to solve problems associated with compression ratio variability and prediction problems. Sizes of the partition L 3 and partition L 4 of each level are suitably adjusted dynamically to achieve desired performance characteristics, as described herein with reference to FIGS. 1 - 4 .
  • the L 3 of cache level 80 suitably stores data that is changing, while the L 3 of 82 and 84 suitably is used to store pre-fetched uncompressed data.
  • the pre-fetch data comprises the excessive amount of data read in many environments.
  • a specific command to pre-fetch large blocks of data units, or purge a large block of data in the cache, or freeze a block of data in cache may be employed as is well understood by those skilled in the art.
  • the victim data unit i.e., tracks, records, or lines
  • the IL 3 victim data unit is suitably stored directly on DASD 86 , represented by arrow 92 .
  • the victim data unit at each level of compressed partition L 4 is pushed down to the next lower level compressed partition L 4 , i.e., the victim of IL 4 is pushed down to IIL 4 (arrow 94 ), and the victim of IIL 4 is pushed down to IIIL 4 (arrow 96 ).
  • the victim of IIIL 4 preferably is uncompressed(decompressed) and stored to DASD 86 (arrow 98 ).
  • the missed data is preferably fed directly from either: compressed partition IL 4 (arrow 99 ) after decompression; a lower level prefetch L 3 (arrows 100 or 102 ); a lower level L 4 (arrows 104 or 106 ) after decompression; or DASD 86 (arrow 102 ).
  • FIG. 1 is described as partitioning a single physical cache, the partitions are suitably capable of being achieved with separate physical memory devices.
  • a greater performance penalty results for misses in L 3 and hits in L 4 when on separate memory devices. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.

Abstract

Aspects for caching storage data include partitioning a storage cache to include a compressed data partition and an uncompressed data partition, and adjusting a size of the compressed data partition and the uncompressed data partition for chosen performance characteristics. A data caching system aspect in a data processing system having a host system in communication with a storage system includes at least one storage device and at least one partially compressed cache. The at least one partially compressed cache further includes an uncompressed partition and a compressed partition, where the compressed partition stores at least a victim data unit from the uncompressed partition.

Description

    FIELD OF THE INVENTION
  • The present invention relates generally to storage caches, and more particularly to partially compressed storage caches. [0001]
  • BACKGROUND OF THE INVENTION
  • As processing speeds of computer systems continue to increase, the ability to efficiently retrieve data from memory remains vital. The use of memory caches has been effective in compensating for speed mismatches between two levels of storage access, e.g., between a processor and main memory. Caches generally provide higher speed memory storage for recently/frequently used data in a computer system. [0002]
  • Improving the performance and utilization of caches remains an important aspect of computer system design. Typically, cache organizations and algorithms attempt to utilize the spatial and temporal localities of the storage access. Success in caches is usually measured by the hit ratio (i.e., the number of times that the needed data is found in the cache), as well as average access time (i.e., average time to locate and retrieve a piece of information and return it for processing), maximal throughput (i.e., maximal rate of data transfer), etc. At times, attempts to achieve better performance involve changes to the cache organization, which often improve hit ratios and access times at a slight expense of the maximal throughput due to cache replacing overhead. [0003]
  • For example, one typical method of improving performance by increasing the hit ratio involves expanding the size of the cache. Unfortunately, as the cache size is increased, an equivalent increase in the hit ratio percentage is not always achieved. For example, doubling a 4 GB (gigabytes) cache with a 75% hit ratio to 8 GB does not result in a doubling of the hit ratio. While a small percentage of improvement in the hit ratio occurs, the doubling in size results in considerable cost expenses. [0004]
  • Alternatively, with a fully compressed cache, an increase in storage capacity is achieved without increasing the cache size. When the compressed cache is used in a read-only environment, normally few problems in data integrity result. However, when used in an environment of changing data, significant problems result, mainly due to the need to have random access to the compressed data in the cache. Forming smaller, uniform-sized chunks within the compressed cache is sometimes used to allow more random access to portions of data. However, further complications in updating are created, since the compressed data may not elegantly fit within each chunk due to the size variations in the data. Further, compressing small chunks usually result in lower compression. [0005]
  • Accordingly, a need exists for a cache organization and algorithm that achieves results at least as effective as increasing a cache's size without the concomitant expense incurred by size increases. [0006]
  • SUMMARY OF THE INVENTION
  • The present invention meets these needs through a partially compressed cache organization. A method aspect for caching storage data includes partitioning a storage cache to include a compressed data partition and an uncompressed data partition, and adjusting the compressed data partition and the uncompressed data partition for chosen performance characteristics, including overall cache size. A data caching system aspect in a data processing system having a host system in communication with a storage system includes at least one storage device and at least one partially compressed cache. The at least one partially compressed cache further includes an uncompressed partition and a compressed partition, where the compressed partition stores at least a victim data unit from the uncompressed partition. [0007]
  • With the present invention, alternative caching organizations and algorithms are introduced that allow for dynamically adjusting partition sizes of uncompressed and compressed cache data according to hit-ratio, response time, compression ratios, and throughput (max IO) objectives. Further, the utilization of sub-partitioning a cache in order to achieve a partially compressed cache is readily applicable in multi-level caching of storage subsystems. In addition, the partially compressed cache organization achieves improved performance on par with increasing a cache's size without incurring the cost expense of cache size increase. These and other advantages of the aspects of the present invention will be more readily understood in conjunction with the following detailed description and accompanying drawings. [0008]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a block diagram of a data processing system arrangement that utilizes caching in accordance with the present invention. [0009]
  • FIGS. 2 and 3 present performance modeling results that illustrate response time changes and maximum IO values with varying levels of compression of the storage cache of FIG. 1. [0010]
  • FIG. 4 illustrates an alternate arrangement of the system of FIG. 1 that realizes significant bandwidth advantages through the use of partial cache compression. [0011]
  • FIG. 5([0012] a) illustrates a block diagram of typical multi-level caching.
  • FIG. 5([0013] b) illustrates a multi-level caching arrangement in accordance with a further aspect of the present invention.
  • DESCRIPTION OF THE INVENTION
  • The present invention relates to partially compressed data caching. The following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements. Various modifications to the preferred embodiment will be readily apparent to those skilled in the art and the generic principles herein may be applied to other embodiments. Thus, the present invention is not intended to be limited to the embodiment shown but is to be accorded the widest scope consistent with the principles and features described herein. [0014]
  • FIG. 1 illustrates a block diagram of a data processing system arrangement that utilizes caching in accordance with the present invention. Included as a part of the system are a [0015] host system 10, e.g., a central processing unit (CPU), storage cache 12, and storage (DASD, direct access storage device) 14. As in typical similar arrangements, storage cache 12 is used to provide quicker access to storage data than storage 14, e.g., on the order of microseconds per access, rather than milliseconds. In contrast to typical caching arrangements, however, the storage cache 12 in accordance with the present invention optimizes cache organization and algorithms dynamically to achieve maximum performance without requiring additional storage cache. It should be appreciated that although the discussion focuses primarily on storage controller cache applications, the invention is equally applicable to CPU cache and other caching applications, as is well appreciated by those skilled in the art.
  • Through the present invention, storage cache [0016] 12 is partitioned into at least two partitions, partition L3 and partition L4. Preferably, one partition, e.g., partition L3, stores uncompressed data, while a second partition, e.g., partition L4, stores compressed data. Suitably then, a compressor unit 16 and decompressor unit 18 are also included in the system of FIG. 1. For illustrative purposes, partitions L3 and L4 are shown as distinct partitions in the storage cache 12. However, this is demonstrative of the logical appearance of the partitions. In a preferred embodiment, the partitions L3 and L4 are part of the same physical memory, so that compressed and uncompressed data are intermingled in the storage cache 12.
  • By way of example, with a compression ratio of about 3:1, and storage cache [0017] 12 having a total size of 4 GB, partitioning the storage cache 12 into two partitions of 2 GB each results in a logical size of 8 GB for storage cache 12 (2 GB uncompressed and 2 GB compressed to store 6 GB worth of data). The partition L3 suitably stores the 2 GB of uncompressed data (unit) lines, while the partition L4 stores the 2 GB of compressed data (unit) lines. For purposes of this disclosure, a “line” suitably refers to a line cache replacement unit (or a replaceable data unit) in a cache. Thus, for storage cache 12, a cache line suitably refers to a storage device 14 record, track, or a fraction of a track.
  • With partitions L[0018] 3 and L4 occupying the same storage cache 12, preferably a L3 and L4 cache directory 19 tracks the status of compressed and uncompressed locations/data blocks within the storage cache 12 by storing a listing of the uncompressed blocks and a listing of compressed blocks, with the formation and updating of the directory achieved according to methods well understood by those skilled in the art. Of course, the listings could be provided as two separate directories, and a third directory (not shown) may also be included to store a listing of empty blocks, if desired.
  • In terms of data flow in the storage cache [0019] 12, preferably partition L3 stores data that is changing. When new data is added to partition L3, preferably a least recently used (LRU) algorithm is employed to make room for the new data by removing the least recently used data, which takes advantage of the storage access temporal locality and determines which uncompressed line becomes the removed data/victim line, as is well-known to those skilled in the art. Suitably, the victim data unit, i.e., victim line of a record, track, or fraction of a track, of partition L3 is compressed through compressor unit 16 and written to the same location in the storage cache 12. With the victim line of partition L3 compressed and stored in partition L4, the original uncompressed location now stores the compressed line and opens two empty compressed blocks in the storage cache 12. Directory 19 is suitably updated to indicate the change in status of the block from uncompressed (i.e., part of partition L3) to compressed (i.e., part of partition L4). The uncompressed victim line is also preferably written/destaged to storage 14. If L3 and L4 are provided as separate physical memory devices, and the L4 physical distance and interface is different than L3, then compressed data to L4 is physically moved to the empty slot in L4, and the directory of L4 stored in L3 or L4 is appropriately updated.
  • The exact compressed data size is usually unpredictable. Within a class of data type and application, the compressed size statistics are usually fairly stable. Varying methods exist for estimating space allocation for partition L[0020] 4 (e.g., for a victim line from partition L3 or for prefetched data from storage 14). A compression sniffer 20 is useful for trail compression for a data line which has been seen once. Suitably, a sniffer is a device and/or software routine that monitors data lines as they are transferred through the caching system, i.e., that monitors the ‘trail’ the data lines travel, to identify how the data lines are compressed. Thus, the compression sniffer/trail compression unit 20 suitably gives actual compressed size. If a compression sniffer is not available, alternative methods for estimating space allocation include: assigning enough space to hold the worst case expanded data (e.g., 1.125 times the line size), storing the compressed data, and reclaiming the left over space; allocating space based on a priori compression statistics, e.g., using 2 sigma the compression statistics (or other threshold) for allocation and keeping the overflow if compressed size is exceeded; or assigning the threshold (e.g., about 3:1) and merely dropping the compressed data if it does not fit (with the assumption that the data is destaged to DASD).
  • Partition L[0021] 4 suitably uses a LRU replacement policy and compression weighted replacement (CWR) to determine its victim lines, which takes into account that data that is directly interpretable by human and machine usually compresses well and that data that can be compressed moderately well has a higher chance of being reused than data that does not compress well. The victim line of partition L4 may suitably be discarded/removed via a bit discarder mechanism 24, i.e., a ‘bit bucket’ or universal data sink, to make space in the storage cache 12 when new data is written to partition L4.
  • When a read attempt results in a read miss in partition L[0022] 3 with a read hit in partition L4, partition L4 suitably passes the uncompressed line (decompressed through decompressor unit 18) to partition L3 and to the host system 10. The partition L4 entry for that line suitably becomes invalidated. When a miss occurs in both partitions L3 and L4, suitably the requested data is read back from storage 14 to the partition L3 for transfer to host system 10.
  • Commonly, the source block size is 512 Bytes or 1024 Bytes and a cache line, i.e., 2K or 4K Bytes, track, or half track, consists of multiple blocks, with a read or modify write to a specific block in a cache line done through the partition L[0023] 3. However, a performance penalty may result if a read or modify write occurs to a specific block that is kept in the partition L4. The primary impacts are the extra reads and writes between L3 and L4 which consumes additional data bandwidth. Preferably, therefore, when the input block size is 1.5K, 2.0K Bytes or greater, the cache line contains only a single block. In this manner, more cache lines are able to be kept in partition L4, resulting in less transfers between partitions L3 and L4. For an adaptive compression algorithm, a block size of 2KB allows a good compression region (i.e., about 3.0), since a block size of 512 B usually achieves less than 2.0 compression.
  • In addition to LRU replacement, the compression ratio, i.e., the ratio of uncompressed size/compressed size, is a factor in the cache replacement algorithm. Thus, the compression factor acts as a discrimination factor for cache replacement (the term discrimination factor being known commonly in Pattern Recognition (PR) research), with the objectives of the caching algorithm being a classification process that keeps the class of most useful data in the cache through some discrimination process (as less well-compressed data usually is used less often). [0024]
  • Suitably, the size of partitions L[0025] 3 and L4 are adjusted according to cache size, storage device size, compression/decompression hardware speed, and performance objectives, e.g., access time, response time, maximum rate of input and output required (MAX IO requirements), etc. With the use of dynamic partitioning, the storage cache 12 of the present invention realizes significant advantages over straight caching. FIGS. 2 and 3 present performance modeling results that illustrate response time changes and maximum IO values as adjustments are made to the percentage of the storage cache 12 that is compressed. The performance model of FIG. 2 suitably illustrates a storage cache 12 of size 4 GB, while the performance model of FIG. 3 suitably illustrates a storage cache 12 of size 8 GB, both used with a storage 14 of 360 GB and a compression ratio of about 3:1.
  • With reference to FIG. 2, when the entire storage cache [0026] 12 is used for uncompressed data storage, i.e., the partition L3 comprises 100% and the partition L4 comprises 0% of the storage cache 12, the base response time (BASE RT) is 2.9 ms (milliseconds) with a maximum IO per second (MAX IO) of 9830. With L3 at 75% and L4 at 25%, the response time is 2.8 ms, and the maximum 10 is 10530. With L3 at 50% and L4 at 50%, the response time drops to 2.6 ms, and the maximum IO is 11190. With L3 at 25% and L4 at 75%, the response time is 2.5 ms, and the maximum IO is 10620.
  • With reference to FIG. 3, when L[0027] 3 is 100%, and L4 is 0%, i.e., all 8 GB are uncompressed, response time is 2.7 ms and maximum IO is 10770. As the percentage changes, i.e., partition L3 goes from 75% to 50% to 25% and while partition L4 goes from 25% to 50% to 75%, the response time is 2.2 ms and the maximum IO decreases from 13590 to 13440 to 13020. The results of performance modeling, as illustrated in FIGS. 2 and 3, reveals that the partitioning of a 4 GB storage cache into two partitions of equal percentage for uncompressed and compressed data (i.e., FIG. 2, (c)) provides better response time and maximum 10 than an 8 GB storage cache with 100% uncompressed data (i.e., FIG. 3, (a)). For these conditions, the response time for the 4 GB arrangement is 2.6 ms with a maximum IO of 11190, while the response time for the 8 GB arrangement is 2.7 ms with a maximum of 10770. Thus, the 50% partitioning approach provides even better performance gains for a 4 GB cache than merely doubling the size to 8 GB with no compression would. FIGS. 2 and 3 further reveal that the most beneficial partition percentages are dependent on cache size. Thus, for the 4 GB cache, a 50% L3 and 50% L4 partitioning provides overall optimum performance. Contrastly, for the 8 GB cache, a 75% L3 and 25% L4 partitioning provides overall optimum performance.
  • Improvements in performance that realizes significant bandwidth advantages through the use of the partitions L[0028] 3 and L4 is achieved with an alternate arrangement of the system of FIG. 1, as illustrated in FIG. 4. In this alternate embodiment, a system 50 interfaces with a storage subsystem 52 via a system adapter 54. The system adapter 54 suitably includes compressor logic 56 and decompressor logic 58. Further included is storage cache 60 and storage 62 (e.g., DASD). Preferably, partition L3 of storage cache 60 includes a plurality of ‘home areas’, where each ‘home area’ refers to an uncompressed data block, as described with reference to FIG. 1. However, in contrast to the preferred embodiment, the system adapter 54 sends compressed data to these home areas at a preferred compression ratio, e.g., more than approximately 2:1. The portion of each home area required to store data varies depending upon the amount of data being compressed, of course, but will never exceed the entire home area due to the compression.
  • With the system adapter [0029] 54 sending compressed data to the storage cache 60, an effective increase in bandwidth for data transfer results by a factor dependent upon the compression ratio. For example, with a compression ratio of about 3:1, data transfer can now effectively occur at three times the normal transfer rate, since three times as much data is able to be transferred, e.g., a rate of 100 MB/s (megabyte/second) now appears as 300 MB/s. Of course, a similar increase in bandwidth may be realized for data transfers from partition L3 to storage 62, since the data being transferred from partition L3 has already been compressed.
  • In terms of data flow, again an LRU algorithm suitably determines which data from partition L[0030] 3 becomes a victim line for storage in partition L4 with writing to storage 62, and similarly, which lines are victims from partition L4. The size of the blocks in partition L4 are smaller in comparison to the data block storage area/home areas of partition L3. Thus, a data sniffer 64 suitably determines whether a victim line from L3 will fit in partition L4. Compacting of data may occur in order to realize line transfer from L3 to L4. Data is suitably discarded from partition L4 via a bit bucket/discarder 65. When a hit occurs in partition L4, suitably the line is returned to partition L3 and to system adapter 54. The decompressor logic 58 suitably decompresses the data before being returned to system 50. If a miss occurs in both partitions L3 and L4, the data is returned from storage 62 to partition L3 and directly to system adapter 54 for decompression and return to system 50.
  • The caching arrangement aspects of the present invention are further capably utilized in systems that comprise multiple levels of caching. By way of example, FIG. 5([0031] a) illustrates a block diagram of typical multi-level caching, such as in a RAMAC storage system from IBM Corporation, Armonk, N.Y. An upper level cache 70 is shared among all of the lower level caches 72, each of which is shared among all of the drawer level caches 74, each of which is then shared among respective disk level caches 76. Conventionally, the LRU algorithm is employed at each of the levels. However, the addition of new data at the upper level cache 70 normally results in replacement of the data in each of the next lower cache levels, as well. Unfortunately, this results in duplication of cache lines, i.e., cache pollution, among the caches.
  • In accordance with a further aspect of the present invention, a multi-level caching arrangement utilizes dynamic partitioning, as described with reference to FIG. 5([0032] b). As shown in FIG. 5(b), each level of the multi-level caching arrangement employs dynamic partitioning of uncompressed and compressed data partitions. Thus, an upper, level one system cache 80 is partitioned into an uncompressed partition, IL3, and compressed partition, IL4. Similarly, a drawer level cache 82 comprises an uncompressed partition IIL3, and compressed partition IIL4, while a drive cache 84 associated with a DASD 86 includes a compressed partition IIIL4 and an uncompressed partition IIIL3. A compression sniffer 88 is further included and operates in a manner as previously described with reference to FIG. 1 to solve problems associated with compression ratio variability and prediction problems. Sizes of the partition L3 and partition L4 of each level are suitably adjusted dynamically to achieve desired performance characteristics, as described herein with reference to FIGS. 1-4.
  • The L[0033] 3 of cache level 80 suitably stores data that is changing, while the L3 of 82 and 84 suitably is used to store pre-fetched uncompressed data. The pre-fetch data comprises the excessive amount of data read in many environments. Alternatively, a specific command to pre-fetch large blocks of data units, or purge a large block of data in the cache, or freeze a block of data in cache, may be employed as is well understood by those skilled in the art.
  • In terms of data flow, the victim data unit (i.e., tracks, records, or lines) chosen at top-level IL[0034] 3 is compressed and then pushed to the top-level compressed partition IL4, represented by arrow 90. Alternatively, for compression ratio lower than the predetermined threshold or for a fast write operation, the IL3 victim data unit is suitably stored directly on DASD 86, represented by arrow 92. The victim data unit at each level of compressed partition L4 is pushed down to the next lower level compressed partition L4, i.e., the victim of IL4 is pushed down to IIL4 (arrow 94), and the victim of IIL4 is pushed down to IIIL4 (arrow 96). The victim of IIIL4 preferably is uncompressed(decompressed) and stored to DASD 86 (arrow 98). When a read data miss occurs in the IL3 partition, the missed data is preferably fed directly from either: compressed partition IL4 (arrow 99) after decompression; a lower level prefetch L3 (arrows 100 or 102); a lower level L4 (arrows 104 or 106) after decompression; or DASD 86 (arrow 102).
  • Thus, through the dynamic partitioning and data organization with multiple levels of caching as represented in FIG. 5([0035] b), duplication of data among the cache levels is readily avoided, while the data requesting path is made faster. Further, retaining victim lines in the described manner creates better performance than normal methods which throw victim lines away.
  • Although the present invention has been described in accordance with the embodiments shown, one of ordinary skill in the art will readily recognize that there could be variations to the embodiments and those variations would be within the spirit and scope of the present invention. For example, although the embodiment of FIG. 1 is described as partitioning a single physical cache, the partitions are suitably capable of being achieved with separate physical memory devices. Of course, a greater performance penalty results for misses in L[0036] 3 and hits in L4 when on separate memory devices. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.

Claims (34)

What is claimed is:
1. A method for caching data in a data processing system, the method comprising:
partitioning a storage cache to include a compressed data partition and an uncompressed data partition; and
adjusting the compressed data partition and the uncompressed data partition for chosen performance characteristics, including overall cache size.
2. The method of
claim 1
wherein the performance characteristics further include cache partition ratios, a storage device size and performance objectives.
3. The method of
claim 1
wherein a trail compression unit determines space allocation for data in the compressed data partition to avoid problems associated with compression ratio variability and prediction.
4. The method of
claim 3
further comprising utilizing a predetermined compression ratio to form compressed data and storing the compressed data in the compressed data partition.
5. The method of
claim 4
wherein the compressed data partition stores a victim data unit from the uncompressed data partition, the victim data unit comprising a track, record, or line of data.
6. The method of
claim 5
wherein the victim data unit comprises least recently used data from the uncompressed data partition.
7. The method of
claim 4
wherein the compressed data partition stores a pre-fetch data unit from a storage device.
8. The method of
claim 4
further comprising storing the compressed data in-place in the storage cache.
9. The method of
claim 4
wherein the predetermined compression ratio comprises a ratio of more than approximately 2:1.
10. The method of
claim 5
further comprising storing the victim data unit directly to storage when the trail compression unit determines a compression ratio falls below a given threshold.
11. The method of
claim 7
wherein the storage device comprises a direct access storage device (DASD).
12. The method of
claim 4
further comprising tracking a status of data lines in the storage cache with a directory to identify whether the data lines are compressed or uncompressed.
13. A data caching system in a data processing system, the data processing system including a host system in communication with a storage system, the data caching system comprising:
at least one storage device; and
at least one cache coupled to the at least one storage device for caching data from the at least one storage device, the at least one cache partitioned to include a compressed partition and an uncompressed partition, wherein the uncompressed partition stores at least a victim data unit from the uncompressed partition.
14. The system of
claim 13
wherein the storage system further comprises a compression sniffer coupled to the at least one partially compressed cache to determine space allocation in the compressed partition.
15. The system of
claim 13
wherein the at least one partially compressed cache comprises an upper level cache of uncompressed data and at least one lower level cache of compressed data.
16. The system of
claim 13
wherein the compressed partition further stores prefetch storage data.
17. The system of
claim 13
wherein the victim data unit comprises least recently used data from the uncompressed partition.
18. The system of
claim 13
wherein the storage system further comprises a compressor and a decompressor coupled to the at least one partially compressed cache.
19. The system of
claim 17
wherein the compressor compresses the victim data unit at a ratio of more than approximately 2:1.
20. A method for effectively increasing data transfer bandwidth between a host system and a storage system, the method comprising:
utilizing a system adapter in the storage system for storage data transfer, the system adapter including compression logic and decompression logic;
coupling at least one partially compressed cache to the system adapter, the at least one partially compressed cache comprising an uncompressed partition and a compressed partition; and
transferring data through the compression logic to the uncompressed partition to effectively increase data transfer from the system adapter to the partially compressed cache by a factor corresponding to a compression ratio of the compression logic.
21. The method of
claim 20
wherein the compression ratio comprises more a compression ratio of more than approximately 2:1.
22. The method of
claim 20
further comprising storing victim lines from the uncompressed partition in the compressed partition.
23. The method of
claim 22
further comprising writing the victim lines to a storage device in the storage system.
24. A method for caching data in a multi-level caching arrangement of a data processing system, the method comprising:
partitioning each cache level to include a compressed data partition and an uncompressed data partition;
adjusting the compressed data partition and the uncompressed data partition of each cache level for chosen performance characteristics; and
utilizing the compressed and uncompressed data partitions of each cache level to cache data from a storage device, wherein cache pollution is substantially avoided and a data requesting path becomes faster.
25. The method of
claim 24
further comprising storing pre-fetched uncompressed storage data in the uncompressed data partition of each cache level.
26. The method of
claim 25
further comprising compressing a victim data unit of an uncompressed data partition of a top level cache and storing the victim data unit in a compressed data partition of the top level cache.
27. The method of
claim 26
further comprising utilizing a predetermined compression ratio to form compressed data for the compressed data partition.
28. The method of
claim 26
further comprising utilizing a trail compression unit for estimating space allocation for data in the compressed data partition to avoid problems associated with compression ratio variability and prediction.
29. The method of
claim 26
further comprising storing a victim data unit of the compressed data partition of one cache level in a compressed data partition of a next lower cache level.
30. The method of
claim 29
further comprising decompressing a victim data unit of a compressed data partition of a lowest level cache for storage back in the storage device.
31. A data processing system comprising:
a host system, the host system performing accesses on storage data; and
a storage system, the storage system coupled to the host system and storing the storage data, the storage system comprising multiple cache levels, each cache level comprising a compressed data partition and an uncompressed data partition.
32. The system of
claim 31
wherein a compressed data partition of a top level cache stores a victim data unit of a top level uncompressed data partition.
33. The system of
claim 32
wherein a compressed data partition of each lower level cache stores a victim data unit of a compressed data partition of a next higher level cache.
34. The system of
claim 33
wherein the storage subsystem further comprises a storage device coupled to the multiple cache levels, the storage device storing a victim data unit from a compressed data partition of a lowest cache level.
US09/090,028 1998-06-10 1998-06-10 Data caching with a partially compressed cache Expired - Fee Related US6324621B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US09/090,028 US6324621B2 (en) 1998-06-10 1998-06-10 Data caching with a partially compressed cache

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US09/090,028 US6324621B2 (en) 1998-06-10 1998-06-10 Data caching with a partially compressed cache

Publications (2)

Publication Number Publication Date
US20010001872A1 true US20010001872A1 (en) 2001-05-24
US6324621B2 US6324621B2 (en) 2001-11-27

Family

ID=22220877

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/090,028 Expired - Fee Related US6324621B2 (en) 1998-06-10 1998-06-10 Data caching with a partially compressed cache

Country Status (1)

Country Link
US (1) US6324621B2 (en)

Cited By (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2834574A1 (en) * 2002-01-10 2003-07-11 Hewlett Packard Co DEVICE AND METHODS FOR COMPRESSING A COVER LINE
US20030145170A1 (en) * 2002-01-31 2003-07-31 Kever Wayne D. Dynamically adjusted cache power supply to optimize for cache access or power consumption
US20040225854A1 (en) * 2003-05-08 2004-11-11 International Business Machines Corporation Apparatus and method to write information to one or more information storage media
US20050071566A1 (en) * 2003-09-30 2005-03-31 Ali-Reza Adl-Tabatabai Mechanism to increase data compression in a cache
US20050071599A1 (en) * 2003-09-30 2005-03-31 Modha Dharmendra Shantilal Storage system and method for dynamically allocating cache space among different workload classes
US20050071151A1 (en) * 2003-09-30 2005-03-31 Ali-Reza Adl-Tabatabai Compression-decompression mechanism
US20050144388A1 (en) * 2003-12-31 2005-06-30 Newburn Chris J. Processor and memory controller capable of use in computing system that employs compressed cache lines' worth of information
US20050144386A1 (en) * 2003-12-29 2005-06-30 Ali-Reza Adl-Tabatabai Mechanism to store reordered data with compression
US20050160234A1 (en) * 2004-01-15 2005-07-21 Newburn Chris J. Multi-processor computing system that employs compressed cache lines' worth of information and processor capable of use in said system
US7143240B2 (en) 2003-10-31 2006-11-28 International Business Machines Corporation System and method for providing a cost-adaptive cache
US20070005911A1 (en) * 2005-07-01 2007-01-04 Nec Laboratories America, Inc. Operating System-Based Memory Compression for Embedded Systems
US20070005901A1 (en) * 2004-06-14 2007-01-04 Quickshift, Inc. Adaptive input / output compressed system and data cache and system using same
US20080244214A1 (en) * 2007-03-28 2008-10-02 Flemming Diane G Workload management in virtualized data processing environment
US20080244568A1 (en) * 2007-03-28 2008-10-02 Flemming Diane G Method to capture hardware statistics for partitions to enable dispatching and scheduling efficiency
US20080244215A1 (en) * 2007-03-28 2008-10-02 Flemming Diane G Workload management in virtualized data processing environment
US20080244213A1 (en) * 2007-03-28 2008-10-02 Flemming Diane G Workload management in virtualized data processing environment
US20090100225A1 (en) * 2007-10-16 2009-04-16 Sony Corporation Data processing apparatus and shared memory accessing method
US20090193040A1 (en) * 2008-01-29 2009-07-30 Mitel Networks Corporation System and method for storing a program using partial compression
US7587572B1 (en) * 2004-08-31 2009-09-08 Sun Microsystems, Inc. Method and system for managing process memory configured in resizable uncompressed and compressed regions
US20090228664A1 (en) * 2008-03-04 2009-09-10 Borkenhagen John M Memory Compression Implementation in a System With Directly Attached Processor Memory
US20090282064A1 (en) * 2008-05-07 2009-11-12 Veeramanikandan Raju On the fly compression and storage device, system and method
US20100122058A1 (en) * 2008-11-12 2010-05-13 Oracle International Corporation Memory page eviction based on present system operation
US20120084531A1 (en) * 2010-09-30 2012-04-05 International Business Machines Corporation Adjusting memory allocation of a partition using compressed memory paging statistics
US8738962B2 (en) 2010-11-17 2014-05-27 International Business Machines Corporation Memory mirroring with memory compression
US20140258609A1 (en) * 2011-11-21 2014-09-11 Huawei Technologies Co., Ltd. Quality of service control method for storage system
US8934867B2 (en) 2012-07-10 2015-01-13 Symbol Technologies, Inc. On-demand access tunnel between service provider network and wireless communication network
US20150089126A1 (en) * 2013-09-25 2015-03-26 Intel Corporation Data Compression In Processor Caches
US20150278098A1 (en) * 2012-12-21 2015-10-01 Tencent Technology (Shenzhen) Company Limited Method For Managing Cache Data Of A Mobile Terminal, Mobile Terminal, And Server
US20150370715A1 (en) * 2014-06-19 2015-12-24 Lsi Technologies Inc. Safe and Efficient Dirty Data Flush for Dynamic Logical Capacity Based Cache in Storage Systems
US20160004636A1 (en) * 2014-07-07 2016-01-07 Samsung Electronics Co., Ltd. Electronic device with cache memory and method of operating the same
US20170161193A1 (en) * 2015-12-02 2017-06-08 International Business Machines Corporation Hybrid cache
CN107430554A (en) * 2015-03-27 2017-12-01 英特尔公司 Storage cache performance is improved as the standard that cache is inserted by using the compressibility of data
US20180276143A1 (en) * 2016-07-19 2018-09-27 Nutanix, Inc. Dynamic cache balancing
EP3391227A4 (en) * 2015-12-17 2019-08-28 Advanced Micro Devices, Inc. Hybrid cache
US11086793B2 (en) * 2020-01-15 2021-08-10 EMC IP Holding Company LLC Data reduction techniques for use with caching
US11119981B2 (en) * 2017-10-27 2021-09-14 Hewlett Packard Enterprise Development Lp Selectively redirect-on-write data chunks in write-in-place file systems
US11455256B2 (en) * 2019-09-13 2022-09-27 Kioxia Corporation Memory system with first cache for storing uncompressed look-up table segments and second cache for storing compressed look-up table segments

Families Citing this family (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7386671B2 (en) * 2000-06-09 2008-06-10 Texas Instruments Incorporated Smart cache
US6694393B1 (en) * 2000-06-30 2004-02-17 Lucent Technologies Inc. Method and apparatus for compressing information for use in embedded systems
US6721814B1 (en) * 2001-05-15 2004-04-13 Emc Corporation System and method of improving host performance characteristics by intelligent shift of data caching from host to cache-enabled mass storage subsystem
US6654867B2 (en) * 2001-05-22 2003-11-25 Hewlett-Packard Development Company, L.P. Method and system to pre-fetch compressed memory blocks using pointers
US6678814B2 (en) * 2001-06-29 2004-01-13 International Business Machines Corporation Method and apparatus for allocating data usages within an embedded dynamic random access memory device
US7656877B1 (en) 2001-08-21 2010-02-02 Cisco Technology, Inc. Apparatus and methods for sniffing data in a cable head end
US7092397B1 (en) 2001-08-31 2006-08-15 Cisco Technology, Inc. Method and apparatus for mapping an MPLS tag to a data packet in a headend
KR20030054110A (en) * 2001-12-24 2003-07-02 한국전자통신연구원 Method and apparatus of caching multiple Java-database connectivity
US6857047B2 (en) * 2002-06-10 2005-02-15 Hewlett-Packard Development Company, L.P. Memory compression for computer systems
US9342459B2 (en) * 2002-08-06 2016-05-17 Qualcomm Incorporated Cache management in a mobile device
US6981119B1 (en) * 2002-08-29 2005-12-27 Advanced Micro Devices, Inc. System and method for storing performance-enhancing data in memory space freed by data compression
US7912299B2 (en) * 2004-10-08 2011-03-22 Microsoft Corporation System and method for efficiently encoding data
US7075535B2 (en) * 2003-03-05 2006-07-11 Sand Codex System and method for exact rendering in a zooming user interface
US7546419B2 (en) * 2004-06-01 2009-06-09 Aguera Y Arcas Blaise Efficient data cache
US7042455B2 (en) * 2003-05-30 2006-05-09 Sand Codex Llc System and method for multiple node display
US7930434B2 (en) * 2003-03-05 2011-04-19 Microsoft Corporation System and method for managing communication and/or storage of image data
US7133054B2 (en) * 2004-03-17 2006-11-07 Seadragon Software, Inc. Methods and apparatus for navigating an image
US7254271B2 (en) * 2003-03-05 2007-08-07 Seadragon Software, Inc. Method for encoding and serving geospatial or other vector data as images
US7190367B2 (en) * 2003-03-25 2007-03-13 Mitsubishi Electric Research Laboratories, Inc. Method, apparatus, and system for rendering using a progressive cache
US7107400B2 (en) 2003-03-31 2006-09-12 International Business Machines Corporation System, apparatus, and process for evaluating projected cache sizes
US7181457B2 (en) * 2003-05-28 2007-02-20 Pervasive Software, Inc. System and method for utilizing compression in database caches to facilitate access to database information
US20050204104A1 (en) * 2004-03-15 2005-09-15 Tatsundo Aoshima Server and method for managing volume storing digital archive
US7225297B2 (en) * 2004-05-28 2007-05-29 International Business Machines Corporation Compressed cache lines incorporating embedded prefetch history data
US7827355B1 (en) * 2004-07-08 2010-11-02 Globalfoundries Inc. Data processor having a cache with efficient storage of predecode information, cache, and method
US7412564B2 (en) * 2004-11-05 2008-08-12 Wisconsin Alumni Research Foundation Adaptive cache compression system
US7457940B2 (en) * 2004-11-16 2008-11-25 International Business Machines Corporation System and method for managing data
US20060235941A1 (en) * 2005-03-29 2006-10-19 Microsoft Corporation System and method for transferring web page data
US7962700B2 (en) * 2006-09-06 2011-06-14 International Business Machines Corporation Systems and methods for reducing latency for accessing compressed memory using stratified compressed memory architectures and organization
US7840786B2 (en) * 2007-04-16 2010-11-23 Advanced Micro Devices, Inc. Techniques for storing instructions and related information in a memory hierarchy
US20090210622A1 (en) * 2008-02-19 2009-08-20 Stefan Birrer Compressed cache in a controller partition
US8447948B1 (en) * 2008-04-25 2013-05-21 Amazon Technologies, Inc Dynamic selective cache compression
US8108442B2 (en) * 2008-07-22 2012-01-31 Computer Associates Think, Inc. System for compression and storage of data
US8125809B2 (en) 2009-09-24 2012-02-28 International Business Machines Corporation Adjustable write bins for multi-level analog memories
US8533550B2 (en) 2010-06-29 2013-09-10 Intel Corporation Method and system to improve the performance and/or reliability of a solid-state drive
US9053018B2 (en) 2012-06-29 2015-06-09 International Business Machines Corporation Compressed memory page selection based on a population count of a dataset
US20150193342A1 (en) * 2012-09-25 2015-07-09 Hitachi, Ltd. Storage apparatus and method of controlling the same
US11132300B2 (en) * 2013-07-11 2021-09-28 Advanced Micro Devices, Inc. Memory hierarchy using page-based compression
TWI493446B (en) 2013-09-23 2015-07-21 Mstar Semiconductor Inc Method and apparatus for managing memory

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5490260A (en) * 1990-12-14 1996-02-06 Ceram, Inc. Solid-state RAM data storage for virtual memory computer using fixed-sized swap pages with selective compressed/uncompressed data store according to each data size
US5237460A (en) * 1990-12-14 1993-08-17 Ceram, Inc. Storage of compressed data on random access storage devices
US5450562A (en) 1992-10-19 1995-09-12 Hewlett-Packard Company Cache-based data compression/decompression
US5574952A (en) * 1994-05-11 1996-11-12 International Business Machines Corporation Data storage system and method for operating a disk controller including allocating disk space for compressed data
US5537588A (en) * 1994-05-11 1996-07-16 International Business Machines Corporation Partitioned log-structured file system and methods for operating the same
US5812817A (en) * 1994-10-17 1998-09-22 International Business Machines Corporation Compression architecture for system memory application

Cited By (59)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2834574A1 (en) * 2002-01-10 2003-07-11 Hewlett Packard Co DEVICE AND METHODS FOR COMPRESSING A COVER LINE
US20030145170A1 (en) * 2002-01-31 2003-07-31 Kever Wayne D. Dynamically adjusted cache power supply to optimize for cache access or power consumption
US7765377B2 (en) 2003-05-08 2010-07-27 International Business Machines Corporation Apparatus and method to write information to one or more information storage media
US20040225854A1 (en) * 2003-05-08 2004-11-11 International Business Machines Corporation Apparatus and method to write information to one or more information storage media
US7107403B2 (en) 2003-09-30 2006-09-12 International Business Machines Corporation System and method for dynamically allocating cache space among different workload classes that can have different quality of service (QoS) requirements where the system and method may maintain a history of recently evicted pages for each class and may determine a future cache size for the class based on the history and the QoS requirements
US20050071566A1 (en) * 2003-09-30 2005-03-31 Ali-Reza Adl-Tabatabai Mechanism to increase data compression in a cache
US20050071599A1 (en) * 2003-09-30 2005-03-31 Modha Dharmendra Shantilal Storage system and method for dynamically allocating cache space among different workload classes
US20050071151A1 (en) * 2003-09-30 2005-03-31 Ali-Reza Adl-Tabatabai Compression-decompression mechanism
US7143240B2 (en) 2003-10-31 2006-11-28 International Business Machines Corporation System and method for providing a cost-adaptive cache
US7162583B2 (en) 2003-12-29 2007-01-09 Intel Corporation Mechanism to store reordered data with compression
WO2005066797A1 (en) * 2003-12-29 2005-07-21 Intel Corporation A mechanism to store reordered data with compression
US20050144386A1 (en) * 2003-12-29 2005-06-30 Ali-Reza Adl-Tabatabai Mechanism to store reordered data with compression
US7512750B2 (en) * 2003-12-31 2009-03-31 Intel Corporation Processor and memory controller capable of use in computing system that employs compressed cache lines' worth of information
US20050144388A1 (en) * 2003-12-31 2005-06-30 Newburn Chris J. Processor and memory controller capable of use in computing system that employs compressed cache lines' worth of information
US20050160234A1 (en) * 2004-01-15 2005-07-21 Newburn Chris J. Multi-processor computing system that employs compressed cache lines' worth of information and processor capable of use in said system
US7257693B2 (en) 2004-01-15 2007-08-14 Intel Corporation Multi-processor computing system that employs compressed cache lines' worth of information and processor capable of use in said system
US20070005901A1 (en) * 2004-06-14 2007-01-04 Quickshift, Inc. Adaptive input / output compressed system and data cache and system using same
US7430638B2 (en) 2004-06-14 2008-09-30 Mossman Holdings Llc Adaptive input / output compressed system and data cache and system using same
US7587572B1 (en) * 2004-08-31 2009-09-08 Sun Microsystems, Inc. Method and system for managing process memory configured in resizable uncompressed and compressed regions
US20070005911A1 (en) * 2005-07-01 2007-01-04 Nec Laboratories America, Inc. Operating System-Based Memory Compression for Embedded Systems
US7698531B2 (en) * 2007-03-28 2010-04-13 International Business Machines Corporation Workload management in virtualized data processing environment
US20080244215A1 (en) * 2007-03-28 2008-10-02 Flemming Diane G Workload management in virtualized data processing environment
US7617375B2 (en) * 2007-03-28 2009-11-10 International Business Machines Corporation Workload management in virtualized data processing environment
US7698530B2 (en) * 2007-03-28 2010-04-13 International Business Machines Corporation Workload management in virtualized data processing environment
US20080244568A1 (en) * 2007-03-28 2008-10-02 Flemming Diane G Method to capture hardware statistics for partitions to enable dispatching and scheduling efficiency
US20080244213A1 (en) * 2007-03-28 2008-10-02 Flemming Diane G Workload management in virtualized data processing environment
US20080244214A1 (en) * 2007-03-28 2008-10-02 Flemming Diane G Workload management in virtualized data processing environment
US8010746B2 (en) * 2007-10-16 2011-08-30 Sony Corporation Data processing apparatus and shared memory accessing method
US20090100225A1 (en) * 2007-10-16 2009-04-16 Sony Corporation Data processing apparatus and shared memory accessing method
US20090193040A1 (en) * 2008-01-29 2009-07-30 Mitel Networks Corporation System and method for storing a program using partial compression
US8214425B2 (en) * 2008-01-29 2012-07-03 Mitel Networks Corporation System and method for storing a program using partial compression
US20090228664A1 (en) * 2008-03-04 2009-09-10 Borkenhagen John M Memory Compression Implementation in a System With Directly Attached Processor Memory
US7984240B2 (en) * 2008-03-04 2011-07-19 International Business Machines Corporation Memory compression implementation in a system with directly attached processor memory
US20090282064A1 (en) * 2008-05-07 2009-11-12 Veeramanikandan Raju On the fly compression and storage device, system and method
US20100122058A1 (en) * 2008-11-12 2010-05-13 Oracle International Corporation Memory page eviction based on present system operation
US8375178B2 (en) * 2008-11-12 2013-02-12 Oracle International Corporation Memory page eviction based on present system operation
US8904145B2 (en) * 2010-09-30 2014-12-02 International Business Machines Corporation Adjusting memory allocation of a partition using compressed memory paging statistics
US20120084531A1 (en) * 2010-09-30 2012-04-05 International Business Machines Corporation Adjusting memory allocation of a partition using compressed memory paging statistics
US8738962B2 (en) 2010-11-17 2014-05-27 International Business Machines Corporation Memory mirroring with memory compression
US20140258609A1 (en) * 2011-11-21 2014-09-11 Huawei Technologies Co., Ltd. Quality of service control method for storage system
US8909868B2 (en) * 2011-11-21 2014-12-09 Huawei Technologies Co., Ltd. Quality of service control method for storage system
US8934867B2 (en) 2012-07-10 2015-01-13 Symbol Technologies, Inc. On-demand access tunnel between service provider network and wireless communication network
US9619388B2 (en) * 2012-12-21 2017-04-11 Tencent Technology (Shenzhen) Company Limited Method for managing cache data of a mobile terminal, mobile terminal, and server
US20150278098A1 (en) * 2012-12-21 2015-10-01 Tencent Technology (Shenzhen) Company Limited Method For Managing Cache Data Of A Mobile Terminal, Mobile Terminal, And Server
US9251096B2 (en) * 2013-09-25 2016-02-02 Intel Corporation Data compression in processor caches
US20150089126A1 (en) * 2013-09-25 2015-03-26 Intel Corporation Data Compression In Processor Caches
US9524201B2 (en) * 2014-06-19 2016-12-20 Avago Technologies General Ip (Singapore) Pte. Ltd. Safe and efficient dirty data flush for dynamic logical capacity based cache in storage systems
US20150370715A1 (en) * 2014-06-19 2015-12-24 Lsi Technologies Inc. Safe and Efficient Dirty Data Flush for Dynamic Logical Capacity Based Cache in Storage Systems
US9684604B2 (en) * 2014-07-07 2017-06-20 Samsung Electronics Co., Ltd. Electronic device with cache memory and method of operating the same
KR20160007874A (en) * 2014-07-07 2016-01-21 삼성전자주식회사 Electronic device having cache memory and method for operating thereof
US20160004636A1 (en) * 2014-07-07 2016-01-07 Samsung Electronics Co., Ltd. Electronic device with cache memory and method of operating the same
KR102336528B1 (en) * 2014-07-07 2021-12-07 삼성전자 주식회사 Electronic device having cache memory and method for operating thereof
CN107430554A (en) * 2015-03-27 2017-12-01 英特尔公司 Storage cache performance is improved as the standard that cache is inserted by using the compressibility of data
US20170161193A1 (en) * 2015-12-02 2017-06-08 International Business Machines Corporation Hybrid cache
EP3391227A4 (en) * 2015-12-17 2019-08-28 Advanced Micro Devices, Inc. Hybrid cache
US20180276143A1 (en) * 2016-07-19 2018-09-27 Nutanix, Inc. Dynamic cache balancing
US11119981B2 (en) * 2017-10-27 2021-09-14 Hewlett Packard Enterprise Development Lp Selectively redirect-on-write data chunks in write-in-place file systems
US11455256B2 (en) * 2019-09-13 2022-09-27 Kioxia Corporation Memory system with first cache for storing uncompressed look-up table segments and second cache for storing compressed look-up table segments
US11086793B2 (en) * 2020-01-15 2021-08-10 EMC IP Holding Company LLC Data reduction techniques for use with caching

Also Published As

Publication number Publication date
US6324621B2 (en) 2001-11-27

Similar Documents

Publication Publication Date Title
US6324621B2 (en) Data caching with a partially compressed cache
US6360300B1 (en) System and method for storing compressed and uncompressed data on a hard disk drive
US8255630B1 (en) Optimization of cascaded virtual cache memory
US8176251B2 (en) Dynamic optimization of cache memory
EP3210121B1 (en) Cache optimization technique for large working data sets
US7831796B2 (en) Apparatus, system, and method for dynamically allocating main memory among a plurality of applications
US9619180B2 (en) System method for I/O acceleration in hybrid storage wherein copies of data segments are deleted if identified segments does not meet quality level threshold
US20030105926A1 (en) Variable size prefetch cache
US6901477B2 (en) Provision of a victim cache within a storage cache hierarchy
WO2012116369A2 (en) Apparatus, system, and method for managing contents of a cache
US20150032942A1 (en) System for increasing utilization of storage media
US7062608B2 (en) Storage device adapter equipped with integrated cache
WO2012106362A2 (en) Apparatus, system, and method for managing eviction of data
US11868271B2 (en) Accessing compressed computer memory
US11204870B2 (en) Techniques for determining and using caching scores for cached data
US11922016B2 (en) Managing free space in a compressed memory system
Wang et al. ADAPT: Efficient workload-sensitive flash management based on adaptation, prediction and aggregation
Benveniste et al. Cache-memory interfaces in compressed memory systems
Cheng et al. AMC: an adaptive multi‐level cache algorithm in hybrid storage systems
Yoon et al. Access characteristic-based cache replacement policy in an SSD
KR20090076265A (en) An prefetched data management method data in a computer memory
Xiao et al. Dual queues cache replacement algorithm based on sequentiality detection
KR102529333B1 (en) Cache management apparatus and method
Li et al. SPOPB: Reducing solid state drive write traffic for flash‐based key‐value caching
Benveniste et al. RC 21662 (Log 97617) 4 February 2000 Computer Science/Mathematics

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, CALIF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SINGH, SHANKER;CHENG, JOE-MING;BEARDSLEY, BRENT CAMERON;AND OTHERS;REEL/FRAME:009249/0336;SIGNING DATES FROM 19980515 TO 19980608

AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SINGH, SHANKER;CHENG, JOE-MING;BEARDSLEY, BRENT CAMERON;AND OTHERS;REEL/FRAME:009682/0495;SIGNING DATES FROM 19980515 TO 19980608

FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Expired due to failure to pay maintenance fee

Effective date: 20131127