|Publication number||US5729713 A|
|Application number||US 08/411,257|
|Publication date||Mar 17, 1998|
|Filing date||Mar 27, 1995|
|Priority date||Mar 27, 1995|
|Publication number||08411257, 411257, US 5729713 A, US 5729713A, US-A-5729713, US5729713 A, US5729713A|
|Original Assignee||Texas Instruments Incorporated|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (17), Non-Patent Citations (2), Referenced by (19), Classifications (7), Legal Events (4)|
|External Links: USPTO, USPTO Assignment, Espacenet|
The invention relates generally to microprocessor based data processing systems and, more particularly, to systems which have various levels of data storage including a first level cache and which use the microprocessor to transfer and modify long data blocks.
Emerging trends in microprocessor based systems like video and audio processing for multimedia applications and peripheral adaptation on the microprocessor bus require a very efficient processor-to-memory interface. FIG. 1 depicts a common architecture of a computer system. The microprocessor (or CPU) MP including a first level cache resides on the mainboard together with the system logic, main memory and adapter cards for peripherals. Optionally the mainboard offers a second level cache (not shown) to speed up system performance. Most of today's Personal Computers (PC) and Engineering Workstations are based on a two-level cache memory design.
With the faster growing speed evaluation on process technology compared to processor bus speed, the internal memory bus of the processor is typically 2-3 times as fast as the external memory bus, so the performance of the on-chip first level cache (1LC) becomes more important. The overall goal of a first level cache is to supply the processor core with information from on-chip memory with no wait states. The information transferred to the core can include instructions and data, and data can be written from the core to the system. Enhanced cache architectures are using a write-back mechanism to update data in the first level cache without updating the external memory system. This allows the core to write data into the first level cache without wait states. The slower external memory is updated from the core only if the data written by the core into the first level cache gets replaced by new data and would otherwise be lost. The processor works most efficiently when more than 90% of read and write requests match the content of the 1LC. A high hit rate of more than 90% can be achieved on processor benchmark programs using data calculation and transportation in blocks smaller than the size of the 1LC. With application software the hit rate of the 1LC is typically less than 90% due to larger programs and data blocks which exceed the size of the 1LC. It is of course important for the system performance to keep the hit rate of the 1LC as high as possible.
To keep pace with larger data blocks transferred from and to the peripherals, several techniques are used to increase the data throughput. For example, peripheral devices such as graphic adapters and hard disk controllers are connected to the microprocessor (MP) local bus to use the full bandwidth of the processor bus. In addition large memory buffers or cache memories are implemented on the peripherals to increase the data throughput of the system. For example, a hard disk might implement a hard disk (HD) cache. There are also programs that use a part of the main memory as a software (SW) cache to store frequently used data. The size of such caches is in general more than 10 times the size of the first level cache (1LC).
FIG. 2 shows a typical example of memory levels involved in hard disk operations with related memory size and maximum transfer rate. The second level cache (2LC) is based on static memory (SRAM) devices which have a faster access time than the dynamic memory (DRAM) devices of the main memory. Thus the SRAM can respond to CPU requests with no wait states whereas the DRAM of the main memory typically requires 1 or more wait states. Another memory level is inserted by hard disk cache programs that use part of the main memory as a software cache to store frequently used memory blocks from the hard disk drive. Larger disk drives use an on-drive cache memory (hard disk cache) which stores data blocks in advance. The slowest transfer rate in the FIG. 2 memory model comes from the magnetic disk drive which is limited by mechanical rotation. In general there is a compromise between memory size and transfer rate. The various memory levels mentioned above are used to improve the overall system performance. However, the function of data transfer from the hard disk to the MP could be also achieved without any cache memory, but at the cost of lower system performance.
FIG. 3 shows the basic data flow required to execute a typical application in the system of FIG. 1. This example indicates that for all tasks of the application the CPU is the device controlling the data flow. Every time the CPU reads information from the external system, the information is allocated in the 1LC. Transfers of data blocks which are multiple times the size of the 1LC (which is the case for all transfers shown in FIG. 3) have the following effect: 1LC gets completely filled with data which will be replaced by new data blocks before there is a CPU request on the same location; the effective 1LC hit rate is near 0% during transfers of large data blocks; and 1LC with write-back mode has to update dirty data from 1LC into slower external memory before the whole 1LC is filled with data.
It is therefore desirable to improve the 1LC hit rate during large block transfers to the CPU, and to reduce the need to update dirty data to external memory during large block transfers to CPUs having a write-back feature.
A data processing arrangement according to the invention prevents a long data read sequence from being allocated to a memory level if the block size of the data read sequence exceeds the size of the memory level.
FIG. 1 is a block diagram of a conventional data processing system.
FIG. 2 shows size and data transfer characteristics of various memory levels within a data processing system such as shown in FIG. 1.
FIG. 3 illustrates the data flow within the data processing system of FIG. 1 during execution of various application tasks.
FIG. 4 is a block diagram which illustrates how non-cache long data read (NCLDR) logic according to the present invention can be applied to the core logic and first level cache of a microprocessor.
FIG. 5 diagrammatically illustrates one exemplary embodiment of the NCLDR logic of FIG. 4.
FIG. 6 is a block diagram which illustrates how another exemplary embodiment of the NCLDR logic is used in a data processing system.
FIG. 7 diagrammatically illustrates the exemplary NCLDR logic of FIG. 6.
As an example for describing the invention, an on-chip 1LC is considered with a cache size of 16 KBytes and a write-back mode. The CPU controls the memory allocation on 1LC with a cache enable signal. Once a long data read sequence is recognized, the control logic of the invention disables memory allocation on 1LC by driving the cache enable signal inactive. This method avoids "flushing" the 1LC with ineffective data. In addition, the most frequently used instruction and data patterns remain in 1LC, the fastest memory of the computer system. Data which is flagged as dirty in the 1LC does not need to be updated into slower external memory.
Microprocessors conventionally have the capability to execute data block instructions. For example, microprocessors based on the so-called x86 architecture have an instruction prefix called REP which allows the repetition of a data transfer command. In a 16 bit device driver software the data transfer command can be repeated 65536 times, which results in a data block size of up to 512 KBytes for a single instruction. The present invention is applicable with microprocessors that have this capability of data block transfers with single instructions. Evaluation of conventional peripheral device driver and software cache routines confirms that data blocks of 8 KBytes to 64 KBytes are typically transferred in combination with the REP instruction prefix during execution of tasks such as those shown in FIG. 3.
In contrast, block transfers of a few hundred bytes up to about 8 K Bytes (half the size of a 16 K Byte 1LC) are used very seldom and have negligible effect on the system performance whether they are allocated in 1LC or not. When a program jumps to and returns from subroutines, the CPU internal registers are typically saved or restored with a single instruction, e.g. push or pop all registers (POPA instruction on x86 architectures). The size of the data blocks involved in such program flow control are dependent on the operating system. For a 16 bit operating system, the block size of this kind of data transfer is typically less than 64 bytes.
Based on the evaluation above, the size of a data block relative to the 1LC size can be predicted by counting the consecutive data read bus cycles. Once the count exceeds the block size typically used for program flow control, the probability of a data block read sequence at least as large as the 16 KByte 1LC of this example is very high. The probability of at least an 8 KByte sequence (half the size of the 16 K Byte 1LC of this example) is even higher.
FIG. 4 is a block diagram which illustrates an example of the present invention. FIG. 4 shows a core 41 and a first level cache 43 of a microprocessor, and a bus 51 for transferring information such as control signals, data, addresses and instructions between the microprocessor and system components external to the microprocessor, such as the system logic, main memory and other peripherals illustrated in FIG. 1. Also connected to bus 51 is non-cache long data read (NCLDR) logic 45. The logic 45 is responsive to information received from bus 51 to produce a processor cache enable signal PKEN which indicates to the microprocessor whether or not the first level cache 43 is to be utilized to cache a block transfer. FIG. 4 also illustrates that the logic 45 may be implemented on-chip with the core 41 and first level cache 43 of the microprocessor, or may be alternatively implemented externally of the microprocessor. More specifically, the dashed lines 47 and 49 illustrate alternative boundaries between the microprocessor and the remainder of the data processing system. If the logic 45 is implemented on-chip with the microprocessor, then the microprocessor/system boundary is defined by dashed line 47 in FIG. 4. If the logic 45 is implemented externally of the microprocessor, then the microprocessor/system boundary is defined by the broken line at 49 in FIG. 4. As will be apparent from the following description, the logic 45 functions the same whether it is implemented on-chip with the microprocessor or off-chip with the remainder of the system components.
FIG. 5 shows one example of the NCLDR logic. The `BUS COMMAND DECODE` block 53 has several CPU control signals from bus 51 as input. The address strobe signal (ADS) indicates that the CPU is starting an external bus cycle. This bus cycle is defined by the three control signals memory/io (MIO), data/code (DC) and write/read (WR). The latter three control signals are valid with ADS. The clock input CLK is the reference signal to sample the signals and check their status. The logic in this block 53 decodes two different types of bus cycles. The data read (data-- rd) output signal indicates that the CPU requests data information from the external system. The code read (code-- rd) output signal indicates that the CPU fetches code information, i.e. instructions. All other bus cycles such as input/output cycles, write cycles or special cycles are not decoded by this block 53. The output signals from logic block 53 feed the input of a `6 BIT COUNTER/DECODER` block 55. FIG. 5 shows that the data-- rd signal is connected with the count input of block 55 and the code-- rd signal is connected with the reset input of block 55.
The 6-bit counter in block 55 has three input signals with the following functions. The count input increases the counter state by one only when the hold input is not asserted. The reset input sets the counter state to zero independent of the other inputs. The counter remains in its current state if hold is asserted. The hold input of logic block 55 is connected with counter state 32 (c32) output from the block 55. The active high c32 output thus represents the maximum attainable count (32) in this embodiment. This means that the counter counts up to state 32 on data read cycles and remains at this state until a code read cycle is initiated by the CPU. So there must be 32 consecutive data read cycles without an instruction fetch (which would activate code-- rd and thus reset the counter) to activate the c32 signal. The c32 signal is inverted at 57 to produce the PKEN signal, and the microprocessor will not allocate any read cycles to 1LC while PKEN is low. Upon the next code read cycle, code-- rd will reset the counter to drive c32 low and PKEN high, thus permitting read cycle allocation to 1LC until the next instance of 32 consecutive data read cycles without an instruction fetch. In a 16 bit operating system, a data block of 64 bytes is required to activate the c32 signal (2 bytes/read cycle X 32 read cycles).
Referring again to FIG. 1, the system logic in many conventional computer systems generates a system cache enable signal SKEN which is input to the microprocessor and which normally enables the first level cache of the microprocessor, but also disables the first level cache when the microprocessor is accessing non-cacheable areas such as video memory. Thus, if the NCLDR logic is implemented externally of the microprocessor, and if the SKEN signal is provided as shown in FIG. 1, then the output of inverter 57 can be used to gate the SKEN signal and thereby provide a suitable PKEN1 signal to the microprocessor, as shown at 45A in FIGS. 6 and 7. The cache enable signal to the processor (PKEN1) is the output of an AND gate 59 which has the system cache enable (SKEN) signal and the inverted version of the c32 signal as input. When the counter state reaches state 32 indicated by an active high level on c32 , then PKEN1 is driven low by the inverted c32 signal output from inverter 57, and the 1LC does not allocate any data from read cycles until the next code fetch.
The exemplary logic 45 and 45A of FIGS. 5-7 can be implemented with, for example, a single programmable logic array (PLA) device, which PLA can be located within the system logic chip set, or separately therefrom as in FIG. 6. Although it is more cost effective to integrate the logic 45 into the microprocessor where the 1LC is located, the system performance is not affected regardless of whether logic 45 is integrated into the microprocessor, or logic 45 or 45A is provided externally of the microprocessor either within the system logic chip set or elsewhere.
In view of the foregoing, it should be clear that the present invention: (1) increases computer system performance with easy implementation by suppressing CPU first level cache on long data read sequences; (2) gets more efficient on systems using a) CPU local bus peripherals (video, audio, mass storage, network), b) Multimedia applications, c) CPUs with internal multiple clock operation (2-3 times as fast as external memory bus), d) CPUs with on chip write back cache, e) 2nd level cache (>64 kByte), f) large main memory (>8 Mbyte), and g) software cache for peripherals; and (3) reduces CPU power consumption on long data read sequences (no 1LC operations).
Although exemplary embodiments of the present invention are described above, this description does not limit the scope of the invention, which can be practiced in a variety of embodiments.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US4268907 *||Jan 22, 1979||May 19, 1981||Honeywell Information Systems Inc.||Cache unit bypass apparatus|
|US4433374 *||Nov 14, 1980||Feb 21, 1984||Sperry Corporation||Cache/disk subsystem with cache bypass|
|US4500954 *||Oct 15, 1981||Feb 19, 1985||International Business Machines Corporation||Cache bypass system with post-block transfer directory examinations for updating cache and/or maintaining bypass|
|US4701844 *||Oct 16, 1986||Oct 20, 1987||Motorola Computer Systems, Inc.||Dual cache for independent prefetch and execution units|
|US4740889 *||Jul 14, 1986||Apr 26, 1988||Motorola, Inc.||Cache disable for a data processor|
|US4942518 *||Nov 12, 1985||Jul 17, 1990||Convex Computer Corporation||Cache store bypass for computer|
|US5247639 *||Jun 20, 1990||Sep 21, 1993||Nec Corporation||Microprocessor having cache bypass signal terminal|
|US5301295 *||May 22, 1991||Apr 5, 1994||Analog Devices, Inc.||Data processor apparatus and method with selective caching of instructions|
|US5371872 *||Oct 28, 1991||Dec 6, 1994||International Business Machines Corporation||Method and apparatus for controlling operation of a cache memory during an interrupt|
|US5390299 *||Dec 27, 1991||Feb 14, 1995||Digital Equipment Corporation||System for using three different methods to report buffer memory occupancy information regarding fullness-related and/or packet discard-related information|
|US5473756 *||Dec 30, 1992||Dec 5, 1995||Intel Corporation||FIFO buffer with full/empty detection by comparing respective registers in read and write circular shift registers|
|US5537552 *||Aug 17, 1993||Jul 16, 1996||Canon Kabushiki Kaisha||Apparatus for selectively comparing pointers to detect full or empty status of a circular buffer area in an input/output (I/O) buffer|
|US5539895 *||May 12, 1994||Jul 23, 1996||International Business Machines Corporation||Hierarchical computer cache system|
|US5561823 *||Mar 14, 1994||Oct 1, 1996||Conner Peripherals, Inc.||Monitor system for determining the available capacity of a READ buffer and a WRITE buffer in a disk drive system|
|US5581731 *||Oct 14, 1994||Dec 3, 1996||King; Edward C.||Method and apparatus for managing video data for faster access by selectively caching video data|
|US5594868 *||Mar 21, 1995||Jan 14, 1997||Hitachi, Ltd.||Processor unit for a parallel processor system discards a received packet when a reception buffer has insufficient space for storing the packet|
|US5625793 *||Apr 15, 1991||Apr 29, 1997||International Business Machines Corporation||Automatic cache bypass for instructions exhibiting poor cache hit ratio|
|1||Bates, Ken, "I/O subsystem performanc; cache implemented in the HSC can dramatically affect disk drive performance", DEC Professional, v12, n1, p.40(7), Jan. 1993.|
|2||*||Bates, Ken, I/O subsystem performanc; cache implemented in the HSC can dramatically affect disk drive performance , DEC Professional, v12, n1, p.40(7), Jan. 1993.|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US6055650 *||Apr 6, 1998||Apr 25, 2000||Advanced Micro Devices, Inc.||Processor configured to detect program phase changes and to adapt thereto|
|US6134634 *||Dec 19, 1997||Oct 17, 2000||Texas Instruments Incorporated||Method and apparatus for preemptive cache write-back|
|US6922754 *||Dec 8, 2003||Jul 26, 2005||Infabric Technologies, Inc.||Data-aware data flow manager|
|US7028143||Apr 15, 2003||Apr 11, 2006||Broadcom Corporation||Narrow/wide cache|
|US7076612 *||Apr 10, 2001||Jul 11, 2006||Koninklijke Philips Electronics N.V.||Cache interface circuit for automatic control of cache bypass modes and associated power savings|
|US7165144 *||Mar 19, 2004||Jan 16, 2007||Intel Corporation||Managing input/output (I/O) requests in a cache memory system|
|US7228388||Nov 19, 2004||Jun 5, 2007||International Business Machines Corporation||Enabling and disabling cache bypass using predicted cache line usage|
|US7930484 *||Apr 19, 2011||Advanced Micro Devices, Inc.||System for restricted cache access during data transfers and method thereof|
|US8606998||Aug 24, 2006||Dec 10, 2013||Advanced Micro Devices, Inc.||System and method for instruction-based cache allocation policies|
|US9235517 *||Aug 12, 2013||Jan 12, 2016||Globalfoundries Inc.||Implementing dynamic cache enabling and disabling based upon workload|
|US20040019747 *||Apr 15, 2003||Jan 29, 2004||Alphamosaic Limited||Narrow/wide cache|
|US20040117441 *||Dec 8, 2003||Jun 17, 2004||Infabric Technologies, Inc.||Data-aware data flow manager|
|US20050210202 *||Mar 19, 2004||Sep 22, 2005||Intel Corporation||Managing input/output (I/O) requests in a cache memory system|
|US20060112233 *||Nov 19, 2004||May 25, 2006||Ibm Corporation||Enabling and disabling cache bypass using predicted cache line usage|
|US20060179228 *||Feb 7, 2005||Aug 10, 2006||Advanced Micro Devices, Inc.||System for restricted cache access during data transfers and method thereof|
|US20060179231 *||Feb 7, 2005||Aug 10, 2006||Advanced Micron Devices, Inc.||System having cache memory and method of accessing|
|US20060179240 *||Feb 9, 2005||Aug 10, 2006||International Business Machines Corporation||System and method for algorithmic cache-bypass|
|US20080052466 *||Aug 24, 2006||Feb 28, 2008||Advanced Micro Devices, Inc.||System and method for instruction-based cache allocation policies|
|US20150046648 *||Aug 12, 2013||Feb 12, 2015||International Business Machines Corporation||Implementing dynamic cache enabling and disabling based upon workload|
|U.S. Classification||711/138, 711/E12.021, 711/118|
|Cooperative Classification||Y02B60/1225, G06F12/0888|
|Jun 2, 1995||AS||Assignment|
Owner name: TEXAS INSTRUMENTS INCORPORATED, TEXAS
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LEYRER, THOMAS;REEL/FRAME:007664/0994
Effective date: 19950515
|Aug 29, 2001||FPAY||Fee payment|
Year of fee payment: 4
|Aug 26, 2005||FPAY||Fee payment|
Year of fee payment: 8
|Aug 21, 2009||FPAY||Fee payment|
Year of fee payment: 12