|Publication number||US6675238 B1|
|Application number||US 09/390,041|
|Publication date||Jan 6, 2004|
|Filing date||Sep 3, 1999|
|Priority date||Sep 3, 1999|
|Publication number||09390041, 390041, US 6675238 B1, US 6675238B1, US-B1-6675238, US6675238 B1, US6675238B1|
|Inventors||Jerrie L. Coffman, Arlin R. Davis|
|Original Assignee||Intel Corporation|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (11), Referenced by (12), Classifications (11), Legal Events (6)|
|External Links: USPTO, USPTO Assignment, Espacenet|
The present invention describes a method and apparatus by which interrupts may be eliminated in a pipelined computer system using a communication system and a polling system to efficiently utilize computer time and memory usage.
Modern microprocessor instruction execution involves the use of a series of discrete stages called a “pipeline ” to break problems into logically organized pieces. By executing processes on a pipeline, the processor (e.g., a “CPU ”) can work on several different problems at once, thus greatly increasing the rate of process completion.
In prior art systems, input/output (“I/O ”) requests are not processed in the normal CPU pipeline system. Instead, pending I/O processes obtain the attention of the CPU by signaling an interrupt, which causes the CPU to stop whatever process it is currently executing in order to address the needs of the pending I/O process. When an interrupt is signaled, a “context switch ” is triggered which forces the CPU to save its current status to registers and non-local memory before processing the pending I/O request. After the CPU processes the pending I/O request, the status of the CPU that existed prior to the interrupt is retrieved from the registers and non-local memory and the CPU resumes processing using the normal pipeline system. These context switches result in a significant amount of execution time loss because they force the CPU to spend time in memory access and force the operating system (“OS ”) to reschedule other active processes to make room for I/O completion processing. Furthermore, typical I/O completion processing uses slow, non-cacheable memory-mapped registers located on I/O adapters across an I/O expansion bus, thus loading the system and I/O buses with further overhead.
Consequently, there is a need in the art for a method and apparatus of I/O process completion that does not require an interruption in the normal activities of the OS, the processor pipeline, or the system and I/O buses.
Embodiments of the present invention provide for an apparatus for input/output processing that includes a plurality of descriptors where each descriptor includes a completion indicator and data associated with an input/output request. The plurality of descriptors includes a head descriptor and a tail descriptor. Embodiments of the present invention further include a plurality of address holders associated with an input/output processor, and each the plurality of address holders is uniquely affiliated with one of the plurality of descriptors. Embodiments of the present invention further include a polling mechanism for evaluating the completion indicator of the head descriptor and a completion processor for interfacing with the head descriptor. Finally, embodiments of the present invention include connectors between the tail descriptor and address holder and between the input/output processor and the head descriptor.
FIG. 1 details an apparatus of an I/O transfer request flow mechanism.
FIG. 2 details a flowchart of an I/O completion polling thread algorithm.
The present invention solves the problem of pipeline disruption via I/O interrupts by utilizing a communications system and a software polling system to efficiently utilize time and memory allocated to I/O completion processing.
An I/O transfer request flow diagram 100 illustrating the performance of software polling in accordance with an embodiment of the present invention is shown in FIG. 1. FIG. 1 illustrates a host memory 140 including a host-in-process queue 110 containing a plurality of data transfer descriptors 150, 160, 170, 180, 190 arranged in linked list format from the head data transfer descriptor 150 to the tail data transfer descriptor 190. The host memory 140 may contain a kernel mode device driver.
FIG. 1 also illustrates an I/O device 120 including an I/O processor 260 and an I/O address list 130 with a plurality of addresses 210, 220, 230, 240, 250 arranged in a first-in-first-out (“FIFO ”) format. The I/O address list 130 may consist of memory mapped registers and is arranged in a FIFO format such that the I/O processor 260 first processes the address 250 that has been in the I/O address list 130 the longest before processing the address 240 that has been in the I/O address list the second longest, and so on throughout the addresses 210-250. The addresses 210-250 operate as pointers because their values represent locations in the host memory 140 where their corresponding data transfer descriptors 150-190 may be found.
The I/O address list 130 may receive via connector 200 new addresses 210 that correspond to the most recently added data transfer descriptor 190 to the host-in-process queue. Additionally, once the I/O processor 260 completes processing the first address 250 in the I/O address list, the I/O processor 260 may update its corresponding data transfer descriptor 150 via connector 270 to indicate that the processing of the information in the first data transfer descriptor is complete.
To begin an I/O request, the CPU builds a data transfer descriptor 190 that contains all of the information the I/O device 120 needs to process the I/O request. Such information may include the type of I/O device utilized, the nature of the data required to be input into or output from the CPU, the length and address of the data on which the I/O will operate, the address of the next data transfer descriptor 210 in the host in-process queue 110 and any other pertinent information. The data transfer descriptor 190 also includes an indicator for determining whether the CPU's I/O request has been completed or not. The data transfer descriptors 150-190 are located within the host memory 140. The host memory 140 may be located within the main CPU memory, which would allow the CPU to take full advantage of CPU caching and limit the amount of system bus traffic generated during I/O completion checking.
After the CPU builds a new data transfer request 190, the host in-process queue 110 chains the data transfer descriptor 190 on the tail of the next preexisting data transfer descriptor 180. Subsequently, the host in-process queue 110 writes the memory address of the new data transfer descriptor 190 in the last address 210 of the I/O address list 130.
The operation of the I/O device 120 is triggered by the presence of addresses 210-250 in the I/O address list 130. Running concurrently with the CPU's creation of data transfer descriptors 150-190, the I/O device 120 removes the first data transfer descriptor address 250 from the I/O address list. The I/O processor 260 subsequently accesses the data transfer descriptor 150 located at the first data transfer descriptor address 250. When the I/O processor 260 completes the I/O operation as directed by the data transfer descriptor 150, the I/O device 120 updates a field in the data transfer descriptor 150 via connector 270 to indicate that the I/O transaction is complete. The I/O device 120 continues processing the next address 240 in the I/O address list 130 in the same manner until the I/O address list 130 is empty or a server error occurs.
An I/O transfer completion polling thread 10 illustrating the performance of software polling in accordance with an embodiment of the present invention is shown in FIG. 2. The I/O transfer completion polling thread 10 may be a kernel thread created by a host device driver to provide the necessary polling time slice and system task load balancing. The I/O transfer completion polling thread 10 consists of a polling loop 20 and an I/O completion branch 30. The polling loop 20 determines whether the I/O operation detailed in the data transfer descriptor 150 at the head of the host in-process queue 110 has been completed by the I/O processor 260 (Step 40). If the I/O operation detailed in the data transfer descriptor 150 is not present, or is not complete, the polling loop 20 yields to other system threads running in the CPU (Step 50) until it is rescheduled by the OS (Step 80).
If the I/O operation detailed in the head data transfer descriptor 150 is present and is complete (Step 40), the I/O completion branch 30 is chosen. The CPU then removes the data transfer descriptor 150 from the head of the host in-process queue 110 (Step 60) and performs any necessary I/O completion processing (Step 70) prompted by the result of the I/O request found in the data transfer descriptor 150. The I/O completion branch 30 concludes (Step 90) by returning to the polling loop 20 to determine whether the I/O operation detailed in the next data transfer descriptor 160 in the host in-process queue is present and is complete (Step 40).
In one embodiment of the present invention, the Next Generation I/O (“NGIO ”) communication interface may be utilized for implementing the I/O completion polling thread 10. NGIO guarantees that completed data transfer descriptors 150-190 will appear at the head of the host in-process queue 110, thereby saving search time when the polling thread 10 is ready to remove the head data transfer descriptor 150.
The software polling, as implemented by the present invention, makes the CPU more efficient in a number of ways. First, software polling provides an I/O communication mechanism that significantly decreases both system and I/O bus overhead by polling in cacheable memory space, which is significantly faster than the non-cacheable memory used in the prior art context switching system. This is because the host memory 140, which contains the host in-process queue 110 may be placed in the main, cacheable CPU memory. Second, software polling eliminates the need to service interrupts for I/O devices, thus allowing the CPU to devote more time to efficiently executing processes on a pipeline. Third, the software polling, because it runs at the same priority level as applications, allows the CPU to perform other tasks when there are no I/O transfers to complete. Fourth, software polling provides the capability for efficient multi-threaded processing that can direct I/O completion to the CPU that initiated the I/O request. Multiple software polling threads 10 can run simultaneously and the distribution of these threads are managed based on the number of system CPUs. Completing the I/O on the CPU that initiated the request increases the probability that the process context data is still in the relatively fast CPU cache, resulting in greater system performance.
Accordingly, the present invention allows for the efficient processing of I/O requests for a CPU executing a pipeline of instructions. It will be appreciated by those skilled in the art that the specific embodiments disclosed above may be readily utilized as a basis for modifying or designing other methods and techniques for carrying out the same purposes of the present invention. It should also be realized by those skilled in the art that such equivalent constructions do not depart from the spirit and scope of the invention as set forth in the following claims.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US5594862 *||Jul 20, 1994||Jan 14, 1997||Emc Corporation||XOR controller for a storage subsystem|
|US5606665 *||Jul 1, 1994||Feb 25, 1997||Digital Equipment Corporation||Buffer descriptor prefetch in network and I/O design|
|US5745790 *||Mar 6, 1997||Apr 28, 1998||Sun Microsystems, Inc.||Method and apparatus for reporting the status of asynchronous data transfer|
|US5751951 *||Oct 30, 1995||May 12, 1998||Mitsubishi Electric Information Technology Center America, Inc.||Network interface|
|US5828901 *||Dec 21, 1995||Oct 27, 1998||Cirrus Logic, Inc.||Method and apparatus for placing multiple frames of data in a buffer in a direct memory access transfer|
|US6006275 *||Oct 6, 1998||Dec 21, 1999||Compaq Computer Corporation||Network connector operable in bridge mode and bypass mode|
|US6031843 *||Nov 21, 1996||Feb 29, 2000||Alcatel Data Networks Inc.||Digital communications switching fabric|
|US6049842 *||May 1, 1997||Apr 11, 2000||International Business Machines Corporation||Efficient data transfer mechanism for input/output devices|
|US6070219 *||Oct 9, 1996||May 30, 2000||Intel Corporation||Hierarchical interrupt structure for event notification on multi-virtual circuit network interface controller|
|US6333929 *||Aug 24, 1998||Dec 25, 2001||Intel Corporation||Packet format for a distributed system|
|US6351474 *||Jan 14, 1998||Feb 26, 2002||Skystream Networks Inc.||Network distributed remultiplexer for video program bearing transport streams|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US6973513 *||Apr 9, 2001||Dec 6, 2005||3Com Corporation||Method for efficient use of a transmit engine|
|US6988268 *||Apr 30, 2002||Jan 17, 2006||Microsoft Corporation||IO completion architecture for user-mode networking|
|US7260661 *||Sep 3, 2004||Aug 21, 2007||Intel Corporation||Processing replies to request packets in an advanced switching context|
|US7370123 *||Oct 11, 2005||May 6, 2008||Nec Electronics Corporation||Information processing apparatus|
|US7478394 *||Jun 4, 2001||Jan 13, 2009||Hewlett-Packard Development Company, L.P.||Context-corrupting context switching|
|US8495261 *||Dec 12, 2008||Jul 23, 2013||International Business Machines Corporation||Redispatching suspended tasks after completion of I/O operations absent I/O interrupts|
|US20030204552 *||Apr 30, 2002||Oct 30, 2003||Microsoft Corporation||IO completion architecture for user-mode networking|
|US20060050693 *||Sep 3, 2004||Mar 9, 2006||James Bury||Building data packets for an advanced switching fabric|
|US20060050694 *||Sep 3, 2004||Mar 9, 2006||James Bury||Processing replies to request packets in an advanced switching context|
|US20060050722 *||Sep 3, 2004||Mar 9, 2006||James Bury||Interface circuitry for a receive ring buffer of an as fabric end node device|
|US20060080479 *||Oct 11, 2005||Apr 13, 2006||Nec Electronics Corporation||Information processing apparatus|
|US20100153605 *||Dec 12, 2008||Jun 17, 2010||International Business Machines Corporation||Redispatching suspended tasks after completion of i/o operations absent i/o interrupts|
|U.S. Classification||710/46, 710/5, 711/E12.019, 710/10|
|International Classification||G06F3/00, G06F9/46, G06F12/08|
|Cooperative Classification||G06F12/0866, G06F9/52|
|European Classification||G06F9/52, G06F12/08B12|
|Sep 3, 1999||AS||Assignment|
Owner name: INTEL CORPORATION, CALIFORNIA
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:COFFMAN, JERRIE L.;DAVIS, ARLIN R.;REEL/FRAME:010234/0736
Effective date: 19990901
|Jun 29, 2007||FPAY||Fee payment|
Year of fee payment: 4
|Jun 29, 2011||FPAY||Fee payment|
Year of fee payment: 8
|Aug 14, 2015||REMI||Maintenance fee reminder mailed|
|Jan 6, 2016||LAPS||Lapse for failure to pay maintenance fees|
|Feb 23, 2016||FP||Expired due to failure to pay maintenance fee|
Effective date: 20160106