WO2003058434A1 - Suspending execution of a thread in a multi-threaded - Google Patents

Suspending execution of a thread in a multi-threaded Download PDF

Info

Publication number
WO2003058434A1
WO2003058434A1 PCT/US2002/039790 US0239790W WO03058434A1 WO 2003058434 A1 WO2003058434 A1 WO 2003058434A1 US 0239790 W US0239790 W US 0239790W WO 03058434 A1 WO03058434 A1 WO 03058434A1
Authority
WO
WIPO (PCT)
Prior art keywords
thread
processor
resources
instruction
selected amount
Prior art date
Application number
PCT/US2002/039790
Other languages
French (fr)
Inventor
Deborah Marr
Scott Rodgers
David Hill
Shivnandan Kaushik
James Crossland
David Koufaty
Original Assignee
Intel Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intel Corporation filed Critical Intel Corporation
Priority to JP2003558678A priority Critical patent/JP2005514698A/en
Priority to KR1020047010393A priority patent/KR100617417B1/en
Priority to DE10297597T priority patent/DE10297597T5/en
Priority to AU2002364559A priority patent/AU2002364559A1/en
Publication of WO2003058434A1 publication Critical patent/WO2003058434A1/en
Priority to HK05107419A priority patent/HK1075109A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30098Register arrangements
    • G06F9/3012Organisation of register space, e.g. banked or distributed register file
    • G06F9/30123Organisation of register space, e.g. banked or distributed register file according to context, e.g. thread buffers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30003Arrangements for executing specific machine instructions
    • G06F9/30076Arrangements for executing specific machine instructions to perform miscellaneous control operations, e.g. NOP
    • G06F9/3009Thread control instructions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3836Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution
    • G06F9/3851Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution from multiple instruction streams, e.g. multistreaming

Definitions

  • the present disclosure pertains to the field of processors. More particularly, the
  • present disclosure pertains to multi-threaded processors and techniques for temporarily suspending the processing of one thread in a multi-threaded processor.
  • a multi-threaded processor is capable of processing multiple different instruction
  • a primary motivating factor driving execution of multiple instruction streams within a single processor is the resulting improvement in processor utilization.
  • Highly parallel architectures have developed over the years, but it is often difficult to extract sufficient parallelism from a single stream of instructions to utilize the multiple execution units.
  • Simultaneous multi-threading processors allow multiple 25 instruction streams to execute concurrently in the different execution resources in an attempt to better utilize those resources. Multi-threading can be particularly advantageous for programs that encounter high latency delays or which often wait for events to occur. When one thread is waiting for a high latency task to complete or for a particular event, a different thread may be processed.
  • Figure 1 illustrates one embodiment of a multi-threaded processor having logic to suspend a thread in response to an instruction and to relinquish resources associated with that thread.
  • Figure 2 is a flow diagram illustrating operation of the multi-threaded processor of
  • Figure 1 illustrates various options for specifying an amount of time a multithreading processor may be suspended.
  • Figure 3b illustrates a flow diagram in which the suspended state may be exited by either the elapse of a selected amount of time or the occurrence of an event.
  • Figure 4 illustrates resource patititioning, sharing, and duplication according to one embodiment.
  • Figure 5 illustrates various design representations or formats for simulation, emulation, and fabrication of a design using the disclosed techniques.
  • the disclosed techniques may allow a programmer to implement a suspend mechanism in one thread while letting other threads harness processing resources. Thus, partitions previously dedicated to the suspended thread may be relinquished while the
  • Figure 1 illustrates one embodiment of a multi-threaded processor 100 having suspend logic 110 to allow a thread to be suspended in response to an instruction.
  • a "processor” may be formed as a single integrated circuit in some embodiments. In other 15 embodiments, multiple integrated circuits may together form a processor, and in yet other embodiments, hardware and software routines (e.g., binary translation routines) may together form the processor.
  • the suspend logic may be microcode, various forms of control logic, or other implementation of the described functionality, possibly including translation, software, etc.
  • the processor 100 is coupled to a memory 195 to allow the processor to retrieve instructions from the memory 195 and to execute these instructions.
  • the memory and the processor may be coupled in a point-to-point fashion, via bus bridges, via a memory controller or via other known or otherwise available techniques.
  • the memory 195 stores various program threads, including a first thread 196 and a second thread 198.
  • the first 25 thread 196 includes a SUSPEND instruction.
  • a bus/memory controller 120 provides instructions for execution to a front end 130.
  • the front end 130 directs the retrieval of instructions from various threads according to instruction pointers 170. Instruction pointer logic is replicated to support multiple threads.
  • the front end 130 feeds instructions into thread 5 partitionable resources 140 for further processing.
  • the thread partitionable resources 140 include logically separated partitions dedicated to particular threads when multiple threads are active within the processor 100. In one embodiment, each separate partition only contains instructions from the thread to which that portion is dedicated.
  • the thread partitionable resources 140 may include, for example, instruction queues. When in a 10 single thread mode, the partitions of the thread partitionable resources 140 may be combined to form a single large partition dedicated to the one thread.
  • the processor 100 also includes replicated state 180.
  • the replicated state 180 includes state variables sufficient to maintain context for a logical processor. With replicated state 180, multiple threads can execute without competition for state variable 15 storage. Additionally, register allocation logic may be replicated for each thread. The replicated state-related logic operates with the appropriate resource partitions to prepare incoming instructions for execution.
  • the thread partitionable resources 140 pass instructions along to shared resources
  • the shared resources 150 operate on instructions without regard to their origin.
  • scheduler and execution units may be thread-unaware shared resources.
  • the partitionable resources 140 may feed instructions from multiple threads to the shared resources 150 by alternating between the threads in a fair manner that provides continued progress on each active thread.
  • the shared resources may execute the provided instructions on the appropriate state without concern for the thread mix.
  • the shared resources 150 may be followed by another set of thread partitionable resources 160.
  • the thread partitionable resources 160 may include retirement resources such as a re-order buffer and the like. Accordingly, the thread partitionable resources 160 may ensure that execution of instructions from each thread concludes properly and that the appropriate state for that thread is appropriately updated.
  • the processor 100 of Figure 1 includes the suspend logic 110.
  • the suspend logic 110 may be programmable to provide a particular duration for which the thread is to be suspended or to provide a fixed delay.
  • the suspend 10 logic 110 includes pipeline flush logic 112 and partition/anneal logic 114.
  • the instruction set of the processor 100 includes a SUSPEND opcode (instruction) to cause thread suspension.
  • the SUSPEND opcode is received as a part of the sequence of instructions of a 15 first thread (Tl).
  • Thread Tl execution is suspended as indicated in block 210.
  • the thread suspend logic 110 includes pipeline flush logic 112, which drains the processor pipeline in order to clear all instructions as indicated in block 220.
  • partition/anneal logic 114 causes any partitioned resources associated exclusively with thread Tl to be relinquished for use by other threads as 20 indicated in block 230. These relinquished resources are annealed to form a set of larger resources for the remaining active threads to utilize.
  • processor resources may continue to be utilized, substantially without interference from
  • the processor may relinquish some or all of the partitions of partitionable resources 140 and 160 that were dedicated to Tl.
  • different permutations of the SUSPEND opcode or settings associated therewith may indicate which resources to relinquish, if any. For example, when a programmer
  • the thread may be suspended, but maintain most resource partitions. Throughput is still enhanced because the shared resources may be used exclusively by other threads during the thread suspension period. When a longer wait is anticipated, relinquishing all partitions associated with the suspended thread allows other threads to have additional resources, potentially increasing the throughput of the other
  • a test is performed to determine if the suspend state should be exited.
  • the thread may
  • a processor 300 may include a delay time (Dl) specified by a routine of microcode 310.
  • a timer or counter 312 may implement the delay and signal the microcode when the specified amount of time has elapsed.
  • one or more fuses 330 may be used to specify a delay (D2), or a register 340
  • a delay (D3) may store a delay (D3).
  • a delay (D4) may be specified by a register or storage location such as a configuration register in a bridge or memory controller 302 which is coupled to the processor.
  • a delay (D5) may also be specified by the basic input/output system (BIOS) 322.
  • the delay (D6) could be stored in a memory 304 which is coupled to the memory controller 302.
  • the processor 300 may retrieve the delay value as 5 an implicit or explicit operand to the SUSPEND opcode as it is executed by an execution unit 320. Other known or otherwise available or convenient techniques of specifying a value may be used to specify the delay as well. [0026] Referring back to Figure 2, if the delay time has not elapsed, then the timer, counter, or other delay-measuring mechanism used continues to track the delay, and the
  • thread Tl 10 thread remains suspended, as indicated by the return to block 240. If the delay time has elapsed, then thread Tl resumption begins in block 250. As indicated in block 250, the pipeline is flushed, to free resources for thread Tl. In block 260, resources are re- partitioned such that thread Tl has portions of the thread-partitionable resources with which to perform operations. Finally, thread Tl re-starts execution, as indicated in block
  • Figures 1 and 2 provide techniques to allow a thread to be suspended by a program for a particular duration.
  • other events also cause Tl to be resumed.
  • an interrupt may cause Tl to resume.
  • 3b illustrates a flow diagram for one embodiment that allows other events to cause the
  • 25 state-breaking events are detected in blocks 370 and 375.
  • block 370 tests whether any (and in some embodiments which) events are enabled to break the suspend state. If no events are enabled to break the suspend state, then the process returns to block
  • thread Tl is resumed, as indicated in block 380. Otherwise, the processor remains with thread Tl in the suspended state, and the process returns to block 365.
  • Figure 4 illustrates the partitioning, duplication, and sharing of resources according to one embodiment. Partitioned resources may be partitioned and annealed (fused back
  • duplicated resources include instruction pointer logic in the instruction fetch portion of the pipeline, register renaming logic in the rename portion of the pipeline, state variables (not shown, but referenced in various stages in the pipeline), and an interrupt controller (not shown, generally asynchronous to pipeline).
  • Shared resources in the embodiment of Figure 4 include schedulers in the schedule stage of the pipeline, a pool of registers in the register read and write portions of the pipeline, execution resources in the execute portion of the pipeline. Additionally, a trace cache and an LI data cache may be shared resources populated according to memory accesses without regard to thread context. In other embodiments, consideration of thread context
  • Partitioned resources in the embodiment of Figure 4 include two queues in queuing stages of the pipeline, a re-order buffer in a retirement stage of the pipeline, and a store buffer. Thread selection multiplexing logic alternates between the various duplicated and partitioned resources to provide reasonable access to both threads.
  • the partitionable resources may not be strictly partitioned, but rather may allow some instructions to cross partitions or may allow partitions to vary in size depending on the thread being executed in that partition or the total number of threads being executed. Additionally, different mixes of resources may be designated as shared, duplicated, and partitioned resources.
  • Figure 5 illustrates various design representations or formats for simulation, emulation, and fabrication of a design using the disclosed techniques.
  • Data representing a design may represent the design in a number of manners.
  • the hardware may be represented using a hardware description language or another functional description language which essentially provides a computerized model of how the designed hardware is expected to perform.
  • the hardware model 1110 may be stored in a storage medium 1100 such as a computer memory so that the model may be simulated using simulation software 1120 that applies a particular test suite 1130 to the hardware model 1110 to determine if it indeed functions as intended.
  • the simulation software is not recorded, captured, or contained in the medium.
  • a circuit level model with logic and/or transistor gates may be produced at some stages of the design process.
  • This model may be similarly simulated, sometimes by dedicated hardware simulators that form the model using programmable logic. This type of simulation, taken a degree further, may be an emulation technique.
  • re-configurable hardware is another embodiment that may involve a machine readable medium storing a model employing the disclosed techniques.
  • the data representing the hardware model may be the data specifying the presence or absence of various features on different mask layers for masks used to produce the integrated circuit.
  • this data representing the integrated circuit embodies the techniques disclosed in that the circuitry or logic in the data can be simulated or fabricated to perform these techniques.
  • the data may be stored in any form of a computer readable medium.
  • An optical or electrical wave 1160 modulated or otherwise generated to transmit such information, a memory 1150, or a magnetic or optical storage 1140 such as a disc may be the medium.
  • the set of bits describing the design or the particular part of the design are an article that may be sold in and of itself or used by others for further design or fabrication. Thus, techniques for suspending execution of a thread in a multi-threaded processor are disclosed.

Abstract

Techniques for suspending execution of a thread in a multi-threaded processor. In one embodiment, a processor includes resources that can be partitioned between multiple threads. Processor logic receives an instruction in a first thread of execution, and, in response to that instruction, relinquishes portions of the portioned resources for use by other threads.

Description

SUSPENDINGEXECUTIONOFATHREADINAMULTI-THREADED
PROCESSOR
RELATED APPLICATIONS
[0001] This application is related to Application Serial Number __/___, , entitled "A
5 Method and Apparatus for Suspending Execution of a Thread Until a Specified Memory
Access Occurs"; Application Serial Number __/___, , entitled "Coherency Techniques for Suspending Execution of a Thread Until a Specified Memory Access Occurs";
Application Serial Number __/__, , entitled "Instruction Sequences for Suspending
Execution of a Thread Until a Specified Memory Access Occurs" all filed on the same date 10 as the present application.
BACKGROUND
1. Field
[0002] The present disclosure pertains to the field of processors. More particularly, the
15 present disclosure pertains to multi-threaded processors and techniques for temporarily suspending the processing of one thread in a multi-threaded processor.
2. Description of Related Art
[0003] A multi-threaded processor is capable of processing multiple different instruction
20 sequences concurrently. A primary motivating factor driving execution of multiple instruction streams within a single processor is the resulting improvement in processor utilization. Highly parallel architectures have developed over the years, but it is often difficult to extract sufficient parallelism from a single stream of instructions to utilize the multiple execution units. Simultaneous multi-threading processors allow multiple 25 instruction streams to execute concurrently in the different execution resources in an attempt to better utilize those resources. Multi-threading can be particularly advantageous for programs that encounter high latency delays or which often wait for events to occur. When one thread is waiting for a high latency task to complete or for a particular event, a different thread may be processed.
[0ft04] Many different techniques have been proposed to control when a processor switches between threads. For example, some processors detect particular long latency events such as L2 cache misses and switch threads in response to these detected long latency events. While detection of such long latency events may be effective in some circumstances, such event detection is unlikely to detect all points at which it may be efficient to switch threads. 10 In particular, event based thread switching may fail to detect points in a program where delays are intended by the programmer.
[0005] In fact, often, the programmer is in the best position to determine when it would be efficient to switch threads to avoid wasteful spin-wait loops or other resource-consuming delay techniques. Thus, allowing programs to control thread switching may enable 15 programs to operate more efficiently. Explicit program instructions that affect thread selection may be advantageous to this end. For example, a "Pause" instruction is described in US Patent Application No. 09/489,130, filed 1/21/2000. The Pause instruction allows a thread of execution to be temporarily suspended either until a count is reached or until an instruction has passed through the processor pipeline. The Pause instruction described in 20 the above-referenced application, however, does not specify that thread partitionable resources are to be relinquished. Different techniques may be useful in allowing programmers to more efficiently harness the resources of a multi-threaded processor. Brief Description of the Figures
[0006] The present invention is illustrated by way of example and not limitation in the
5 Figures of the accompanying drawings. [0007] Figure 1 illustrates one embodiment of a multi-threaded processor having logic to suspend a thread in response to an instruction and to relinquish resources associated with that thread. [0008] Figure 2 is a flow diagram illustrating operation of the multi-threaded processor of
10 Figure 1 according to one embodiment. [0009] Figure 3 a illustrates various options for specifying an amount of time a multithreading processor may be suspended. [0010] Figure 3b illustrates a flow diagram in which the suspended state may be exited by either the elapse of a selected amount of time or the occurrence of an event. [0M1] Figure 4 illustrates resource patititioning, sharing, and duplication according to one embodiment. [0012] Figure 5 illustrates various design representations or formats for simulation, emulation, and fabrication of a design using the disclosed techniques.
Detailed Description 20
[0013] The following description describes techniques for suspending execution of a thread in a multi-threaded processor. In the following description, numerous specific details such as logic implementations, opcodes, means to specify operands, resource partitioning/sharing/duplication implementations, types and interrelationships of system
25 components, and logic partitioning/integration choices are set forth in order to provide a more thorough understanding of the present invention. It will be appreciated, however, by one skilled in the art that the invention may be practiced without such specific details. In other instances, control structures, gate level circuits and full software instruction sequences have not been shown in detail in order not to obscure the invention. Those of 5 ordinary skill in the art, with the included descriptions, will be able to implement appropriate functionality without undue experimentation.
[0014] The disclosed techniques may allow a programmer to implement a suspend mechanism in one thread while letting other threads harness processing resources. Thus, partitions previously dedicated to the suspended thread may be relinquished while the
10 thread is suspended. These and/or other disclosed techniques may advantageously improve overall processor throughput.
[0015] Figure 1 illustrates one embodiment of a multi-threaded processor 100 having suspend logic 110 to allow a thread to be suspended in response to an instruction. A "processor" may be formed as a single integrated circuit in some embodiments. In other 15 embodiments, multiple integrated circuits may together form a processor, and in yet other embodiments, hardware and software routines (e.g., binary translation routines) may together form the processor. The suspend logic may be microcode, various forms of control logic, or other implementation of the described functionality, possibly including translation, software, etc.
[0-016] The processor 100 is coupled to a memory 195 to allow the processor to retrieve instructions from the memory 195 and to execute these instructions. The memory and the processor may be coupled in a point-to-point fashion, via bus bridges, via a memory controller or via other known or otherwise available techniques. The memory 195 stores various program threads, including a first thread 196 and a second thread 198. The first 25 thread 196 includes a SUSPEND instruction. [0017] In the embodiment of Figure 1, a bus/memory controller 120 provides instructions for execution to a front end 130. The front end 130 directs the retrieval of instructions from various threads according to instruction pointers 170. Instruction pointer logic is replicated to support multiple threads. The front end 130 feeds instructions into thread 5 partitionable resources 140 for further processing. The thread partitionable resources 140 include logically separated partitions dedicated to particular threads when multiple threads are active within the processor 100. In one embodiment, each separate partition only contains instructions from the thread to which that portion is dedicated. The thread partitionable resources 140 may include, for example, instruction queues. When in a 10 single thread mode, the partitions of the thread partitionable resources 140 may be combined to form a single large partition dedicated to the one thread.
[0018] The processor 100 also includes replicated state 180. The replicated state 180 includes state variables sufficient to maintain context for a logical processor. With replicated state 180, multiple threads can execute without competition for state variable 15 storage. Additionally, register allocation logic may be replicated for each thread. The replicated state-related logic operates with the appropriate resource partitions to prepare incoming instructions for execution.
[0019] The thread partitionable resources 140 pass instructions along to shared resources
150. The shared resources 150 operate on instructions without regard to their origin. For 20 example, scheduler and execution units may be thread-unaware shared resources. The partitionable resources 140 may feed instructions from multiple threads to the shared resources 150 by alternating between the threads in a fair manner that provides continued progress on each active thread. Thus, the shared resources may execute the provided instructions on the appropriate state without concern for the thread mix.
[Qffi.0] The shared resources 150 may be followed by another set of thread partitionable resources 160. The thread partitionable resources 160 may include retirement resources such as a re-order buffer and the like. Accordingly, the thread partitionable resources 160 may ensure that execution of instructions from each thread concludes properly and that the appropriate state for that thread is appropriately updated.
[0021] As previously mentioned, it may be desirable to provide programmers with a technique to implement a delay without requiring constant polling of a memory location or even execution of a loop of instructions. Thus, the processor 100 of Figure 1 includes the suspend logic 110. The suspend logic 110 may be programmable to provide a particular duration for which the thread is to be suspended or to provide a fixed delay. The suspend 10 logic 110 includes pipeline flush logic 112 and partition/anneal logic 114.
[0022] The operations of the embodiment of Figure 1 may be further explained with reference to the flow diagram of Figure 2. In one embodiment, the instruction set of the processor 100 includes a SUSPEND opcode (instruction) to cause thread suspension. In block 200, the SUSPEND opcode is received as a part of the sequence of instructions of a 15 first thread (Tl). Thread Tl execution is suspended as indicated in block 210. The thread suspend logic 110 includes pipeline flush logic 112, which drains the processor pipeline in order to clear all instructions as indicated in block 220. In one embodiment, once the pipeline has been drained, partition/anneal logic 114 causes any partitioned resources associated exclusively with thread Tl to be relinquished for use by other threads as 20 indicated in block 230. These relinquished resources are annealed to form a set of larger resources for the remaining active threads to utilize.
[0023] As indicated in block 235, other threads may be executed (assuming instructions are available for execution) during the time in which thread Tl is suspended. Thus, processor resources may continue to be utilized, substantially without interference from
25 thread Tl. Dedication of the processor resources more fully to other threads may advantageously expedite processing of other useful execution streams when thread Tl has little or no useful work to accomplish, or when a program decides that completing tasks in thread Tl is not a priority. [0024] In general, with thread Tl suspended, the processor enters an implementation
5 dependent state which allows other threads to more fully utilize the processor resources. In some embodiments, the processor may relinquish some or all of the partitions of partitionable resources 140 and 160 that were dedicated to Tl. In other embodiments, different permutations of the SUSPEND opcode or settings associated therewith may indicate which resources to relinquish, if any. For example, when a programmer
10 anticipates a shorter wait, the thread may be suspended, but maintain most resource partitions. Throughput is still enhanced because the shared resources may be used exclusively by other threads during the thread suspension period. When a longer wait is anticipated, relinquishing all partitions associated with the suspended thread allows other threads to have additional resources, potentially increasing the throughput of the other
15 threads. The additional throughput, however, comes at the cost of the overhead associated with removing and adding partitions when threads are respectively suspended and resumed.
[0025] In block 240, a test is performed to determine if the suspend state should be exited.
If the specified delay has occurred (i.e., sufficient time has elapsed), then the thread may
20 be resumed. The time for which the thread is suspended may be specified in a number of manners, as shown in Figure 3a. For example, a processor 300 may include a delay time (Dl) specified by a routine of microcode 310. A timer or counter 312 may implement the delay and signal the microcode when the specified amount of time has elapsed. Alternatively, one or more fuses 330 may be used to specify a delay (D2), or a register 340
25 may store a delay (D3). A delay (D4) may be specified by a register or storage location such as a configuration register in a bridge or memory controller 302 which is coupled to the processor. A delay (D5) may also be specified by the basic input/output system (BIOS) 322. Alternatively still, the delay (D6) could be stored in a memory 304 which is coupled to the memory controller 302. The processor 300 may retrieve the delay value as 5 an implicit or explicit operand to the SUSPEND opcode as it is executed by an execution unit 320. Other known or otherwise available or convenient techniques of specifying a value may be used to specify the delay as well. [0026] Referring back to Figure 2, if the delay time has not elapsed, then the timer, counter, or other delay-measuring mechanism used continues to track the delay, and the
10 thread remains suspended, as indicated by the return to block 240. If the delay time has elapsed, then thread Tl resumption begins in block 250. As indicated in block 250, the pipeline is flushed, to free resources for thread Tl. In block 260, resources are re- partitioned such that thread Tl has portions of the thread-partitionable resources with which to perform operations. Finally, thread Tl re-starts execution, as indicated in block
15 270.
[0027] Thus, the embodiments of Figures 1 and 2 provide techniques to allow a thread to be suspended by a program for a particular duration. In one embodiment, other events also cause Tl to be resumed. For example, an interrupt may cause Tl to resume. Figure
3b illustrates a flow diagram for one embodiment that allows other events to cause the
20 suspend state to be exited. In block 360, the thread is already suspended according to previous operations. In block 370, whether sufficient time has elapsed (as previously discussed with respect to Figure 2) is tested. In the event that sufficient time has elapsed, then thread Tl is resumed, as indicated in block 380.
[0028] On the other hand, if insufficient time has elapsed in block 365, then any suspend-
25 state-breaking events are detected in blocks 370 and 375. In some embodiments, there may be operands, configuration setting, permutations of the SUSPEND instruction, etc., that specify which if any events cause the suspend state to be exited. Thus, block 370 tests whether any (and in some embodiments which) events are enabled to break the suspend state. If no events are enabled to break the suspend state, then the process returns to block
5 365. If any of the enabled events occurs, as tested in block 375, then thread Tl is resumed, as indicated in block 380. Otherwise, the processor remains with thread Tl in the suspended state, and the process returns to block 365.
[0029] Figure 4 illustrates the partitioning, duplication, and sharing of resources according to one embodiment. Partitioned resources may be partitioned and annealed (fused back
10 together for re-use by other threads) according to the ebb and flow of active threads in the machine. In the embodiment of Figure 4, duplicated resources include instruction pointer logic in the instruction fetch portion of the pipeline, register renaming logic in the rename portion of the pipeline, state variables (not shown, but referenced in various stages in the pipeline), and an interrupt controller (not shown, generally asynchronous to pipeline).
15 Shared resources in the embodiment of Figure 4 include schedulers in the schedule stage of the pipeline, a pool of registers in the register read and write portions of the pipeline, execution resources in the execute portion of the pipeline. Additionally, a trace cache and an LI data cache may be shared resources populated according to memory accesses without regard to thread context. In other embodiments, consideration of thread context
20 may be used in caching decisions. Partitioned resources in the embodiment of Figure 4, include two queues in queuing stages of the pipeline, a re-order buffer in a retirement stage of the pipeline, and a store buffer. Thread selection multiplexing logic alternates between the various duplicated and partitioned resources to provide reasonable access to both threads.
[0030] In the embodiment of Figure 4, when one thread is suspended, all instructions related to thread 1 are drained from both queues. Each pair of queues is then combined to provide a larger queue to the second thread. Similarly, more registers from the register pool are made available to the second thread, more entries from the store buffer are freed for the second thread, and more entries in the re-order buffer are made available to the second thread. In essence, these structures are returned to single dedicated structures of twice the size. Of course, different proportions may result from implementations using different numbers of threads. In some embodiments, the thread partitionable resources, the replicated resources, and the shared resources may be arranged differently. In some embodiments, there may not be partitionable resources on both ends of the shared resources. In some embodiments, the partitionable resources may not be strictly partitioned, but rather may allow some instructions to cross partitions or may allow partitions to vary in size depending on the thread being executed in that partition or the total number of threads being executed. Additionally, different mixes of resources may be designated as shared, duplicated, and partitioned resources.
Figure 5 illustrates various design representations or formats for simulation, emulation, and fabrication of a design using the disclosed techniques. Data representing a design may represent the design in a number of manners. First, as is useful in simulations, the hardware may be represented using a hardware description language or another functional description language which essentially provides a computerized model of how the designed hardware is expected to perform. The hardware model 1110 may be stored in a storage medium 1100 such as a computer memory so that the model may be simulated using simulation software 1120 that applies a particular test suite 1130 to the hardware model 1110 to determine if it indeed functions as intended. In some embodiments, the simulation software is not recorded, captured, or contained in the medium. Additionally, a circuit level model with logic and/or transistor gates may be produced at some stages of the design process. This model may be similarly simulated, sometimes by dedicated hardware simulators that form the model using programmable logic. This type of simulation, taken a degree further, may be an emulation technique. In any case, re-configurable hardware is another embodiment that may involve a machine readable medium storing a model employing the disclosed techniques.
Furthermore, most designs, at some stage, reach a level of data representing the physical placement of various devices in the hardware model. In the case where conventional semiconductor fabrication techniques are used, the data representing the hardware model may be the data specifying the presence or absence of various features on different mask layers for masks used to produce the integrated circuit. Again, this data representing the integrated circuit embodies the techniques disclosed in that the circuitry or logic in the data can be simulated or fabricated to perform these techniques.
In any representation of the design, the data may be stored in any form of a computer readable medium. An optical or electrical wave 1160 modulated or otherwise generated to transmit such information, a memory 1150, or a magnetic or optical storage 1140 such as a disc may be the medium. The set of bits describing the design or the particular part of the design are an article that may be sold in and of itself or used by others for further design or fabrication. Thus, techniques for suspending execution of a thread in a multi-threaded processor are disclosed. While certain exemplary embodiments have been described and shown in the accompanying drawings, it is to be understood that such embodiments are merely illustrative of and not restrictive on the broad invention, and that this invention not be limited to the specific constructions and arrangements shown and described, since various other modifications may occur to those ordinarily skilled in the art upon studying this disclosure.

Claims

What is claimed is:
1. A processor comprising: a plurality of thread partitionable resources that are each partitionable between a plurality of threads; logic to receive a program instruction from a first thread of said plurality of threads, and in response to said program instruction to cause the processor to suspend execution of the first thread and to relinquish portions of said plurality of thread partitionable resources associated with the first thread for use by other ones of said plurality of threads.
2. The processor of claim 1 wherein the program instruction is a suspend instruction.
3. The processor of claim 1 wherein said logic is to cause the processor to suspend the first thread for a selected amount of time.
4. The processor of claim 3 wherein said selected amount of time is a fixed amount of time.
5. The processor of claim 3 wherein said processor is to execute instructions from a second thread while said first thread is suspended.
6. The processor of claim 3 wherein said selected amount of time is programmable by at least one technique chosen from a set consisting of: providing an operand in conjunction with the program instruction; blowing fuses to set the selected amount; programming the selected amount in a storage location in advance of decoding the program instruction; setting the selected amount in microcode.
7. The processor of claim 1 wherein said plurality of thread partitionable resources comprises: an instruction queue; a register pool.
8. The processor of claim 7 further comprising: a plurality of shared resources, said plurality of shared resources comprising: a plurality of execution units; a cache; a scheduler; a plurality of duplicated resources, said plurality of duplicated resources comprising: a plurality of processor state variables; an instruction pointer; register renaming logic.
9. The processor of claim 8 wherein said plurality of thread partitionable resources further comprises: a plurality of re-order buffers; a plurality of store buffer entries.
10. The processor of claim 1 wherein said logic is further to cause the processor to resume execution of said first thread in response to an event.
11. The processor of claim 3 wherein said logic is further to cause the processor to ignore events until said selected amount of time has elapsed.
12. The processor of claim 1 wherein said processor is embodied in digital format on a computer readable medium.
13. A method comprising: receiving a first opcode in a first thread of execution; suspending said first thread for a selected amount of time in response to said first opcode; relinquishing a plurality of thread partitionable resources in response to said first opcode.
14. The method of claim 13 wherein relinquishing comprises: annealing the plurality of thread partitionable resources to become larger structures usable by fewer threads.
15. The method of claim 14 wherein relinquishing said plurality of thread partitionable resources comprises: relinquishing a partition of an instruction queue; relinquishing a plurality of registers from a register pool.
16. The method of claim 15 wherein relinquishing said plurality of thread partitionable resources further comprises: relinquishing a plurality of store buffer entries; relinquishing a plurality of re-order buffer entries.
17. The method of claim 13 wherein said selected amount of time is programmable by at least one technique chosen from a set consisting of: providing an operand in conjunction with the first opcode; blowing fuses to set the selected amount of time; programming the selected amount of time in a storage location in advance of decoding the program instruction; setting the selected amount of time in microcode.
18. A system comprising: a memory to store a plurality of program threads, including a first thread and a second thread, said first thread including a first instruction; a processor coupled to said memory, said processor including a plurality of thread partitionable resources and a plurality of shared resources, said processor to execute instructions from said memory, said processor, in response to execution of said first instruction to suspend said first thread and to relinquish portions of said plurality of thread partitionable resources.
19. The system of claim 18 wherein said processor is to execute said second thread from said memory while said first thread is suspended.
20. The system of claim 19 wherein said processor is to suspend execution of said first thread in response to said first instruction for a selected amount of time, said selected amount of time is chosen by at least one technique chosen from a set consisting of: providing an operand in conjunction with the program instruction; blowing fuses to set the selected amount of time; programming the selected amount of time in a storage location in advance of decoding the program instruction; setting the select amount of time in microcode.
21. The system of claim 18 wherein said plurality of thread partitionable resources comprises: an instruction queue; a register pool.
22. The system of claim 21 wherein said processor further comprises: a plurality of shared resources, said plurality of shared resources comprising: a plurality of execution units ; a cache; a scheduler; a plurality of duplicated resources, said plurality of duplicated resources comprising: a plurality of processor state variables; an instruction pointer; register renaming logic.
23. The system of claim 22 wherein said plurality of thread partitionable resources further comprises: a plurality of re-order buffers; a plurality of store buffer entries;
24. An apparatus comprising: means for receiving a first instruction from a first thread; means for suspending said first thread in response to said first instruction; means for relinquishing a plurality of partitions of a plurality of resources; means for re-partitioning said plurality of resources after a selected amount of time.
25. The apparatus of claim 24 wherein said first instruction is a macro-instruction from a user-executable program.
26. The apparatus of claim 25 wherein said plurality of resources comprises a register pool and an instruction queue.
PCT/US2002/039790 2001-12-31 2002-12-11 Suspending execution of a thread in a multi-threaded WO2003058434A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
JP2003558678A JP2005514698A (en) 2001-12-31 2002-12-11 Suspend processing of multi-thread processor threads
KR1020047010393A KR100617417B1 (en) 2001-12-31 2002-12-11 Suspending execution of a thread in a multi-threaeded processor
DE10297597T DE10297597T5 (en) 2001-12-31 2002-12-11 Suspending the execution of a thread in a multi-thread processor
AU2002364559A AU2002364559A1 (en) 2001-12-31 2002-12-11 Suspending execution of a thread in a multi-threaded
HK05107419A HK1075109A1 (en) 2001-12-31 2005-08-24 A processor and a method of suspending a thread

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/039,777 2001-12-31
US10/039,777 US20030126416A1 (en) 2001-12-31 2001-12-31 Suspending execution of a thread in a multi-threaded processor

Publications (1)

Publication Number Publication Date
WO2003058434A1 true WO2003058434A1 (en) 2003-07-17

Family

ID=21907295

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2002/039790 WO2003058434A1 (en) 2001-12-31 2002-12-11 Suspending execution of a thread in a multi-threaded

Country Status (9)

Country Link
US (1) US20030126416A1 (en)
JP (1) JP2005514698A (en)
KR (1) KR100617417B1 (en)
CN (1) CN1287272C (en)
AU (1) AU2002364559A1 (en)
DE (1) DE10297597T5 (en)
HK (1) HK1075109A1 (en)
TW (1) TW200403588A (en)
WO (1) WO2003058434A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1716482A2 (en) * 2004-02-04 2006-11-02 Intel Corporation Sharing idled processor execution resources
KR100745477B1 (en) 2003-04-24 2007-08-02 인터내셔널 비지네스 머신즈 코포레이션 Dynamic Switching of Multithreaded Processor between Single Threaded and Simultaneous Multithreded Modes
US8407714B2 (en) 2007-06-20 2013-03-26 Fujitsu Limited Arithmetic device for processing one or more threads
US10740126B2 (en) 2013-03-15 2020-08-11 Intel Corporation Methods, systems and apparatus for supporting wide and efficient front-end operation with guest-architecture emulation
US11163720B2 (en) 2006-04-12 2021-11-02 Intel Corporation Apparatus and method for processing an instruction matrix specifying parallel and dependent operations
US11204769B2 (en) 2011-03-25 2021-12-21 Intel Corporation Memory fragments for supporting code block execution by using virtual cores instantiated by partitionable engines
US11656875B2 (en) 2013-03-15 2023-05-23 Intel Corporation Method and system for instruction block to execution unit grouping

Families Citing this family (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7363474B2 (en) * 2001-12-31 2008-04-22 Intel Corporation Method and apparatus for suspending execution of a thread until a specified memory access occurs
US7127561B2 (en) * 2001-12-31 2006-10-24 Intel Corporation Coherency techniques for suspending execution of a thread until a specified memory access occurs
US7216346B2 (en) * 2002-12-31 2007-05-08 International Business Machines Corporation Method and apparatus for managing thread execution in a multithread application
US7213093B2 (en) * 2003-06-27 2007-05-01 Intel Corporation Queued locks using monitor-memory wait
US7836450B2 (en) 2003-08-28 2010-11-16 Mips Technologies, Inc. Symmetric multiprocessor operating system for execution on non-independent lightweight thread contexts
US7870553B2 (en) * 2003-08-28 2011-01-11 Mips Technologies, Inc. Symmetric multiprocessor operating system for execution on non-independent lightweight thread contexts
JP4818919B2 (en) 2003-08-28 2011-11-16 ミップス テクノロジーズ インコーポレイテッド Integrated mechanism for suspending and deallocating computational threads of execution within a processor
US7711931B2 (en) * 2003-08-28 2010-05-04 Mips Technologies, Inc. Synchronized storage providing multiple synchronization semantics
US7594089B2 (en) 2003-08-28 2009-09-22 Mips Technologies, Inc. Smart memory based synchronization controller for a multi-threaded multiprocessor SoC
US7376954B2 (en) 2003-08-28 2008-05-20 Mips Technologies, Inc. Mechanisms for assuring quality of service for programs executing on a multithreaded processor
US7849297B2 (en) 2003-08-28 2010-12-07 Mips Technologies, Inc. Software emulation of directed exceptions in a multithreading processor
US7418585B2 (en) * 2003-08-28 2008-08-26 Mips Technologies, Inc. Symmetric multiprocessor operating system for execution on non-independent lightweight thread contexts
US9032404B2 (en) 2003-08-28 2015-05-12 Mips Technologies, Inc. Preemptive multitasking employing software emulation of directed exceptions in a multithreading processor
US8533716B2 (en) 2004-03-31 2013-09-10 Synopsys, Inc. Resource management in a multicore architecture
US9189230B2 (en) 2004-03-31 2015-11-17 Intel Corporation Method and system to provide concurrent user-level, non-privileged shared resource thread creation and execution
GB0407384D0 (en) * 2004-03-31 2004-05-05 Ignios Ltd Resource management in a multicore processor
US8041930B2 (en) * 2005-05-11 2011-10-18 Arm Limited Data processing apparatus and method for controlling thread access of register sets when selectively operating in secure and non-secure domains
US8032737B2 (en) * 2006-08-14 2011-10-04 Marvell World Trade Ltd. Methods and apparatus for handling switching among threads within a multithread processor
CN101595456A (en) * 2006-12-27 2009-12-02 莫尔It资源有限公司 The method and system that is used for transaction resource control
US7975272B2 (en) * 2006-12-30 2011-07-05 Intel Corporation Thread queuing method and apparatus
US8725975B2 (en) * 2007-01-03 2014-05-13 Freescale Semiconductor, Inc. Progressive memory initialization with waitpoints
US20080162858A1 (en) * 2007-01-03 2008-07-03 Freescale Semiconductor, Inc. Hardware-based memory initialization with software support
US20080244242A1 (en) * 2007-04-02 2008-10-02 Abernathy Christopher M Using a Register File as Either a Rename Buffer or an Architected Register File
US7707390B2 (en) * 2007-04-25 2010-04-27 Arm Limited Instruction issue control within a multi-threaded in-order superscalar processor
JP5104861B2 (en) * 2007-06-19 2012-12-19 富士通株式会社 Arithmetic processing unit
US20090100249A1 (en) * 2007-10-10 2009-04-16 Eichenberger Alexandre E Method and apparatus for allocating architectural register resources among threads in a multi-threaded microprocessor core
US8131983B2 (en) * 2008-04-28 2012-03-06 International Business Machines Corporation Method, apparatus and article of manufacture for timeout waits on locks
US20120166777A1 (en) * 2010-12-22 2012-06-28 Advanced Micro Devices, Inc. Method and apparatus for switching threads
US8578394B2 (en) * 2011-09-09 2013-11-05 Microsoft Corporation Exempting applications from suspension
CN103389911B (en) * 2012-05-07 2016-08-03 启碁科技股份有限公司 Save the method for system resource and use the operating system of its method
US10140129B2 (en) 2012-12-28 2018-11-27 Intel Corporation Processing core having shared front end unit
US9361116B2 (en) 2012-12-28 2016-06-07 Intel Corporation Apparatus and method for low-latency invocation of accelerators
US9417873B2 (en) 2012-12-28 2016-08-16 Intel Corporation Apparatus and method for a hybrid latency-throughput processor
US10346195B2 (en) * 2012-12-29 2019-07-09 Intel Corporation Apparatus and method for invocation of a multi threaded accelerator
CN103345422B (en) * 2013-07-02 2019-01-29 厦门雅迅网络股份有限公司 A kind of multithreading hard real-time control method based on Linux
US10223026B2 (en) * 2013-09-30 2019-03-05 Vmware, Inc. Consistent and efficient mirroring of nonvolatile memory state in virtualized environments where dirty bit of page table entries in non-volatile memory are not cleared until pages in non-volatile memory are remotely mirrored
US10140212B2 (en) 2013-09-30 2018-11-27 Vmware, Inc. Consistent and efficient mirroring of nonvolatile memory state in virtualized environments by remote mirroring memory addresses of nonvolatile memory to which cached lines of the nonvolatile memory have been flushed
US9515901B2 (en) 2013-10-18 2016-12-06 AppDynamics, Inc. Automatic asynchronous handoff identification
US20160170767A1 (en) * 2014-12-12 2016-06-16 Intel Corporation Temporary transfer of a multithreaded ip core to single or reduced thread configuration during thread offload to co-processor
CN105843592A (en) * 2015-01-12 2016-08-10 芋头科技(杭州)有限公司 System for implementing script operation in preset embedded system
US10552224B2 (en) * 2015-05-14 2020-02-04 Hitachi, Ltd. Computer system including server storage system
US11023233B2 (en) * 2016-02-09 2021-06-01 Intel Corporation Methods, apparatus, and instructions for user level thread suspension
US10353817B2 (en) * 2017-03-07 2019-07-16 International Business Machines Corporation Cache miss thread balancing
CN107690625B (en) * 2017-03-16 2022-02-11 深圳大趋智能科技有限公司 iOS thread recovery method and device
TWI647619B (en) * 2017-08-29 2019-01-11 智微科技股份有限公司 Method for performing hardware resource management in an electronic device, and corresponding electronic device
CN109471673B (en) * 2017-09-07 2022-02-01 智微科技股份有限公司 Method for hardware resource management in electronic device and electronic device
US10481915B2 (en) * 2017-09-20 2019-11-19 International Business Machines Corporation Split store data queue design for an out-of-order processor
GB2569098B (en) * 2017-10-20 2020-01-08 Graphcore Ltd Combining states of multiple threads in a multi-threaded processor
GB201717303D0 (en) * 2017-10-20 2017-12-06 Graphcore Ltd Scheduling tasks in a multi-threaded processor

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0827071A2 (en) * 1996-08-27 1998-03-04 Matsushita Electric Industrial Co., Ltd. Multithreaded processor for processing multiple instruction streams independently of each other by flexibly controlling throughput in each instruction stream
WO1998043193A2 (en) * 1997-03-21 1998-10-01 University Of Maryland Spawn-join instruction set architecture for providing explicit multithreading

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5357617A (en) * 1991-11-22 1994-10-18 International Business Machines Corporation Method and apparatus for substantially concurrent multiple instruction thread processing by a single pipeline processor
JP3678759B2 (en) * 1992-07-21 2005-08-03 アドバンスト・マイクロ・ディバイシズ・インコーポレイテッド Apparatus for generating an interrupt and method for generating an interrupt
US5584031A (en) * 1993-11-09 1996-12-10 Motorola Inc. System and method for executing a low power delay instruction
JPH08320797A (en) * 1995-05-24 1996-12-03 Fuji Xerox Co Ltd Program control system
US5933627A (en) * 1996-07-01 1999-08-03 Sun Microsystems Thread switch on blocked load or store using instruction thread field
US5961639A (en) * 1996-12-16 1999-10-05 International Business Machines Corporation Processor and method for dynamically inserting auxiliary instructions within an instruction stream during execution
US6401155B1 (en) * 1998-12-22 2002-06-04 Philips Electronics North America Corporation Interrupt/software-controlled thread processing
SG65097A1 (en) * 1998-12-28 2001-08-21 Compaq Computer Corp Break event generation during transitions between modes of operation in a computer system
US6535905B1 (en) * 1999-04-29 2003-03-18 Intel Corporation Method and apparatus for thread switching within a multithreaded processor
US6341347B1 (en) * 1999-05-11 2002-01-22 Sun Microsystems, Inc. Thread switch logic in a multiple-thread processor
US6493741B1 (en) * 1999-10-01 2002-12-10 Compaq Information Technologies Group, L.P. Method and apparatus to quiesce a portion of a simultaneous multithreaded central processing unit
US6357016B1 (en) * 1999-12-09 2002-03-12 Intel Corporation Method and apparatus for disabling a clock signal within a multithreaded processor
US6496925B1 (en) * 1999-12-09 2002-12-17 Intel Corporation Method and apparatus for processing an event occurrence within a multithreaded processor
US6931639B1 (en) * 2000-08-24 2005-08-16 International Business Machines Corporation Method for implementing a variable-partitioned queue for simultaneous multithreaded processors
AU2002326378A1 (en) * 2001-07-13 2003-01-29 Sun Microsystems, Inc. Facilitating efficient join operations between a head thread and a speculative thread

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0827071A2 (en) * 1996-08-27 1998-03-04 Matsushita Electric Industrial Co., Ltd. Multithreaded processor for processing multiple instruction streams independently of each other by flexibly controlling throughput in each instruction stream
WO1998043193A2 (en) * 1997-03-21 1998-10-01 University Of Maryland Spawn-join instruction set architecture for providing explicit multithreading

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
MENDELSON A ET AL: "DESIGN ALTERNATIVES OF MULTITHREADED ARCHITECTURE", INTERNATIONAL JOURNAL OF PARALLEL PROGRAMMING, PLENUM PRESS, NEW YORK, US, vol. 27, no. 3, June 1999 (1999-06-01), pages 161 - 193, XP000849200, ISSN: 0885-7458 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100745477B1 (en) 2003-04-24 2007-08-02 인터내셔널 비지네스 머신즈 코포레이션 Dynamic Switching of Multithreaded Processor between Single Threaded and Simultaneous Multithreded Modes
EP1716482A2 (en) * 2004-02-04 2006-11-02 Intel Corporation Sharing idled processor execution resources
JP2007520825A (en) * 2004-02-04 2007-07-26 インテル・コーポレーション Sharing processor execution resources in standby state
JP2012104140A (en) * 2004-02-04 2012-05-31 Intel Corp Sharing processor execution resources in waiting state
EP3048527A1 (en) * 2004-02-04 2016-07-27 Intel Corporation Sharing idled processor execution resources
US11163720B2 (en) 2006-04-12 2021-11-02 Intel Corporation Apparatus and method for processing an instruction matrix specifying parallel and dependent operations
US8407714B2 (en) 2007-06-20 2013-03-26 Fujitsu Limited Arithmetic device for processing one or more threads
US11204769B2 (en) 2011-03-25 2021-12-21 Intel Corporation Memory fragments for supporting code block execution by using virtual cores instantiated by partitionable engines
US10740126B2 (en) 2013-03-15 2020-08-11 Intel Corporation Methods, systems and apparatus for supporting wide and efficient front-end operation with guest-architecture emulation
US11656875B2 (en) 2013-03-15 2023-05-23 Intel Corporation Method and system for instruction block to execution unit grouping

Also Published As

Publication number Publication date
KR20040069352A (en) 2004-08-05
AU2002364559A1 (en) 2003-07-24
US20030126416A1 (en) 2003-07-03
CN1608246A (en) 2005-04-20
TW200403588A (en) 2004-03-01
DE10297597T5 (en) 2005-01-05
HK1075109A1 (en) 2005-12-02
CN1287272C (en) 2006-11-29
JP2005514698A (en) 2005-05-19
KR100617417B1 (en) 2006-08-30

Similar Documents

Publication Publication Date Title
US20030126416A1 (en) Suspending execution of a thread in a multi-threaded processor
US7363474B2 (en) Method and apparatus for suspending execution of a thread until a specified memory access occurs
US7127561B2 (en) Coherency techniques for suspending execution of a thread until a specified memory access occurs
US7254697B2 (en) Method and apparatus for dynamic modification of microprocessor instruction group at dispatch
US7213093B2 (en) Queued locks using monitor-memory wait
US7020871B2 (en) Breakpoint method for parallel hardware threads in multithreaded processor
US6256775B1 (en) Facilities for detailed software performance analysis in a multithreaded processor
US20030126379A1 (en) Instruction sequences for suspending execution of a thread until a specified memory access occurs
US20050120194A1 (en) Apparatus, method, and instruction for initiation of concurrent instruction streams in a multithreading microprocessor
US7603543B2 (en) Method, apparatus and program product for enhancing performance of an in-order processor with long stalls
US8635621B2 (en) Method and apparatus to implement software to hardware thread priority
US10599428B2 (en) Relaxed execution of overlapping mixed-scalar-vector instructions
US20040215945A1 (en) Method for changing a thread priority in a simultaneous multithread processor
US20040210742A1 (en) Method and circuit for modifying pipeline length in a simultaneous multithread processor
WO2005022384A1 (en) Apparatus, method, and instruction for initiation of concurrent instruction streams in a multithreading microprocessor
WO2001040935A1 (en) Method and apparatus for constructing a pre-scheduled instruction cache
US7337304B2 (en) Processor for executing instruction control in accordance with dynamic pipeline scheduling and a method thereof
EP3433724B1 (en) Processing vector instructions
JP2001142702A (en) Mechanism for fast access to control space of pipeline processor
GB2441903A (en) Resuming control of resources by a processor on exiting a sleep mode and disabling an associated monitor.
JPH04308930A (en) Electronic computer

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SD SE SG SK SL TJ TM TN TR TT TZ UA UG UZ VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LU MC NL PT SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 20028261585

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 2003558678

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 1020047010393

Country of ref document: KR

RET De translation (de og part 6b)

Ref document number: 10297597

Country of ref document: DE

Date of ref document: 20050105

Kind code of ref document: P

WWE Wipo information: entry into national phase

Ref document number: 10297597

Country of ref document: DE

122 Ep: pct application non-entry in european phase
REG Reference to national code

Ref country code: DE

Ref legal event code: 8607