WO2002069238A2 - Managing coherence via put/get windows - Google Patents
Managing coherence via put/get windows Download PDFInfo
- Publication number
- WO2002069238A2 WO2002069238A2 PCT/US2002/005587 US0205587W WO02069238A2 WO 2002069238 A2 WO2002069238 A2 WO 2002069238A2 US 0205587 W US0205587 W US 0205587W WO 02069238 A2 WO02069238 A2 WO 02069238A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- cache
- processor
- data
- memory
- activities
- Prior art date
Links
- 238000012545 processing Methods 0.000 claims abstract description 26
- 238000000034 method Methods 0.000 claims abstract description 25
- 238000004891 communication Methods 0.000 claims description 26
- 230000000694 effects Effects 0.000 claims description 22
- 230000001427 coherent effect Effects 0.000 claims description 10
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000004140 cleaning Methods 0.000 claims description 4
- 230000004888 barrier function Effects 0.000 description 7
- 230000008901 benefit Effects 0.000 description 4
- 230000003139 buffering effect Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 238000013461 design Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000001360 synchronised effect Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013481 data capture Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 230000011664 signaling Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05K—PRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS
- H05K7/00—Constructional details common to different types of electric apparatus
- H05K7/20—Modifications to facilitate cooling, ventilating, or heating
- H05K7/20709—Modifications to facilitate cooling, ventilating, or heating for server racks or cabinets; for data centers, e.g. 19-inch computer racks
- H05K7/20836—Thermal management, e.g. server temperature control
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F12/00—Accessing, addressing or allocating within memory systems or architectures
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F04—POSITIVE - DISPLACEMENT MACHINES FOR LIQUIDS; PUMPS FOR LIQUIDS OR ELASTIC FLUIDS
- F04D—NON-POSITIVE-DISPLACEMENT PUMPS
- F04D25/00—Pumping installations or systems
- F04D25/16—Combinations of two or more pumps ; Producing two or more separate gas flows
- F04D25/166—Combinations of two or more pumps ; Producing two or more separate gas flows using fans
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F04—POSITIVE - DISPLACEMENT MACHINES FOR LIQUIDS; PUMPS FOR LIQUIDS OR ELASTIC FLUIDS
- F04D—NON-POSITIVE-DISPLACEMENT PUMPS
- F04D27/00—Control, e.g. regulation, of pumps, pumping installations or pumping systems specially adapted for elastic fluids
- F04D27/004—Control, e.g. regulation, of pumps, pumping installations or pumping systems specially adapted for elastic fluids by varying driving speed
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/16—Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
- G06F15/163—Interprocessor communication
- G06F15/167—Interprocessor communication using a common memory, e.g. mailbox
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/16—Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
- G06F15/163—Interprocessor communication
- G06F15/173—Interprocessor communication using an interconnection network, e.g. matrix, shuffle, pyramid, star, snowflake
- G06F15/17356—Indirect interconnection networks
- G06F15/17368—Indirect interconnection networks non hierarchical topologies
- G06F15/17381—Two dimensional, e.g. mesh, torus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/14—Fourier, Walsh or analogous domain transformations, e.g. Laplace, Hilbert, Karhunen-Loeve, transforms
- G06F17/141—Discrete Fourier transforms
- G06F17/142—Fast Fourier transforms, e.g. using a Cooley-Tukey type algorithm
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/52—Program synchronisation; Mutual exclusion, e.g. by means of semaphores
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/52—Program synchronisation; Mutual exclusion, e.g. by means of semaphores
- G06F9/526—Mutual exclusion algorithms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/54—Interprogram communication
- G06F9/544—Buffers; Shared memory; Pipes
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/003—Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
- G09G5/006—Details of the interface to the display terminal
- G09G5/008—Clock recovery
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L7/00—Arrangements for synchronising receiver with transmitter
- H04L7/02—Speed or phase control by the received code signals, the signals containing no special synchronisation information
- H04L7/033—Speed or phase control by the received code signals, the signals containing no special synchronisation information using the transitions of the received signal to control the phase of the synchronising-signal-generating means, e.g. using a phase-locked loop
- H04L7/0337—Selecting between two or more discretely delayed clocks or selecting between two or more discretely delayed received code signals
- H04L7/0338—Selecting between two or more discretely delayed clocks or selecting between two or more discretely delayed received code signals the correction of the phase error being performed by a feed forward loop
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F24—HEATING; RANGES; VENTILATING
- F24F—AIR-CONDITIONING; AIR-HUMIDIFICATION; VENTILATION; USE OF AIR CURRENTS FOR SCREENING
- F24F11/00—Control or safety arrangements
- F24F11/70—Control systems characterised by their outputs; Constructional details thereof
- F24F11/72—Control systems characterised by their outputs; Constructional details thereof for controlling the supply of treated air, e.g. its pressure
- F24F11/74—Control systems characterised by their outputs; Constructional details thereof for controlling the supply of treated air, e.g. its pressure for controlling air flow rate or air velocity
- F24F11/77—Control systems characterised by their outputs; Constructional details thereof for controlling the supply of treated air, e.g. its pressure for controlling air flow rate or air velocity by controlling the speed of ventilators
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02B—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO BUILDINGS, e.g. HOUSING, HOUSE APPLIANCES OR RELATED END-USER APPLICATIONS
- Y02B30/00—Energy efficient heating, ventilation or air conditioning [HVAC]
- Y02B30/70—Efficient control or regulation technologies, e.g. for control of refrigerant flow, motor or heating
Definitions
- This invention relates to the field of distributed-memory message-passing parallel computer design and system software, as applied for example to computation in the field of life sciences.
- Each processing node of the supercomputer architecture is a semiconductor device that includes two electronic processors (among other components).
- One of these processors is designated the “Compute Processor” and, in the common made operation, is dedicated to application computation.
- the other processor is the “I/O Processor,” which, in the common mode of operation, is a service processor dedicated to performing activities in support of message-passing communication.
- Each of these processors contains a separate first-level cache (LI) which may contain a copy of data stored in a common memory accessed by both processors. If one processor changes its LI copy of a memory location, and the other processor has a copy of the same location, the two copies become "coherent” if they are made to be the same.
- LI first-level cache
- Message passing is a commonly-known form of computer communication wherein processors explicitly copy data from their own memory to that of another node.
- the I/O Processor is principally used to facilitate message passing between the common memory of a node and the common memory of other nodes. Therefore, it both produces data (when a message is received) that is consumed by the Compute Processor, and consumes data (in order to send a message) that is produced by the Compute Processor.
- both processors it is very common for both processors to have a copy of the same memory location in their Lls. If the messages passed are small and many, then the problem is exacerbated. Thus, there is a clear need to find a way to make the Lls of each processor coherent, without extensive circuitry, and with minimal impact on performance.
- One-sided communication uses a
- put/get message-passing paradigm, where messages carry the source (for get) or the destination (for put) memory address.
- puts and gets are typically assembled in batches and issued together. This keeps the independently operating processors in rough synchronization, maximizing performance.
- the time during which puts and gets occur is termed the put/get window. This window extends both in time (when it occurs) and in memory (over the range of memory addresses carried by the put or get messages).
- Figure 2 shows a put/get window 30 having a number of distinct messages.
- Put/get windows extend the concept of coherence to processors on different processing nodes of the massively parallel supercomputer. Implementations of put/get windows must insure that all messages put to a window during the time it is open are received into the memory of the window before the window is closed. Similarly, a get on the memory of the window is only allowed during the time the window is open. Therefore, put/get windows are simply a mechanism for a node to synchronize with remote processors operating on its memory.
- the management of a put/get window is currently accomplished by either buffering the put/get messages or by using explicit synchronization messages. Buffering the messages consumes memory, which is always in limited supply. Explicit synchronization for each window suffers from the long latency of round-trip messages between all the nodes accessing the window. Therefore, on large-scale machines such as the one described in copending patent application no.
- An object of this invention is to provide an improved procedure for managing coherence in a parallel processing computer system.
- Another object of the present invention is to achieve coherency between the first- level caches of the processors of a multi-processor node without extensive circuitry and with minimal impact on the performance of each processor.
- a further object of the invention is to provide a method and apparatus, working in conjunction with software algorithms, to accomplish efficient high speed message- passing communications between processors or a direct memory access (DMA) device, which maintains coherence without significantly reducing performance.
- DMA direct memory access
- the invention provides a software algorithm that simplifies and significantly speeds the management of cache coherence in a message passing massively parallel supercomputer (such as the one described in copending patent application no. (attorney Docket
- the invention provides a mechanism to ensure that the data being "put" (sent) is not in the cache of either processor, and that the data being "gotten” (received) is also not in the cache of either processor.
- the invention reduces the total amount of work required to achieve coherence and allow that work to be amortized over a large number of individual messages.
- both processing elements within a node must perform this work, the invention enables this to happen concurrently. Further, when required, these activities can be coordinated over a large number of independent nodes in the massively parallel machine by employing the Global Barrier Network described in copending patent application no. (attorney Docket 15275).
- the invention provides a hardware apparatus that assists the above-described cache coherence software algorithm, and limits the total time (or latency) required to achieve cache coherence over the Put/Get Window.
- This apparatus is a simple extension to the hardware address decoder that creates, in the physical memory address space of the node, an area of memory that (a) does not actually exist, and (b) is therefore able to respond instantly to read and write requests from the processing elements.
- This further speeds the coherence activities because it allows garbage data (which the processor will never use) to be pulled into the processor's cache, thereby evicting just the modified data and displacing unmodified data with optimal performance. The performance is faster because this garbage data does not actually need to be fetched from memory, rather, the memory controller need only instantly reply.
- data cache block flush and invalidate may be used to write data from the memory area of the first processor into the shared memory area, while at the same time, preventing the first processor from using data the data written in its memory area.
- data cache block zero may be used to write data from the memory area of the first processor into the shared memory.
- Figure 1 shows a two processor node embodying this invention.
- Figure 2 illustrates a put/get window that may be used in the practice of this invention.
- the present invention relates to a method and apparatus for managing coherence of a multi-processor computer system.
- Figure 1 illustrates a node 10 that may embody this invention.
- Each of the processors 12, 14 of node 10 has a respective cache memory area 16, 20, and the two processors share a third memory area 22.
- the present invention relates to a software algorithm that simplifies and significantly speeds the management of cache memory coherence in a message passing parallel computer, and to hardware apparatus that assists this cache coherence algorithm.
- the software algorithm uses the opening and closing of put/get windows to coordinate the activities required to achieve cache coherence.
- the hardware apparatus may be an extension to the hardware address decode, that creates, in the physical memory address space of the node, an area of physical memory that (a) does not actually exist, and (b) is therefore able to respond instantly to read and write requests from the processing elements.
- this invention utilizes a principal referred to as "put/get" data transfer.
- put/get data transfer
- typical application messaging traffic involves an increasing number of messages, where each such message contains a piece of work performed by other nodes in the multi -computer.
- one node scatters locally produced work items to numerous other nodes (a "put"), while assembling numerous remotely produced work items into its local memory (a "get”).
- put locally produced work items
- get assembling numerous remotely produced work items into its local memory
- One-Sided Communication uses a "put/get" message-passing paradigm, where messages carry the source (for "get”) or destination (for "put") memory address.
- put/get window This time during which puts and gets occur is termed the put/get window. This window extends both in time (when it occurs) and in memory (over which range of memory addresses does the data in the put or get reside).
- FIG. 2 shows a put/get window 30 having a number of distinct messages.
- algorithms often proceed as a series of steps, where each step consists of a computation phase followed by a communication phase.
- the nodes exchange data produced by the computation phase and required for the next step of the algorithm.
- the phases are usually tightly synchronized, sot that the communication happens all at once over the entire machine. Therefore, the cost of managing the synchronization of put/get windows can be amortized over a large number of nodes at the start and end of each communication phase.
- a global operation can be used to open many put/get windows at the start of a communication.
- the present invention utilizes this put/get window to provide a simple means to manage memory coherence.
- a software algorithm is provided that simplifies and significantly speeds the management of cache coherence in a message passing massively parallel supercomputer (such as the one described in copending patent application no. (attorney Docket
- this algorithm uses the opening and closing of "Put/Get Windows" to coordinate the activities required to achieve memory coherence.
- this invention provides a mechanism to ensure that the data being "put" (sent) is not in the cache of either processor, and that the data being "gotten” (received) is also not in the cache of either processor.
- this invention reduces the total amount of work required to achieve coherence and allow that work to be amortized over a large number of individual messages. Also, since both processing elements within a node must perform this work, this invention enables this to happen concurrently. Further, when required, these activities can be coordinated over a large number of independent nodes in the massively parallel machine by employing the Global Barrier Network described in copending patent application no. (attorney Docket 1527).
- a novel hardware apparatus that assists the above-described cache coherence algorithm, and limits the total time (or latency) required to achieve cache coherence over the Put/Get Window.
- This apparatus is a simple extension to the hardware address decoder that creates, in the physical memory address space of the node, an area of virtual memory that (a) does not actually exist, and (b) is therefore able to respond instantly to read and write requests from the processing elements.
- This further speeds the coherence activities because it allows garbage data (which the processor will never use) to be pulled into the processor's cache, thereby evicting just the modified data and displacing unmodified data with optimal performance. The performance is faster because this garbage data does not actually need to be fetched from memory, rather, the memory controller need only instantly reply.
- the first embodiment may be preferred if the size of the message being received is smaller than the size of LI, while the second embodiment may be preferred if the size of the message received is larger than LI .
- DCBF data cache block flush and invalidate
- the software then allows the computer processor to continue to process instructions, and closes the put/get window using a global synchronization operation such as the global barrier described in copending application copending application D#15272 Global Interrupt and Barrier Networks.
- Second embodiment If the size of the message received is larger than the size of LI.
- the invention makes use of an instruction termed "data cache block zero" (DCBZ), on a reserved contiguous physical address range equal in size to LI .
- DCBZ creates a new cache line with contents of zero. If a new cache line is not available, then another cache line in LI (for example, the least recently used line), has its data written back to the common memory, and is then zero'ed with the address given by the DCBZ instruction.
- DCBZ is a PowerPC BookE instruction; similar instructions exist for other processors.
- the software executes DCBZ to each line of the reserved area consecutively, where a line of the reserved area is equal in size to a cache line and like-aligned. This causes all lines in the LI to be flushed, i.e., all modified lines are written back to common memory, because the entire contents of LI is replaced quickly from the reserved area.
- the software then allows the compute processor to continue to process instructions, and closes the put/get window using a global synchronization operation such as the global barrier described in copending application copending application D#l 5272 Global Interrupt and Barrier Networks.
- DCBF instructions are slower than DCBZ, then the operating system may use the DCBZ instruction for messages smaller then LI and vice-versa.
- the I/O Processor need not flush its cache at all if the communication memory space is marked write-through to its LI cache.
- the making of the above-mentioned global "and" in a short interval of time, which allows the put/get window to be made temporarily narrow, is discussed in detail in related patent application no. (Attorney Docket: 15258 ).
Abstract
Description
Claims
Priority Applications (13)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AT02713681T ATE437402T1 (en) | 2001-02-24 | 2002-02-25 | COHERENCE MANAGEMENT VIA PUT/GET WINDOWS |
KR1020037010817A KR100615411B1 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
DE60233055T DE60233055D1 (en) | 2001-02-24 | 2002-02-25 | COHERENCE MANAGEMENT VIA PUT / GET WINDOW |
CNB028054318A CN1286016C (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
PCT/US2002/005587 WO2002069238A2 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
AU2002245518A AU2002245518A1 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
IL15751702A IL157517A0 (en) | 2001-02-24 | 2002-02-25 | Managing colerence via put/get windows |
US10/468,995 US7870343B2 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
CA002437663A CA2437663A1 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
JP2002568284A JP3954969B2 (en) | 2001-02-24 | 2002-02-25 | Coherence management via put / get windows |
EP02713681A EP1410216B1 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
US12/543,890 US8122197B2 (en) | 2001-02-24 | 2009-08-19 | Managing coherence via put/get windows |
US12/953,770 US8161248B2 (en) | 2001-02-24 | 2010-11-24 | Simplifying and speeding the management of intra-node cache coherence |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US27112401P | 2001-02-24 | 2001-02-24 | |
US60/271,124 | 2001-02-24 | ||
PCT/US2002/005587 WO2002069238A2 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
Related Child Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/468,995 A-371-Of-International US7870343B2 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
US12/543,890 Division US8122197B2 (en) | 2001-02-24 | 2009-08-19 | Managing coherence via put/get windows |
US12/953,770 Continuation US8161248B2 (en) | 2001-02-24 | 2010-11-24 | Simplifying and speeding the management of intra-node cache coherence |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2002069238A2 true WO2002069238A2 (en) | 2002-09-06 |
WO2002069238A3 WO2002069238A3 (en) | 2003-11-27 |
Family
ID=68463256
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2002/005587 WO2002069238A2 (en) | 2001-02-24 | 2002-02-25 | Managing coherence via put/get windows |
Country Status (11)
Country | Link |
---|---|
US (3) | US7870343B2 (en) |
EP (1) | EP1410216B1 (en) |
JP (1) | JP3954969B2 (en) |
KR (1) | KR100615411B1 (en) |
CN (1) | CN1286016C (en) |
AT (1) | ATE437402T1 (en) |
AU (1) | AU2002245518A1 (en) |
CA (1) | CA2437663A1 (en) |
DE (1) | DE60233055D1 (en) |
IL (1) | IL157517A0 (en) |
WO (1) | WO2002069238A2 (en) |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7804504B1 (en) | 2004-12-13 | 2010-09-28 | Massachusetts Institute Of Technology | Managing yield for a parallel processing integrated circuit |
US7882307B1 (en) | 2006-04-14 | 2011-02-01 | Tilera Corporation | Managing cache memory in a parallel processing environment |
US7853752B1 (en) * | 2006-09-29 | 2010-12-14 | Tilera Corporation | Caching in multicore and multiprocessor architectures |
JP2009104300A (en) * | 2007-10-22 | 2009-05-14 | Denso Corp | Data processing apparatus and program |
US9164554B2 (en) | 2010-04-12 | 2015-10-20 | Sandisk Enterprise Ip Llc | Non-volatile solid-state storage system supporting high bandwidth and random access |
US8725951B2 (en) * | 2010-04-12 | 2014-05-13 | Sandisk Enterprise Ip Llc | Efficient flash memory-based object store |
US8954385B2 (en) | 2010-06-28 | 2015-02-10 | Sandisk Enterprise Ip Llc | Efficient recovery of transactional data stores |
US8874515B2 (en) | 2011-04-11 | 2014-10-28 | Sandisk Enterprise Ip Llc | Low level object version tracking using non-volatile memory write generations |
US9135064B2 (en) | 2012-03-07 | 2015-09-15 | Sandisk Enterprise Ip Llc | Fine grained adaptive throttling of background processes |
US10152450B2 (en) * | 2012-08-09 | 2018-12-11 | International Business Machines Corporation | Remote processing and memory utilization |
US9037669B2 (en) | 2012-08-09 | 2015-05-19 | International Business Machines Corporation | Remote processing and memory utilization |
FR3048795A1 (en) * | 2016-03-11 | 2017-09-15 | Commissariat Energie Atomique | ON-CHIP SYSTEM AND METHOD OF EXCHANGING DATA BETWEEN NODES OF CALCULATIONS OF SUCH SYSTEM ON CHIP |
US10691609B2 (en) | 2017-07-24 | 2020-06-23 | International Business Machines Corporation | Concurrent data erasure and replacement of processors |
US10116436B1 (en) * | 2017-09-26 | 2018-10-30 | Intel Corporation | Techniques for preventing memory timing attacks |
CN110567112B (en) * | 2019-09-16 | 2021-07-13 | 重庆美的通用制冷设备有限公司 | Control method of air conditioning equipment, control system of air conditioning equipment and air conditioning system |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5864738A (en) * | 1996-03-13 | 1999-01-26 | Cray Research, Inc. | Massively parallel processing system using two data paths: one connecting router circuit to the interconnect network and the other connecting router circuit to I/O controller |
US6148146A (en) * | 1998-01-07 | 2000-11-14 | Poore; Bobby L. | Water heater |
US6223269B1 (en) * | 1997-09-27 | 2001-04-24 | Emc Corporation | Stacked mapped storage system |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5379391A (en) * | 1991-03-01 | 1995-01-03 | Storage Technology Corporation | Method and apparatus to access data records in a cache memory by multiple virtual addresses |
JP2819982B2 (en) * | 1993-03-18 | 1998-11-05 | 株式会社日立製作所 | Multiprocessor system with cache match guarantee function that can specify range |
US5537588A (en) * | 1994-05-11 | 1996-07-16 | International Business Machines Corporation | Partitioned log-structured file system and methods for operating the same |
US5860110A (en) * | 1995-08-22 | 1999-01-12 | Canon Kabushiki Kaisha | Conference maintenance method for cache memories in multi-processor system triggered by a predetermined synchronization point and a predetermined condition |
US5841973A (en) * | 1996-03-13 | 1998-11-24 | Cray Research, Inc. | Messaging in distributed memory multiprocessing system having shell circuitry for atomic control of message storage queue's tail pointer structure in local memory |
US5958019A (en) * | 1996-07-01 | 1999-09-28 | Sun Microsystems, Inc. | Multiprocessing system configured to perform synchronization operations |
TW379298B (en) * | 1996-09-30 | 2000-01-11 | Toshiba Corp | Memory updating history saving device and memory updating history saving method |
US6085295A (en) * | 1997-10-20 | 2000-07-04 | International Business Machines Corporation | Method of maintaining data coherency in a computer system having a plurality of interconnected nodes |
US6032228A (en) * | 1997-11-26 | 2000-02-29 | International Business Machines Corporation | Flexible cache-coherency mechanism |
US6490625B1 (en) * | 1997-11-26 | 2002-12-03 | International Business Machines Corporation | Powerful and flexible server architecture |
US6119205A (en) * | 1997-12-22 | 2000-09-12 | Sun Microsystems, Inc. | Speculative cache line write backs to avoid hotspots |
GB9727452D0 (en) * | 1997-12-31 | 1998-02-25 | Northern Telecom Ltd | Method and apparatus for replicating operations on data |
US6631448B2 (en) * | 1998-03-12 | 2003-10-07 | Fujitsu Limited | Cache coherence unit for interconnecting multiprocessor nodes having pipelined snoopy protocol |
JP3959914B2 (en) * | 1999-12-24 | 2007-08-15 | 株式会社日立製作所 | Main memory shared parallel computer and node controller used therefor |
US6772298B2 (en) * | 2000-12-20 | 2004-08-03 | Intel Corporation | Method and apparatus for invalidating a cache line without data return in a multi-node architecture |
JP2003288218A (en) * | 2002-03-27 | 2003-10-10 | Toshiba Corp | Microcomputer and its control method |
-
2002
- 2002-02-25 DE DE60233055T patent/DE60233055D1/en not_active Expired - Lifetime
- 2002-02-25 AT AT02713681T patent/ATE437402T1/en not_active IP Right Cessation
- 2002-02-25 US US10/468,995 patent/US7870343B2/en not_active Expired - Fee Related
- 2002-02-25 AU AU2002245518A patent/AU2002245518A1/en not_active Abandoned
- 2002-02-25 IL IL15751702A patent/IL157517A0/en unknown
- 2002-02-25 EP EP02713681A patent/EP1410216B1/en not_active Expired - Lifetime
- 2002-02-25 CN CNB028054318A patent/CN1286016C/en not_active Expired - Fee Related
- 2002-02-25 KR KR1020037010817A patent/KR100615411B1/en not_active IP Right Cessation
- 2002-02-25 WO PCT/US2002/005587 patent/WO2002069238A2/en active Application Filing
- 2002-02-25 JP JP2002568284A patent/JP3954969B2/en not_active Expired - Fee Related
- 2002-02-25 CA CA002437663A patent/CA2437663A1/en not_active Abandoned
-
2009
- 2009-08-19 US US12/543,890 patent/US8122197B2/en not_active Expired - Fee Related
-
2010
- 2010-11-24 US US12/953,770 patent/US8161248B2/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5864738A (en) * | 1996-03-13 | 1999-01-26 | Cray Research, Inc. | Massively parallel processing system using two data paths: one connecting router circuit to the interconnect network and the other connecting router circuit to I/O controller |
US6223269B1 (en) * | 1997-09-27 | 2001-04-24 | Emc Corporation | Stacked mapped storage system |
US6148146A (en) * | 1998-01-07 | 2000-11-14 | Poore; Bobby L. | Water heater |
Non-Patent Citations (4)
Title |
---|
'Re: non-blocking ops vs. threads', [Online] 11 July 1995, XP002950965 Retrieved from the Internet: <URL:http://www.mpi-forum.org/archives/mail/mpi-core/0057> * |
'Re: non-blocking ops vs. threads', [Online] 11 July 1995, XP002950966 Retrieved from the Internet: <URL:http://www.mpi-forum.org/archives/mail/mpi-core/0056> * |
'Re: non-blocking ops vs. threads', [Online] 12 July 1995, XP002950964 Retrieved from the Internet: <URL:http://www.mpi-forum.org/archives/mail/mpi-core/0063> * |
See also references of EP1410216A2 * |
Also Published As
Publication number | Publication date |
---|---|
IL157517A0 (en) | 2004-03-28 |
US8122197B2 (en) | 2012-02-21 |
EP1410216B1 (en) | 2009-07-22 |
CN1286016C (en) | 2006-11-22 |
US20090313439A1 (en) | 2009-12-17 |
US20110072219A1 (en) | 2011-03-24 |
KR20040004529A (en) | 2004-01-13 |
US20070055825A1 (en) | 2007-03-08 |
WO2002069238A3 (en) | 2003-11-27 |
ATE437402T1 (en) | 2009-08-15 |
EP1410216A2 (en) | 2004-04-21 |
AU2002245518A1 (en) | 2002-09-12 |
CA2437663A1 (en) | 2002-09-06 |
JP2005507097A (en) | 2005-03-10 |
KR100615411B1 (en) | 2006-08-25 |
EP1410216A4 (en) | 2006-03-22 |
JP3954969B2 (en) | 2007-08-08 |
US8161248B2 (en) | 2012-04-17 |
CN1493027A (en) | 2004-04-28 |
US7870343B2 (en) | 2011-01-11 |
DE60233055D1 (en) | 2009-09-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8122197B2 (en) | Managing coherence via put/get windows | |
Iftode et al. | Improving release-consistent shared virtual memory using automatic update | |
US7174434B2 (en) | Low latency memory access and synchronization | |
US6286090B1 (en) | Mechanism for selectively imposing interference order between page-table fetches and corresponding data fetches | |
US7818514B2 (en) | Low latency memory access and synchronization | |
JPH10133917A (en) | Multiprocess system with coherency-relative error logging capability | |
JPH10187645A (en) | Multiprocess system constituted for storage in many subnodes of process node in coherence state | |
JPH10116253A (en) | Multiprocess system executing synchronous operation | |
JPH10143482A (en) | Multiprocessor system for executing efficient write operation | |
JPH10187470A (en) | Multiprocess system provided with device for optimizing spin lock operation | |
JPH10149342A (en) | Multiprocess system executing prefetch operation | |
US20090006769A1 (en) | Programmable partitioning for high-performance coherence domains in a multiprocessor system | |
EP1408416A2 (en) | Extended symmetrical multiprocessor architecture | |
Andrews et al. | Notification and multicast networks for synchronization and coherence | |
Terasawa et al. | A cache coherency protocol for multiprocessor chip | |
WO2002069152A1 (en) | Managing coherence via put/get windows | |
Blumrich et al. | Managing coherence via put/get windows | |
Simoni | Implementing a directory-based cache consistency protocol | |
Blumrich et al. | Simplifying and speeding the management of intra-node cache coherence | |
Oi et al. | A cache coherence protocol for the bidirectional ring based multiprocessor | |
Rhee et al. | A scalable cache coherent scheme exploiting wormhole routing networks | |
Johnson | A multiprocessor architecture for high-rate communication processing | |
Shing et al. | A Novel Approach to the Design of Scalable Shared-Memory Multiprocessors | |
Tanaka et al. | Lightweight hardware distributed shared memory supported by generalized combining |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2437663 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020037010817 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 157517 Country of ref document: IL Ref document number: 2002568284 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 028054318 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2002713681 Country of ref document: EP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
WWP | Wipo information: published in national office |
Ref document number: 1020037010817 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 2002713681 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007055825 Country of ref document: US Ref document number: 10468995 Country of ref document: US |
|
WWP | Wipo information: published in national office |
Ref document number: 10468995 Country of ref document: US |