|Publication number||US20040213220 A1|
|Application number||US 09/749,383|
|Publication date||Oct 28, 2004|
|Filing date||Dec 28, 2000|
|Priority date||Dec 28, 2000|
|Also published as||US7724748, US7983275, US8542689, US20060203846, US20100226375, US20110268117, US20140079065|
|Publication number||09749383, 749383, US 2004/0213220 A1, US 2004/213220 A1, US 20040213220 A1, US 20040213220A1, US 2004213220 A1, US 2004213220A1, US-A1-20040213220, US-A1-2004213220, US2004/0213220A1, US2004/213220A1, US20040213220 A1, US20040213220A1, US2004213220 A1, US2004213220A1|
|Original Assignee||Davis Arlin R.|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (6), Referenced by (59), Classifications (15), Legal Events (1)|
|External Links: USPTO, USPTO Assignment, Espacenet|
 1. Field
 This invention relates to local area networks (LANs), and more specifically to emulation of connectionless LANs over connection-oriented fabrics.
 2. Background
 A number of networks are moving towards a connection-oriented arrangement. An example of a connection-oriented technology is Asynchronous Transfer Mode (ATM). Another example of a proposed technology that includes a connection-oriented (or channel based) capability is known as the Infiniband Architecture (IBA), described in the Infiniband Architecture Specification vol. 1, release 0.9, Mar. 31, 2000, authored by the Infiniband Trade Association. While connection-oriented technologies offer many advantages, in many instances it is desirable to maintain an interoperability between an existing connectionless technology and the connection-oriented technology. It is also desirable to maintain such interoperability, for example, when transitioning from a connectionless technology to a connection-oriented technology (or network) to allow existing software and components (e.g., legacy software) to be used. The Institute for Electrical and Electronics Engineers (IEEE) 802.3 Ethernet local area network (LAN) standard is an example of a common connectionless technology for a network.
 Current approaches to provide LAN emulation over a connection-oriented network (such as ATM) have a number of disadvantages. One example is ATM LAN emulation, which is in a specification provided by the ATM forum for the coexistence of legacy LANs and ATM LANs, ATM forum, “LAN emulation over ATM specification” version 1.0, 1995. The ATM LAN emulation specification is discussed in William Stallings, “Data and Computer Communications,” pages 487-495, fifth edition, 1997.
 As described in Stallings, the ATM LAN emulation specification proposes the use of a centralized LAN emulation service (LES) to perform basic LAN emulation services for nodes in a network, including: to set up connections and to map Media Access Control (MAC) addresses to ATM addresses. The LES also includes a broadcast and unknown server (BUS) service to provide broadcast/multicast of a packet to a plurality of nodes upon request from a client, and to provide a specialized protocol to allow nodes to learn ATM addresses of other nodes (i.e., by sending a LE_ARP_request message).
 Currently, there are no existing 802.3 LAN emulation mechanisms in place for Infiniband fabrics. Moreover, there are a number of disadvantages of systems such as the ATM LAN emulation mentioned previously. First, by using a centralized LES service, the network is prone to a single point of failure. Furthermore, the ATM LAN emulation described above, requires a separate and specialized address resolution protocol (ARP) (which is not compatible with the legacy or existing LAN networks) in order to attain the ATM address of a node corresponding to the node's MAC or LAN address. Moreover, calls through the operating system kernel requiring multiple buffer copies of data is typically required in many such existing computer systems, which can burden a processor with substantial overhead.
 The specialized name service and address resolution protocol maps Internet Protocol (IP) addresses to the mediums connection semantics. This method requires client software on each node and a centralized LAN emulation (LANE) server node that processes the ARPs, broadcast frames and multicast frames. Current LAN emulation architectures that map connection-oriented networks to 802.3 Ethernet generally map the connections to IP network addresses. This restricts the protocol to Transmission Control Protocol/Internet Protocol (TCP/IP) only. Further, in current systems, broadcasting in software over connection-oriented networks typically requires a buffer copy for each channel to send to all remote connected nodes. In addition, multicast traffic is not typically supported over existing connection-oriented networks.
 Therefore, there is a need for an 802.3 LAN emulation mechanism for Infiniband fabrics that solves the above noted problems of current systems.
 The present invention is further described in the detailed description which follows in reference to the noted plurality of drawings by way of non-limiting examples of embodiments of the present invention in which like reference numerals represent similar parts throughout the several views of the drawings and wherein:
FIG. 1 is a diagram of an example system for LAN emulation according to an example embodiment of the present invention;
FIG. 2 is a diagram of an example format of an 802.3 MAC address with embedded Infiniband LID according to an example embodiment of the present invention;
FIG. 3 is a table of an example mapping in an address resolution protocol according to an example embodiment of the present invention;
FIG. 4 is a diagram of an example software stack that resides in an IBLAN emulating node according to an example embodiment of the present invention;
FIG. 5 is a block diagram of an example initialization sequence of an IBLAN driver according to an example embodiment of the present invention; and
FIG. 6 is a system diagram of an example bridge node between an Infiniband fabric and Ethernet network according to an example embodiment of the present invention.
 The particulars shown herein are by way of example and for purposes of illustrative discussion of the embodiments of the present invention. The description taken with the drawings make it apparent to those skilled in the art how the present invention may be embodied in practice.
 Further, arrangements may be shown in block diagram form in order to avoid obscuring the invention, and also in view of the fact that specifics with respect to implementation of such block diagram arrangements is highly dependent upon the platform within which the present invention is to be implemented, i.e., specifics should be well within purview of one skilled in the art. Where specific details (e.g., circuits, flowcharts) are set forth in order to describe example embodiments of the invention, it should be apparent to one skilled in the art that the invention can be practiced without these specific details. Finally, it should be apparent that any combination of hard-wired circuitry and software instructions can be used to implement embodiments of the present invention, i.e., the present invention is not limited to any specific combination of hardware circuitry and software instructions.
 Although example embodiments of the present invention may be described using an example system block diagram in an example host unit environment, practice of the invention is not limited thereto, i.e., the invention may be able to be practiced with other types of systems, and in other types of environments (e.g., servers).
 Reference in the specification to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the invention. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
 The present invention relates to methods and devices for LAN emulation over Infiniband (IB) fabrics. According to the present invention, Infiniband connection-oriented fabrics may be presented to a protocol stack's networking layer as in 802.3 Ethernet network. Therefore, a connectionless LAN (802.3 Ethernet) is emulated over a connection-oriented fabric (Infiniband fabric). The present invention uses a name service to identify all Infiniband LAN emulation (IBLAN) nodes on the fabric. The present invention includes a software service that allows broadcast and multicast frames to be distributed to all nodes. According to the present invention, an Infiniband LID address is embedded in a standard Ethernet MAC header. This allows legacy network support on a local IBA subnet by tunneling standard 802.3 Ethernet frames across the subnet using Infiniband Architecture transport services.
 Infiniband Architectures provide many transport mechanisms (e.g., reliable and unreliable connections, reliable and unreliable datagrams, raw datagrams, and multicast services), for transferring data. In devices and methods according to the present invention, interoperability with all Infiniband Architectures is assured by providing mechanisms that consider the least common denominator of all Infiniband Architecture features. This includes, at a minimum 256 byte packet size, unreliable datagram, unreliable connection, and reliable connection. However, packet sizes of 512, 1024, 2048, and 4096, as well as multicasting and reliable and raw datagram service may also be incorporated according to the present invention. LAN emulation according to the present invention includes broadcasting and multicasting, Ethernet to Infiniband Architecture address mapping, and Infiniband host node discovery.
 Address mapping may be achieved by using the 16 bit base local identifier (LID) assigned to each port on each node of an Infiniband fabric. The host node may use this base LID address as the basis for its 48 bit Ethernet MAC address and treat each port as a separate network interface card (NIC). This address may be used by protocol drivers to update their local address resolution protocol (ARP) table and may be used as the reply to standard ARP requests. Node discovery may be accomplished by using the Subnet Management Administration Interface to query for a complete list of nodes on the fabric. The result of the query, a host node list, may be used an IBLAN node to simulate Ethernet and direct all broadcast and multicast frames. Unicast frames may be directed to specific IBLAN nodes using the embedded LID.
FIG. 1 shows a diagram of an example system for LAN emulation according to an example embodiment of the present invention. Infiniband Architecture fabric 10 has a number of hosts or nodes attached to it. These include nodes 12-18. Nodes 12-18 may all include an IBLAN driver, therefore, allowing the transfer of Ethernet messages amoung nodes 12-18. Although only four nodes are shown in this system diagram, there may be many more nodes that exist on the fabric and still be within the spirit and scope of the present invention. Further, one or more of nodes 12-18 connected to the Infiniband Architecture fabric 10 may be a subnet manager node. The subnet manager manages the subnet and performs initialization processes whereby the subnet manager identifies all nodes on fabric 10. The subnet manager assigns a local identifier (LID) to each port of a host or node and activates the port. A node may have one or more ports, each with a unique LID. The subnet manager stores this fabric topology information whereby it may be accessed by other nodes on the fabric.
 Moreover, one or more of nodes 12-18 that reside on IBA fabric 10 may be a bridge to another subnet or a different network all together. For example, node 16 may not only connect to Infiniband fabric 10, but may also have a port that is connected to a standard Ethernet network. In this situation, node 16 serves as a bridge between Infiniband fabric 10 and an Ethernet network.
FIG. 2 shows a diagram of an example format of an 802.3 MAC address with embedded Infiniband LID according to an example embodiment of the present invention. The 48 bit address includes a base Infiniband LID address of 16 bits, a reserved portion that includes 8 bits, and a vendor ID portion of 24 bits. The base Infiniband LID address is an address associated with a port of a node connected to an Infiniband fabric. The 8 bit reserve section may be used as needed by a particular application or function. The vendor ID is for plug and play and indicates a manufacturer, specific model, and/or version of a device. The vendor ID helps plug and play configure the node with appropriate drivers to run a particular device of the manufacturer.
FIG. 3 shows a table of an example mapping in an address resolution protocol according to an example embodiment of the present invention. An address resolution protocol (ARP) that resides at each node on the Infiniband fabric 10 that includes an IBLAN driver, maps network layer addresses, e.g., IP addresses, to Ethernet 48 bit MAC physical addresses. This mapping may be stored as an address resolution protocol table and is updated based on changes to nodes on the Infiniband fabric. The Ethernet address on the right side of the table shown in FIG. 3 corresponds to the format of the address shown in FIG. 2.
 To illustrate, node 12 on Infiniband fabric 10 may desire to send data to node 14 on Infiniband fabric 10. An application or device at node 12 may generate a network layer address based on a network protocol used at node 12. The address resolution protocol maps the network layer address to a physical Ethernet address. Initially, a broadcast Ethernet address is sent across the Infiniband fabric to all nodes, e.g., 14, 16, 18, etc., that reside on Infiniband fabric 10 and include an IBLAN driver. The Ethernet broadcast address may contain all ones in the 48 bit destination address, whereas the 48 bit source address contains the LID of each node on the Infiniband fabric. All nodes receive the broadcast message and whichever node has the network layer address may respond by sending a unicast message containing the LID of the destination node back to node 12. Node 12 uses this LID and directs a unicast message to the destination node using a known channel. All nodes on the Infiniband fabric, i.e., all NICs, are capable of receiving a destination address of all ones (e.g., broadcast message), a destination address with the most significant bit set to “1” but the rest not all ones (e.g., multicast message), or their unique Infiniband LID address (e.g., unicast message). The network protocol header, e.g., IP header, that resides after the Ethernet header may be used by upper level software at a destination node to determine if this broadcast message is for this particular node. If the message is not for this particular node, the multicast or broadcast message may simply be discarded.
FIG. 4 shows a diagram of an example software stack that resides in an IBLAN emulating node according to an example embodiment of the present invention. The stack consists of a network protocol layer 30, one or more Infiniband LAN (IBLAN) driver(s) 34, 36, a transport services library layer 44, along with an IBA name services and subnet manager interface 54 and IBA bus driver 56, a host channel adaptor (HCA) driver 58, and a host channel adapter 60. The stack may also include an intermediate driver 32 for load balancing and failover. Intermediate driver 32 driver may reside between network protocol layer 30 and the IBLAN driver(s) 34, 36.
 Network protocol stack 30 may include any protocol, for example, TCP/IP, NetWare, Open Systems Interconnections (OSI), DECnet, AppleTalk, etc. Intermediate drivers 32 may be layered between the protocol stacks and multiple IBLAN drivers. Intermediate drivers 32 may consolidate multiple instances of IBLAN drivers into one and may manage the load balances and failover across two or more ports (e.g., two in FIG. 4). Each IBLAN driver 34, 36 may include packet data transfer services 38 for unicast, multicast and broadcast transfers across an Infiniband fabric 10, host to host connection services 40 that discovers and resolves connection paths between hosts (by communicating with subnet manager on fabric), and driver initialization function 42 used to initialize an IBLAN driver. Each IBLAN driver implementation 34, 36 establishes policy for managing connections between nodes based on the destination MAC address. If Infiniband channels are relatively cheap based on hardware and memory requirements, then drivers may wish to establish node to node connections during address resolution protocol processing and keep the channels active indefinitely instead of aging (giving the channels back after use) them. If connection aging is performed at the driver level, it may be desirable to sink up the IBLAN driver with the address resolution protocol aging table process to insure that subsequent address resolution protocol processing is provided to initiate new connections.
 The Infiniband Architecture currently defines multicasting within the fabric as an optional feature. Since multicasting is optional, an IBLAN driver according to the present invention provides multicasting and broadcasting in software to ensure interoperability with all and any hardware, including hardware without multicasting (e.g., first generation hardware).
 Transport services library 44 provides Infiniband transport services which include connection management, work queue management, memory management, and message pool management. The IBLAN driver 34, 36 uses the service layer to establish connections and send data to any peer IBLAN driver on the fabric. Transport services library 44 includes: channel services datagram and connections section 46 which includes message and DMA channels 48; a resource manager that manages the message pools; and a connection manager 52. Channel services 46 performs segmentation and reassembly of datagrams so that the maximum transfer unit (MTU) for IBLAN drivers may exceed the 256 byte limit of minimum size Infiniband Architecture packets. Further, an IBLAN driver is allowed to report one MTU to the protocol drivers that may be used for both messages on unreliable connections (unicast) and messages on unreliable datagrams (multicast, broadcast, etc.). Connection manager 52 discovers the remote node's datagram work queue pair. The name service 54, TSL connection manager 52, and the TSL channel services 46 may be used to support multicasting and broadcasting by the IBLAN driver.
 Infiniband Architecture name services and subnet manager interface 54 may be used by IBLAN driver 34, 36 to get a list of active nodes on the fabric and locate the appropriate port and LID for each remote IBLAN interface. This interface also supports periodic queries or event notification which indicates nodes coming and going. The Infiniband Architecture defines subnet administration that manages a subnet. Subnet administration via a subnet management database (SMDB) provides persistent storage of subnet topology, and events and configuration information. Infiniband Architecture name services and subnet management interface 54 provides class drivers with an application programming interface (API) and interface to querythe SMDB and schedule events. This interface may be used to locate all active remote IBLAN nodes on the fabric. Path information to remote IBLAN nodes on the fabric may be provided via this mechanism so that an IBLAN driver may maintain primary and secondary paths for redundancy. An IBLAN driver according to the present invention may periodically query the SMDB for link and node activity. The following are example API calls from an IBLAN driver to a subnet manager to query and get LID's back: “IbaNsGetPlatformGuidListByDeviceType( )”, “IbaNsGetPortGuidListByPlatformGuid( )”, and “IbaNsGetLidListByPortGuid( )”.
 Infiniband Architecture bus driver 56 loads and IBLAN driver when a local port is initialized with a LID and is set to the active state. Infiniband Architecture bus driver 56 also may provide an interface to the IBLAN driver which returns the LID and the LID mask of this new activated port. In this example embodiment, bus driver 56 loads two instances of the IBLAN driver and gives the first one the LID assigned to port one and the second the LID assigned to port two.
 The Infiniband Architecture defines a configuration manager (CFM) that acts as the agency to manage ownership and sharing of I/O controllers (IOC) by hosts. The CFM provides data maintained in the configuration management database (CMDB). Access to the CMDB may be provided by configuration management class MADS. Each host loads an Infiniband Architecture bus driver that discovers IOCs, generates plug and play objects, and provides drivers with the appropriate Infiniband Architecture information for connectivity. In addition to the remote IOCs, the bus driver may also discover all local host channel adapters (HCAs) and ports for IBLAN driver initialization. A vendor ID and device ID may be used to locate and load the appropriate IBLAN driver at a node. An instance of an IBLAN driver may be expected to be loaded for each active port. Each port is treated like a network interface card (NIC) so that load balancing (multiplexing data between two or more channels which increases performance) and failover (switching between paths or ports) may be done with intermediate network device interface (NDIS) drivers, similar to existing PCI NICs. Intermediate driver 32 may only bundle NICs that are on the same Infiniband Architecture subnet.
 Hardware channel adaptor driver 58 drives host channel adaptor 60. In this example embodiment, host channel adaptor 60 contains two ports 62 and 64. As noted previously, an IBLAN driver 34, 36 may be associated with each port 62, 64 respectively.
 Host channel adaptor driver 58 controls the low level hardware interface. Host channel adaptor driver 58 provides a verbs (defined in the Infiniband Architecture specification) API for upper level layers needing Infiniband transport services.
FIG. 5 shows a block diagram of an example initialization sequence of an IBLAN driver according to an example embodiment of the present invention. Bus driver 56 provides IBLAN driver 34 with adaptor or local port information (example API call—“IbaBdGetLocalEndPointinfoByPdo”). Name service 54 provides destination and path information to IBLAN driver 34 (example API calls noted previously). Subnet driver 90 provides path information to IBLAN driver 34 (example API call—“IbaSnGetPathByPortLids”, once each for getting primary and secondary paths). TSL 44 provides connection and data transfer services to IBLAN driver 34.
FIG. 6 shows a system diagram of an example bridge node between an Infiniband fabric and Ethernet network according to an example embodiment of the present invention. As shown in FIG. 6, an Infiniband fabric 10 includes node devices 12, 14, 16 and 18. However, node device 16 also has another port that connects to an Ethernet network 80. Ethernet network 80 also contains additional node devices 82, 84 and 86. Node devices 12-16 contain IBLAN drivers according to the present invention, therefore, a network protocol in node 12 may send an Ethernet data transfer across IB fabric 10 to node device 16 which then may transfer the Ethernet data transfer onto Ethernet network 80 to one or more of node devices 82-86. This is advantageous in that a network protocol residing at node device 12 need not know that the Ethernet traffic that is being sent to a node on an Ethernet network, e.g., 80, has transferred across an Infiniband fabric to get there.
 The present invention is advantageous in that it is the first implementation of an 802.3 LAN emulation for an Infiniband Architecture. Further, according to the present invention no specialized name servers and address resolution protocol are required.
 Moreover, the present invention is not restricted to a TCP/IP protocol only, but imbeds an Infiniband link level local identifier (LID) address in an 802.3 Ethernet MAC address so that any protocol may run on top of Infiniband (IB) fabrics. Also, regarding broadcasting, the present invention avoids the buffer copy by posting the same buffer to each separate Infiniband channel. The present invention also provides a mechanism to support multicast traffic over Infiniband fabrics. In addition, the present invention provides a mechanism to fail-over to secondary paths via the same port. Moreover, a load balance and fail-over driver may be stacked on top of IB LAN drivers to provide redundancy across multiple ports and/or channel-adaptors. The present invention may use a combination of channel and datagram services to provide scalability even with channel adaptors that have limited channel work queue resources.
 It is noted that the foregoing examples have been provided merely for the purpose of explanation and are in no way to be construed as limiting of the present invention. While the present invention has been described with reference to a preferred embodiment, it is understood that the words which have been used herein are words of description and illustration, rather than words of limitation. Changes may be made within the purview of the appended claims, as presently stated and as amended, without departing from the scope and spirit of the present invention in its aspects. Although the present invention has been described herein with reference to particular methods, materials, and embodiments, the present invention is not intended to be limited to the particulars disclosed herein, rather, the present invention extends to all functionally equivalent structures, methods and uses, such as are within the scope of the appended claims.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US6172981 *||Oct 30, 1997||Jan 9, 2001||International Business Machines Corporation||Method and system for distributing network routing functions to local area network stations|
|US6256314 *||Aug 11, 1998||Jul 3, 2001||Avaya Technology Corp.||Apparatus and methods for routerless layer 3 forwarding in a network|
|US6377990 *||Jun 15, 1998||Apr 23, 2002||Lodgenet Entertainment Corporation||System for providing internet access from locations different from those for which the user's software was configured|
|US6411625 *||Feb 27, 1998||Jun 25, 2002||Nec Corporation||ATM-LAN network having a bridge that establishes communication with or without LAN emulation protocol depending on destination address|
|US6694361 *||Jun 30, 2000||Feb 17, 2004||Intel Corporation||Assigning multiple LIDs to ports in a cluster|
|US6711162 *||Apr 12, 2000||Mar 23, 2004||3Com Corporation||Method and apparatus for providing proxy service, route selection, and protocol conversion for service endpoints within data networks|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US7111101 *||May 7, 2003||Sep 19, 2006||Ayago Technologies General Ip (Singapore) Ptd. Ltd.||Method and system for port numbering in an interconnect device|
|US7245627 *||Apr 23, 2002||Jul 17, 2007||Mellanox Technologies Ltd.||Sharing a network interface card among multiple hosts|
|US7533184 *||Jun 13, 2003||May 12, 2009||Microsoft Corporation||Peer-to-peer name resolution wire protocol and message format data structure for use therein|
|US7584274 *||Jun 15, 2004||Sep 1, 2009||International Business Machines Corporation||Coordinating use of independent external resources within requesting grid environments|
|US7590623||Jan 6, 2005||Sep 15, 2009||International Business Machines Corporation||Automated management of software images for efficient resource node building within a grid environment|
|US7664844||Aug 20, 2008||Feb 16, 2010||International Business Machines Corporation||Managing network errors communicated in a message transaction with error information using a troubleshooting agent|
|US7668741||Jan 6, 2005||Feb 23, 2010||International Business Machines Corporation||Managing compliance with service level agreements in a grid environment|
|US7707288||Jan 6, 2005||Apr 27, 2010||International Business Machines Corporation||Automatically building a locally managed virtual node grouping to handle a grid job requiring a degree of resource parallelism within a grid environment|
|US7712100||Sep 14, 2004||May 4, 2010||International Business Machines Corporation||Determining a capacity of a grid environment to handle a required workload for a virtual grid job request|
|US7724748||May 9, 2006||May 25, 2010||Intel Corporation||LAN emulation over infiniband fabric apparatus, systems, and methods|
|US7734679||Sep 16, 2008||Jun 8, 2010||International Business Machines Corporation||Managing analysis of a degraded service in a grid environment|
|US7739155||May 22, 2008||Jun 15, 2010||International Business Machines Corporation||Automatically distributing a bid request for a grid job to multiple grid providers and analyzing responses to select a winning grid provider|
|US7743142 *||Jan 23, 2009||Jun 22, 2010||International Business Machines Corporation||Verifying resource functionality before use by a grid job submitted to a grid environment|
|US7761557||Jan 6, 2005||Jul 20, 2010||International Business Machines Corporation||Facilitating overall grid environment management by monitoring and distributing grid activity|
|US7765327 *||Sep 1, 2005||Jul 27, 2010||Intel Corporation||Intermediate driver having a fail-over function|
|US7788375||Feb 2, 2009||Aug 31, 2010||International Business Machines Corporation||Coordinating the monitoring, management, and prediction of unintended changes within a grid environment|
|US7793308||Jan 6, 2005||Sep 7, 2010||International Business Machines Corporation||Setting operation based resource utilization thresholds for resource use by a process|
|US7843962 *||Nov 30, 2010||Obsidian Research Corporation||Method to extend the physical reach of an infiniband network|
|US7844715 *||Aug 27, 2008||Nov 30, 2010||Qlogic, Corporation||System and method for a shared I/O subsystem|
|US7921133||Jun 23, 2007||Apr 5, 2011||International Business Machines Corporation||Query meaning determination through a grid service|
|US7962651||Jun 13, 2005||Jun 14, 2011||Microsoft Corporation||Peer-to-peer name resolution protocol (PNRP) and multilevel cache for use therewith|
|US7971187||Apr 24, 2006||Jun 28, 2011||Microsoft Corporation||Configurable software stack|
|US7983275||May 20, 2010||Jul 19, 2011||Intel Corporation||LAN emulation over infiniband fabric apparatus, systems, and methods|
|US8136118||May 21, 2009||Mar 13, 2012||International Business Machines Corporation||Maintaining application operations within a suboptimal grid environment|
|US8165138 *||Dec 4, 2007||Apr 24, 2012||International Business Machines Corporation||Converged infiniband over ethernet network|
|US8228913 *||Sep 29, 2008||Jul 24, 2012||International Business Machines Corporation||Implementing system to system communication in a switchless non-IB compliant environment using InfiniBand multicast facilities|
|US8239498||Oct 28, 2005||Aug 7, 2012||Bank Of America Corporation||System and method for facilitating the implementation of changes to the configuration of resources in an enterprise|
|US8255546 *||Sep 30, 2005||Aug 28, 2012||Microsoft Corporation||Peer name resolution protocol simple application program interface|
|US8275881||May 4, 2009||Sep 25, 2012||International Business Machines Corporation||Managing escalating resource needs within a grid environment|
|US8331381 *||Dec 4, 2007||Dec 11, 2012||International Business Machines Corporation||Providing visibility of Ethernet components to a subnet manager in a converged InfiniBand over Ethernet network|
|US8346591||Jun 9, 2009||Jan 1, 2013||International Business Machines Corporation||Automating responses by grid providers to bid requests indicating criteria for a grid job|
|US8387058||Jun 21, 2008||Feb 26, 2013||International Business Machines Corporation||Minimizing complex decisions to allocate additional resources to a job submitted to a grid environment|
|US8396757||Jun 24, 2009||Mar 12, 2013||International Business Machines Corporation||Estimating future grid job costs by classifying grid jobs and storing results of processing grid job microcosms|
|US8503468 *||Nov 5, 2008||Aug 6, 2013||Fusion-Io, Inc.||PCI express load sharing network interface controller cluster|
|US8542689||Jul 15, 2011||Sep 24, 2013||Intel Corporation||LAN emulation over infiniband fabric apparatus, systems, and methods|
|US8583650||Aug 4, 2009||Nov 12, 2013||International Business Machines Corporation||Automated management of software images for efficient resource node building within a grid environment|
|US8612862||Jun 27, 2008||Dec 17, 2013||Microsoft Corporation||Integrated client for access to remote resources|
|US8683062||Feb 28, 2008||Mar 25, 2014||Microsoft Corporation||Centralized publishing of network resources|
|US8782201 *||Oct 28, 2005||Jul 15, 2014||Bank Of America Corporation||System and method for managing the configuration of resources in an enterprise|
|US20050004916 *||Jun 13, 2003||Jan 6, 2005||Microsoft Corporation||Peer-to-peer name resolution wire protocol and message format data structure for use therein|
|US20050198242 *||Jan 5, 2004||Sep 8, 2005||Viascope Int.||System and method for detection/interception of IP collision|
|US20050267992 *||Jun 13, 2005||Dec 1, 2005||Microsoft Corporation||Peer-to-peer name resolution protocol (PNRP) and multilevel cache for use therewith|
|US20050278441 *||Jun 15, 2004||Dec 15, 2005||International Business Machines Corporation||Coordinating use of independent external resources within requesting grid environments|
|US20060048157 *||May 18, 2004||Mar 2, 2006||International Business Machines Corporation||Dynamic grid job distribution from any resource within a grid environment|
|US20060059492 *||Sep 14, 2004||Mar 16, 2006||International Business Machines Corporation||Determining a capacity of a grid environment to handle a required workload for a virtual grid job request|
|US20060149576 *||Jan 6, 2005||Jul 6, 2006||Ernest Leslie M||Managing compliance with service level agreements in a grid environment|
|US20060149652 *||Jan 6, 2005||Jul 6, 2006||Fellenstein Craig W||Receiving bid requests and pricing bid responses for potential grid job submissions within a grid environment|
|US20060149714 *||Jan 6, 2005||Jul 6, 2006||Fellenstein Craig W||Automated management of software images for efficient resource node building within a grid environment|
|US20060150158 *||Jan 6, 2005||Jul 6, 2006||Fellenstein Craig W||Facilitating overall grid environment management by monitoring and distributing grid activity|
|US20060150190 *||Jan 6, 2005||Jul 6, 2006||Gusler Carl P||Setting operation based resource utilization thresholds for resource use by a process|
|US20060200584 *||Sep 1, 2005||Sep 7, 2006||Intel Corporation||Intermediate driver having a fail-over function|
|US20060203846 *||May 9, 2006||Sep 14, 2006||Intel Corporation||LAN emulation over infiniband fabric apparatus, systems, and methods|
|US20070014308 *||Jul 17, 2006||Jan 18, 2007||Gunthorpe Jason G||Method to extend the physical reach of an infiniband network|
|US20070076630 *||Sep 30, 2005||Apr 5, 2007||Microsoft Corporation||Peer name resolution protocol simple application program interface|
|US20100082853 *||Sep 29, 2008||Apr 1, 2010||International Business Machines Corporation||Implementing System to System Communication in a Switchless Non-IB Compliant Environment Using Infiniband Multicast Facilities|
|US20100115174 *||Nov 5, 2008||May 6, 2010||Aprius Inc.||PCI Express Load Sharing Network Interface Controller Cluster|
|US20110022526 *||Apr 8, 2010||Jan 27, 2011||Bruce Currivan||Method and System for Content Selection, Delivery and Payment|
|US20140177639 *||Dec 20, 2012||Jun 26, 2014||Mellanox Technologies Ltd.||Routing controlled by subnet managers|
|WO2006107133A1 *||Nov 28, 2005||Oct 12, 2006||Chanwoo Kim||Ip management method and apparatus for protecting/blocking specific ip address or specific device on network|
|U.S. Classification||370/389, 370/401, 370/465|
|International Classification||H04L12/46, H04L29/12|
|Cooperative Classification||H04L49/358, H04L61/10, H04L12/4604, H04L29/12018, H04L29/12009, H04L12/4608|
|European Classification||H04L61/10, H04L29/12A, H04L12/46B, H04L29/12A1|
|Mar 19, 2001||AS||Assignment|
Owner name: INTEL CORP., CALIFORNIA
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DAVIS, ARLIN R.;REEL/FRAME:011599/0054
Effective date: 20010313