Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS7620981 B2
Publication typeGrant
Application numberUS 11/139,206
Publication dateNov 17, 2009
Filing dateMay 26, 2005
Priority dateMay 26, 2005
Fee statusPaid
Also published asUS8387132, US8726363, US20060272015, US20100095023, US20120195315
Publication number11139206, 139206, US 7620981 B2, US 7620981B2, US-B2-7620981, US7620981 B2, US7620981B2
InventorsCharles William Frank, Thomas Earl Ludwig, William R. Babbitt, Jr., Mark Adams, Nicholas James Witchey
Original AssigneeCharles William Frank, Thomas Earl Ludwig, Babbitt Jr William R, Mark Adams, Nicholas James Witchey
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Virtual devices and virtual bus tunnels, modules and methods
US 7620981 B2
Abstract
Modules and methods that provide virtual device communication via information packets addressed to the virtual devices. Virtual devices aggregate capabilities of target devices and represent the aggregate as an operational device to remote systems. Physical device elements and virtual devices contained within modules are able to communicate and cooperate over extended distances in a peer-to-peer fashion via a virtual bus tunnel offering addressing and data transport functionality. Such modules and methods can be combined with disk drives to form RAID-like storage structures.
Images(9)
Previous page
Next page
Claims(15)
1. A method comprising:
associating, with a packet processor, a first virtual device with a first physical device element, which is a disk drive, and a second virtual device with a second physical device element;
assigning, by the packet processor, the first virtual device a first host address and the second virtual device a second host address, wherein the first and second host addresses are different network routable addresses that are both associated with a common frame address that corresponds to a network interface that communicatively couples the packet processor to a network;
receiving, by the packet processor, an information packet from the network interface, the information packet having a control portion including the first host address;
routing, by the packet processor, the information packet to the first virtual device; and
associating, by the packet processor, the first virtual device with a plurality of disk drives, wherein the first virtual device is a multicast group of logical partitions distributed over the plurality of disk drives.
2. The method of claim 1, wherein said receiving comprises receiving the information packet from the network interface via a virtual bus tunnel.
3. The method of claim 1, wherein the information packet is a first information packet and the method further comprises:
receiving a second information packet from the second virtual device determining that the second information packet has a device format;
configuring the second information packet with a host address of a third virtual device; and
transmitting the second information packet to the third virtual device via a virtual bus tunnel.
4. The method of claim 1, further comprising:
accessing, by the packet processor, a localized domain name server system; and
routing the information packet based at least in part on said accessing.
5. The method of claim 1, wherein the first and second virtual devices are respectively represented by first and second software constructs stored in a memory of a transform module.
6. A transform module comprising:
a network interface having a frame address and configured to communicatively couple the transform module to a network;
a packet processor coupled with the network interface and configured to associate a first virtual device with a first physical device element, which is a disk drive, and a second virtual device with a second physical device element;
to assign the first virtual device a first host address and the second virtual device a second host address, wherein the first and second host addresses are different network routable address;
to receive an information packet via the network interface, the information packet having a control portion including the first host address;
to route the information packet to the first virtual device; and
to associate the first virtual device with a plurality of disk drives, wherein the first virtual device is a multicast group of logical partitions distributed over the plurality of disk drives.
7. The transform module of claim 6, wherein the first virtual device is an application, a program, a task, a thread, a function, and/or a device driver.
8. The transform module of claim 6, wherein the first host address comprises an IPv4 or an IPv6 address.
9. The transform module of claim 6, further comprising:
a target device interface that provides the transform module access to the first and second physical device elements.
10. The transform module of claim 9, wherein the target device interface includes a serial AT attachment (SATA) interface, a parallel AT attachment (PATA) interface, a small computer system interface (SCSI), and/or fiber channel interface.
11. The transform module of claim 9, wherein the target device interface is a homogenous interface.
12. The transform module of claim 9, wherein the target device interface is a heterogenous interface.
13. The transform module of claim 6, wherein the packet processor is further configured to route the information packet to the first virtual device via a virtual bus tunnel.
14. A computing system comprising:
an operating system configured to be communicatively coupled with an application, to receive a plurality of calls from the application, and to issue, based at least in part on the plurality of calls, a plurality of device-level requests having device formats; and
a transform module coupled with the operating system and configured
to receive the plurality of device-level requests and to generate a first information packet, which includes a first host address, and a second information packet, which includes a second host address, wherein the first and second host addresses are distinct, network-routable addresses that are respectively assigned to first and second virtual devices of a networked device, wherein the first virtual device is associated with a plurality of disk drives and the first virtual device is a multicast group of logical partitions distributed over the plurality of disk drives; and
to transmit the first and second information packets to the networked device through a virtual bus tunnel.
15. The computing system of claim 14, wherein transform module is configured to generate the first information packet with a split identifier including the first host address and a logical block address.
Description
FIELD OF THE INVENTION

The field of the invention is virtual device communication.

BACKGROUND OF THE INVENTION

Computing systems and devices from the smallest of embedded systems to the largest of super computers require resources to accomplish their objectives. Possible resources include CPUs, memory, monitors, hard disk drives, networks, peripherals, and a host of other devices that combine to form a complete computing experience for the user. Over the last decade, users of computing systems and devices have become less sophisticated with respect to the complexity of the computing systems. Users are now more concerned with their computing experience rather than in the details of how their computing system operates. The migration of computing systems toward less sophisticated users has pushed such systems to become more and more complex to compensate for the users lack of computer system expertise. The results of this migration can be seen by the continued increase in complexity and power of current operating systems, including Windows® or Linux.

Computers running Windows® or Linux present a single data portal to a user, where the computer contains nearly all aspects of the user's environment including data, applications, preferred settings, and so on. However, many users currently need multiple portals to their computing environment rather than a single system as they travel from home, to abroad, or to work. Not only do users wish to have access to their data as they move, but they also wish to take their computing experience with them. Laptop computers offer the ability to move, but at great expense. Consequently, numerous systems have been developed to enable a more decentralized user experience approach, including web portals such as Yahoo!®, where a user accesses their email interface from anywhere in the world. Users are no longer concerned about the location of their computing resources, but are rather concerned about having access to their data and having the same computing experience no matter where they are or what device they use as an interface.

A number of problems arise due to users demand for a uniform experience. First, huge computing services including those being developed by Yahoo!™, Microsoft®, and even Google™ require large centralized server systems to house the massive amount of user's data. Such servers are expensive to operate and maintain. Second, users' experiences tend to be limited to the capabilities these services have to offer, (including email, photo organization, web browsing, searching, and so forth), rather than a complete generalized experience. Third, these large centralized systems do not present users with their complete data set, but rather a small fraction of the data. Nor can the systems necessarily run general applications that users desire when working with their data.

In order to address these limitations and provide a cost effective computing experience for users, a fully decentralized approach is required. Some decentralized computing efforts are well known, including for example, Search for Extra-Terrestrial Intelligence (SETI@Home), the Great Internet Mersenne Prime Search (GIMPS), and various file sharing programs such as BitTorrent, and Kazaa. Although average users can access these applications and programs, the applications offer only a limited decentralized user's experience, which focus on specific capabilities or services rather than the generalized computing needs of the user.

Several decentralized computing platforms exist today that are intended to be generalized, including Beowulf Linux clusters and advanced web services including Microsoft .NET. Unfortunately, these and other decentralized computing platforms have several flaws that impact the user. Among other things they require new software that the user must then purchase or license (as in the case of .NET), or write themselves as in the case of clusters. A further problem is that the new software still requires strong centralized computing systems to support the overall computing experience. Such approaches attack the problem from an application or user point of view, which causes the solution space to be too broad. A broad point of view then tends to limit the solution to course grained approach that results in a high cost of implementation to the user.

A better approach to decentralization of computing resources attacks the problem from a more fined grained perspective, where the system scales incrementally in a more cost effective manner rather than at a systems level were incremental costs are high. Rather than approaching the problem from a computing system standpoint, decentralization can be approached from a fundamental computing resource perspective, in which a computing resource distills to the physical device element level (where a device element could include, but not limited to, a CPU, memory, hard disk, or even a monitor). When these devices combine in a decentralized fashion, a wide range of users gain the benefits of decentralization. In addition, decentralization of devices allows devices to focus on doing what they do best and to be responsible for their own services and capabilities; thereby offloading peripheral management work from an operating system and CPU. Consequently, the operating system and CPU can focus on providing more bandwidth to an application. Decentralization also provides the benefit of scalability, such that a computing system scales up at the atomic device level as opposed to at the system level. Users scale their computing systems by purchasing devices which are more cost effective than complete systems. In addition, older generations of decentralized devices still remain useful and contribute to the over computing system without having to be discarded.

Decentralization, by its very nature, includes the concept of virtualization, where devices are distributed over a network but “appear” as locally connected from the perspective of other devices. With device virtualization provided over a network, the computing resources no longer have to be local to the user. This implies a user gains access to their computing experience and resources from any connected location once the user takes appropriate authentication or security measures. Such a decentralized computing system addresses the remaining problems with current architectures: users maintain access to their data, users can use multiple interfaces while maintaining their computing experience, applications do not have to be rewritten because decentralization occurs at the device level rather than the system level, and costs to the user decrease because the user scales their system by purchasing devices at the atomic unit rather than buying at the system level.

Beyond the benefits to the user, decentralization at the device level also creates opportunities for creating new capabilities or services by using decentralized devices as building blocks. Just as devices focus on doing what they do well, new products can be built that aggregate device level capabilities or services into more powerful capabilities or services while maintaining scalability at device level. The new products offer the powerful capabilities or services as a “device” to the rest of the system even though the “devices” are actually virtual in nature. For instance, hard disk drives can be aggregated into a network storage system with RAID capabilities, where the entire storage system appears as a single storage device. The network storage system focuses on storing and retrieving data, without incurring overhead associated with complete system architectures. Because the drives are decentralized and virtualized, the storage system can still be scaled at the atomic level, the hard disk drive, which is much more cost effective to the end user.

To fully realize the benefits and capabilities offered by decentralization, devices and virtual devices need to pass information to each other in an organized way, as if they exist on a virtualized bus. Therefore, there is a need for modules and methods that:

    • Process information packets addressed to groups of devices, individual devices, or virtual devices;
    • Includes interfaces to target devices and to some form of network;
    • Extend device virtual buses over large distances so device can be accessed from any location; and
    • Manage virtual devices offering aggregated behavior.

Such a set of modules combine with target devices to offer aggregated capabilities, and can also integrate into existing computing systems. When installed in a computing system such as Windows®, the module could take the form of a device driver and the aggregate capabilities or services of the remote target devices appear as a single device.

A number of benefits are realized from this approach. First, computing systems employ such modules to distribute the locality of their physical device elements. This is beneficial in cases where physical security is needed or where geographic isolation offers protection against acts of God. Second, users purchase enhanced capabilities by purchasing computing systems resources in discreet units rather than complete systems. If they wish to double the computing power of their system, the consumer purchases a single CPU (assuming an appropriate module is attached to the CPU) and adds it to the decentralized system. The cost for a single device is much less than the cost of a complete system. Along these lines, old components can be retained and still provide useful capabilities even though they are not as powerful as the new component because the old components can still function within the decentralized system. Third, the customer has access to their experience from anywhere in the world because the virtual system bus extends to any length. Once a user authenticates to their environment, the distributed system presents the user with their preferred experience to them no matter the actual location of the discreet resources housing the user's data. Forth, more complex structures can be created from the discreet device building blocks. For instance, sophisticated RAID structures can be created by using hard drives and appropriately programmed modules attached to the drives. The RAID structures appear as a single drive to the user or the user's applications.

Thus, specific compositions and methods of virtual device communication, information packets transformation, and virtual bus tunnels have been disclosed. It should be apparent, however, to those skilled in the art that many more modifications besides those already described are possible without departing from the inventive concepts herein. The inventive subject matter, therefore, is not to be restricted except in the spirit of the disclosure. Moreover, in interpreting the disclosure all terms should be interpreted in the broadest possible manner consistent with the context. In particular the terms “comprises” and “comprising” should be interpreted as referring to the elements, components, or steps in a non-exclusive manner, indicating that the referenced elements, components, or steps can be present, or utilized, or combined with other elements, components, or steps that are not expressly referenced.

SUMMARY OF THE INVENTION

The present invention is directed toward virtual devices and information packets used to communicate among virtual devices, where the information packets are addressed to virtual devices. Virtual devices contained within modules coordinate with other virtual devices within other modules in a peer-to-peer fashion to offer aggregated capabilities. Information packets comprise host addresses of virtual devices and comprise identifiers (IDs). IDs comprise sufficient information for routing the packets over a network to a final destination and comprise sufficient information to determine how the information held within the packet should be conveyed to a physical target device. Modules comprising virtual devices and the ability to create information packets offer a virtual bus tunnel for virtual devices so they can coordinate behavior in a distributed fashion forming. Various embodiments of such modules offer a number of interesting features that allows devices, either local or remote with respect to the target devices, to offer aggregate capabilities to device users than the users would ordinarily have access to at the device level. Modules capabilities include, but not necessarily limited to, distributed data storage systems.

Virtual Devices

In our own prior art we disclosed physical device elements, (i.e. hardware devices) that don't have a unique frame address but do have a host address. Specific examples included disk partitions within a hard drive, RAM, or CPUs. We also disclosed and claimed protocols, methods and systems for using the host addresses to access the physical device elements.

As used herein, the term “virtual device” means anything (hardware, firmware, or software during its execution) that has a host address, but doesn't have a unique frame address, through which to represent itself as an operational device to an external application. A host address is any physical interface independent address, and includes, for example, IP addresses, IPv4 addresses, and IPv6 addresses. Host addresses are also considered herein to include domain names and URLs (web page addresses) because they resolve to a network address based on applications or protocols. A frame address is a physical media interface address within a packet that frames a host address. For example, in Ethernet a frame address includes a MAC address, in Frame Relay a frame address includes the DLCI or the LAN adapter address, and in ATM a frame address includes an AES address. USB usually uses frame addresses, but can also use packets that do not have frame addresses.

Network addressable virtual devices are virtual devices that are addressed at the network level of the OSI model, as opposed for example, to the session, presentation, or application level. Thus, a mapped Windows® drive is addressed at the application level, and therefore would be virtual device, but not a network addressable virtual device. On the other hand, assume that a physical disk drive has partitions P1-Pn. Partition P1 can be a network addressable virtual device using techniques disclosed herein, and it would be a physical device element. A logical partition mapping data block addresses to partition P1 disk locations can also be a network addressable virtual device, but it would not be a physical device element because the mapping of the data blocks to physical disk locations can be arbitrary. A logical group G1 consisting of several logical partitions P1-Pm can also be a network addressable virtual device, and a logical volume V1 consisting of several logical groups G1-Go can also be a virtual device. Of course, neither the logical group G1 nor the logical volume V1 would be physical device elements. Therefore, a collection of network addressable virtual devices can also be considered a network addressable virtual device. Consequently, within this document “virtual device” can refer to both a single virtual device and to a collection of network addressable virtual devices working together.

In preferred embodiments, virtual devices comprise functionality such that the virtual devices can offer remote applications, users, or other devices access to aggregated capabilities of the physical target devices. Such virtual devices do not necessarily have to have a one-to-one correspondence to target devices, but these and other virtual devices do have to represent themselves as operational devices to external applications. For example, an independently addressable logical partition on a disk drive that responds properly to read and write requests based on logical block addresses would be considered a virtual device. Virtual devices also comprise any software construct or functionality including a function, application, program, service, device driver, task, process, thread, or other non-physical device elements. Those skilled in the art of firmware or software development will recognize that a virtual device is a logical construct. Consequently, numerous coding structures for virtual devices are possible including structures that represent multiple virtual devices.

Although multiple virtual devices can share a single frame address, each virtual device within a system must have its own host address to avoid conflict. This is somewhat analogous to “multi-homing” where an individual computer can have multiple IP addresses and multiple network interfaces. But the similarity is not perfect because multi-homing assigns IP addresses only to the physical network interfaces, and not to sets of functionalities as represented by applications or programs. In addition, the multiple IP addresses are usually from different sub-nets to ensure link redundancy should a route fail. Still another difference from multi-homing is the sheer number of addresses available. Because virtual devices can share frame addresses, there may be tens, hundreds, or thousands of addresses. Multi-homed systems usually have only two or three addresses.

In a particularly simple embodiment, a virtual device can merely be an electronic front end to a physical disk drive that includes for example, a processing unit (CPU) and other computer chips, an electronic network interface, an electronic disk drive interface, random access memory (RAM), a printed circuit board (PC board), a power supply, and interconnecting wires or traces. In a more sophisticated embodiment, a virtual device can be a software driver or application, which would of course have to reside on a machine readable memory, and execute on some form of physical processing unit. Non-operating software would not typically have a host address and therefore would not be considered a virtual device.

In still another aspect it is contemplated that one could write software that would execute on a computer or other hardware as a virtual device. From that perspective the inventive subject matter includes methods of writing such software, recording such software on a machine readable medium, licensing, selling, or otherwise distributing such software, installing, and operating such software on suitable hardware. Moreover, the software per se is deemed to fall within the scope of the inventive subject matter.

Information Packets

As used herein an “information packet” is any packet passing through a virtual device containing information. An information packet can include an ID in a split-ID format, a contiguous-ID format, or a device format.

Modules

As used herein, the term “module” means any combination of hardware, software, or firmware that provides: (a) a target device interface; (b) network interface; and (c) packet processor. Packet processor can include a virtual device manager, a virtual device, and a virtual bus tunnel. In instances where the module includes software, the software will, of course, reside at least temporarily on a machine-readable memory, and execute on a processing unit of an electronic device, but either or both of the machine-readable memory and the processing units can be external to the module.

Virtual Device Management

Virtual device management refers to the capability of a module to manage virtual devices within the module that appear to be real devices external to the module, where management comprises all aspects relating to controlling the existence or behavior of the virtual device constructs. It is contemplated that management comprises functionality to create, to destroy, to suspend, to resume, to modify, to secure, or other related functionality.

Virtual Bus Tunnel

In order for the virtual devices to offer their capabilities or render a service to remote systems, the virtual devices need some form of communication bus. Because the virtual devices do not comprise physical interfaces, they need a mechanism for addressing other virtual devices as well as transporting data to the other devices, wherein the mechanism is referred to as a “virtual bus tunnel.” The virtual bus tunnel provides virtual devices the ability to communicate or coordinate their behaviors over extended physical distances such that a collection of devices working together do not necessarily have to be physically located near each other. Furthermore, the virtual bus tunnel allows for packet level communication among physical device elements and virtual devices addressable on the virtual bus tunnel. Communication over the virtual bus tunnel can also include direct communication from an application to virtual devices.

Normal physical buses suffer from bus contention where devices must vie for access to the bus. Virtual bus tunnels resolve contention by grouping virtual devices logically with a group ID such that multiple groups access the same physical transport media without contending for the bus by using the group ID within their packets. Each group ID represents a single logical group of virtual devices sharing a single bus with other logical groups each with their own ID. In addition, each virtual device is individually addressable via its own device ID. It is specifically contemplated that an Internet Protocol network embodies a virtual bus tunnel utilizing IP addresses as host addresses representing device IDs for virtual devices. Under such an embodiment, group IDs can be managed via IP multicasting addresses such that applications, users, or other devices that are members of the group perceive the group as a single device. Because the virtual bus tunnel can offer capabilities including quality of service management, it too can be considered a virtual device and can be addressed via a host address.

Peer-to-Peer Coordination

Virtual devices also need the ability to work together in a coordinated fashion to reach their full potential in providing aggregated capabilities or services. Peer-to-peer communication allows for devices to communicate directly with each other without using multicast where all devices within the multicast group “see” the traffic. Peer-to-peer coordination further provides for virtual devices to be aggregated into yet more powerful capabilities or services. In addition, it is contemplated that modules can be deployed into applications such that the applications have the ability to access the capabilities or services that virtual devices offer. Consequently, modules can be implemented as hardware systems attached to devices or software drivers installed on workstations running operating systems where the virtual devices within each module are programmed to work together in a peer-to-peer fashion.

The teachings herein can be advantageously employed by product developers to create systems where modules coupled with devices or components provide aggregated capabilities or services to multiple applications, users, or other devices in a distributed fashion. Devices that can be enhanced via such modules include hard disk drives, memory, monitors, patient sensors, point of sales scanners, office equipment, building automation sensors, media players, handheld systems, home entertainment systems, or computers. In addition, the modules can be integrated into software applications (if the module comprises all software components), formed into a rack-mount system containing numerous devices including hard disk drives, or single modules attached to a single device. Modules can even be sold as a stand alone product.

Various objects, features, aspects, and advantages of the present invention will become more apparent from the following detailed description of the preferred embodiments of the invention, along with the accompanying drawings in which like numerals represent like components

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1A is a schematic of an unframed information packet in which the host address is an address of a virtual device other than a physical device element.

FIG. 1B is a schematic of a framed information packet including the host address of FIG. 1A.

FIG. 2A is a schematic of a possible embodiment of an information packet utilizing a split-ID format.

FIG. 2B is a schematic of an alternative embodiment of an information packet utilizing a split-ID format.

FIG. 3 is a schematic of a possible embodiment of an information packet utilizing contiguous-ID format.

FIG. 4 is a schematic of a virtual device that aggregates capabilities FIG. 5 is a schematic of a block level logical representation of a module.

FIG. 6 is a schematic of a possible block level physical representation of a module.

FIG. 7 is a schematic of a representation of a virtual bus tunnel providing communication among devices external to a computing system.

FIG. 8 is a schematic of a possible embodiment of a virtual bus tunnel used to create data mirrors, stripes, or spans from partitions on hard drives located in different physical locations.

FIG. 9 is a schematic of a possible set steps used to create a virtual bus tunnel.

DETAILED DESCRIPTION

Virtual devices communicate with remote systems via information packets over a virtual bus tunnel. The information packets originate from a source who wishes to address an entity and direct data to the entity, where an entity on virtual bus tunnel comprises a virtual device. The entity can appear as a single “device” from the perspective of the source.

FIG. 1A illustrates an information packet addressed to a virtual device that is not a physical device element. Information packet 100 comprises a host address 110 used to address a virtual device. Preferred host addresses include routable addresses. Examples of routable addresses including IPv4 address or IPv6 addresses. Consequently, software or firmware employed to create information packets has sufficient intelligence to distinguish the difference between virtual devices, physical device elements, or target devices in order to properly address virtual devices and the capabilities they represent. FIG. 1B illustrates a framed version of the information packet of FIG. 1A. Framing packet 150 comprises frame address 155 and encapsulated information packet 100. An example framing packet includes an Ethernet packet where the MAC addresses corresponds to frame addresses.

Preferred types of information packets include packets utilizing a split-ID format, contiguous-ID format, or device format, where an ID comprises sufficient information for determining where the packets should go or how a packet should interact with a final target device.

IDs

Within the context of an information packet, packets comprise IDs that contain all necessary information for routing the packets to a correct entity comprising a module, object within a module, a device, or virtual device. IDs comprise at least three sub-IDs which include group IDs, device IDs, or target IDs. Group IDs identify a logical grouping of entities such that the logical group can be addressed via a single ID. Using an IP address as a group ID allows multiple entities to be addressed as if they are a single “device.” Device IDs identify specific entities in a group. It is contemplated that specific entities comprise non-physical objects including virtual devices. Target IDs identify a specific entity's capability or characteristic.

Split-ID Format

FIG. 2A depicts an embodiment of a split-ID format within an information packet. Referring to FIG. 2A, information packet 200 comprises several smaller portions, control portion 220 and data portion 215. Nominally, a module uses control portion 220 to determine how to control a packet including figuring out to which entity within the module, if any, the packet should be routed. Data portion 215 comprises an encapsulated packet 230 that can be further used by the module or the entity to determine the final disposition of the information within the packet.

Although FIG. 2A displays a contiguous ID segment in the control portion of each packet and encapsulated packet, it is contemplated that any arrangement or number of ID segments within the control portion or data portion of a packet such that the complete ID information can be reconstructed falls within the scope of the split-ID concept. In addition, the control portion of a packet does not necessarily have to appear at the beginning of a packet, but could reside anywhere within a packet.

FIG. 2B fully represents a split-ID, where information packet 200B has a plurality of encapsulated packets within its data portion. Encapsulated packet 230A further comprises control portion 220B and an encapsulated packet 230B. As shown, the nesting of encapsulated packets is arbitrary. However, the complete ID of information packet 200B comprises split-ID 210A through split-ID 210N. Therefore, “split-ID” refers to the ID of the packet being split among encapsulated packets. One should note each piece of the ID is not necessarily related to the other pieces of the ID. In some embodiments, one part of the ID represents a group ID or device ID comprising routable address information and other ID pieces represent target ID comprising information relating to the physical structure of a target device. In some cases, group IDs comprise tags or names for easier management. In more preferred embodiments, the routable addresses comprise IP addresses assigned to storage areas and other ID pieces comprise Logical Block Addresses (LBAs) associated with logical partitions that map the LBAs to disk partition locations.

Contiguous-ID Format

Information packets using a split-ID format represent a more preferred approach for addressing and for data transport using modules; however, other approaches are also possible. FIG. 3 illustrates information packet 300 comprising a control portion 320 and data portion 315. Control portion 320 further comprises single contiguous-ID 310, where contiguous-ID 310 comprises information for the module to resolve the final disposition of the packet. Contiguous-ID 310 is formed by having all the ID information contained in one unit of data. It is contemplated that contiguous-ID 310 comprises host addresses including IP addresses where the IP address can also map to a device specific characteristic. It is more specifically contemplated that contiguous-ID 310 comprises an IPv4 or IPv6 address that maps directly to an LBA of a logical partition that maps to a disk drive partition. Contiguous-IDs comprise the elements of a group ID, device ID, or target ID.

FIG. 3 displays control portion 300 at the beginning of packet 320. One should note the placement of control portion does not necessarily have to be at the beginning of the packet, but could reside anywhere within the packet. Furthermore, more than one contiguous-ID can be present within packet 320.

Device Format

Information packets using device formats represent data packets received from a target device or sent to a target device. Contemplated packets are raw packets with no additional formatting added to them such that a target device perceives the packets as a naturally presented packet for the target device's interface. Information packets using device formats are device specific, consequently, they can vary from module to module depending on what types of target devices attach to the modules. In some embodiments, a module can be able to handle many different types of device formats directed toward a heterogeneous mix of target devices.

Aggregated Capabilities

FIG. 4 illustrates the aggregation of a set of capabilities. Module 400 comprises virtual device 420 that aggregates a set of capabilities 424, 426, and 428. Capabilities 424, 426, and 428 comprise capabilities programmed into virtual device 420 or capabilities of target devices accessible to virtual device 420. Contemplated virtual devices include logical drive partitions that aggregate the data storage capabilities across multiple hard disk drives. Even though FIG. 4 shows three capabilities, a virtual device aggregates any number of capabilities. In addition, it is contemplated that the capabilities can comprise a heterogeneous mix of capabilities.

Module Capabilities

FIG. 5 illustrates a logical block diagram of a possible module 500 capable of creating information packets addressed to virtual devices used to communicate with remote systems. Module 500 comprises network interface 510, packet processor 520, and target device interface 530. Network oriented data path 505 flows bi-directionally to and from network interface 510. Packet data path 515 also flows bi-directionally to and from network interface 510 and packet processor 520. In addition device data path 535 flows bi-directional between packet processor 520 and target device interface 530. Target device interface 530 provides for physically handing packets from module 500 to a number of target devices. It is contemplated the number of target devices can vary from one to an arbitrary number as represented by devices 540A through 540M. When target device interface 530 interacts with target devices 540A through 540M, the device data passes through target device data paths 545A through 545M that correspond to their respective target devices.

Packet processor 520 comprises the capability for processing the various types of information packets including packets with split-ID, contiguous-ID, or device formats. In order to process the packets, packet processor further comprises virtual device manager 526 and virtual bus tunnel 524 both of which represent further functionality for packet handling. Both of these elements are under the command and control of packet processor 520. Virtual device manager 526 further comprises at least one of a possible plurality of virtual devices 528A through 528N. It is contemplated that the number of virtual devices can vary as shown depending on the requirements of module 500. Virtual devices 528A through 528N each have a bi-directional device data path 537A through 537N, respectively, providing access to target devices 540A through 540M. In addition, virtual devices 528A through 528N each have a bi-directional packet data path 525A through 525N respectively used to access virtual bus tunnel 524. Virtual bus tunnel 524 provides the virtual devices a communication path 517 to other modules, virtual devices, or system external to module 500 via network interface 510. Contemplated data paths within the module include Application Program Interfaces (APIs), memory copies, Inter-Process Communications (IPCs), physical connections, or other data transfer mechanisms.

FIG. 5 represents a logical view of the roles and responsibilities within module 500. In practice the logical constructs can be combined together in various ways that best suite the desired implementation. For instance, each logical block separated into individual tasks or threads, or all the blocks could be combined into a single monolithic executable image.

Network Interface

Network interface 510 has responsibilities comprising providing network oriented data path 505 from module 500 to remote modules and systems that desire access to the capabilities or services offered by module 500 and the objects it manages. It is contemplated that preferred embodiments of network interface 510 comprise use of an internetworking protocol to manage communication. In more preferred embodiments of network interface 510, the internetworking protocol comprises IPv4 or IPv6. In addition, because module 500 can be placed in a general purpose computing system and function as driver level software, network interface 510 is contemplated to comprise an API that provides access to the general purpose computing system's network port. It is also contemplated communications with entities external to module 500 can be secured with respect to confidentiality, integrity, or authentication.

Packet Processor

Packet processor 520 represents a logical group of capabilities to handle information packet handling and to govern some of module 500 main aspects. In preferred embodiments, packet processors contains any combination of hardware, software, or firmware necessary to assign host addresses to various objects associated with module 500, to take action on the information packets passing through network interface 510 or target device interface 530, or to process information packets at rates up to the line-rate of network interface 510 or the line-rate of target device interface 530. Packet processor 520 assigns host addresses to module 500 objects including target devices 5405A through 540M, virtual devices 528A through 528N, module 500 itself, virtual bus tunnel 524, logical groups of module objects, or other real or virtual objects associated with module 500. When packet processor 520 takes action on information packets traveling through module 500, information packets are converted among the various types of ID packets. The packets contain the necessary information for routing to other systems as well as the data for the other systems to properly process the packets. Therefore, specific goals or rules employed by module 500 govern the actions taken on the information packets. Preferred embodiments of packet processor 520 comprise the capability of communicating with other modules. Yet even more preferred embodiments of packet processor 520 comprise the capability of communicating with other modules, where the other modules are separated from module 500 by routers.

In order to facilitate communications with other modules, a preferred embodiment of packet processor 520 further comprises the ability to acquire a plurality of host addresses from internal module resources or from external modules resources. Once acquired, the host addresses can be assigned by packet processor 520 to various module objects. To further facilitate communications, preferred host addresses comprise routable addresses such that systems external to the module determine how to route the information packets containing the host addresses to a final destination. Furthermore, a more preferred version of packet processor 520 comprises a routable node with respect to the routable addresses. Such a packet processor comprises the capability of routing packets with host addresses to internal objects as well as to external objects. In addition host addresses can comprise addressable names used to reference objects, where contemplated names include names that can be used by DNS systems to resolve specific host addresses. Furthermore, for localized systems that do not have access to DNS, a local name resolution system is contemplated, preferably a peer-to-peer mechanism. More preferred host addresses comprise IPv4 or IPv6 addresses.

As a preferred packet processor 520 operates on information packets, it is contemplated that the results of taking action on the packets yield no out-going packets or at least one out-going packet. The case of no out-going packets provides for silently discarding packets not destined for any object within module 500. Contemplated uses for silently discarding include using IP multicasting to communicate with a logical group of objects providing a aggregate set of capabilities or services where information packets are sent to all objects in the group, but only one target object needs to respond and all others can silently discard the information packet. The case of at least one out-going packet provides for transforming from one type of packet to another and for routing packets to their final destination. An even more preferred embodiment of packet processor 520 provides capabilities comprising packets processing uni-directionally, multi-directionally, or for returning packets backward from whence they came. Contemplated uses for uni-directional processing include accepting information packets destined for target devices and transforming the packets to device format. The packets are then delivered directly to a target device. Contemplated uses for multi-directional processing include sending packets to a target device as well as sending a possible acknowledgement to a remote source. In addition, contemplated uses for returning backward packets include responding with an error if information packets can not be processed properly.

Packet processor 520 also comprises additional contemplated functionality including configuring target devices 540A through 540M, configuring target device interface 537, managing virtual bus tunnel 524, or managing virtual device manager 526. Packet processor 520 further comprises virtual bus tunnel 524 and virtual device manager 526. Virtual bus tunnel 524 provides a conduit for virtual devices 528A through 528N used for communication with other systems external to module 500. Virtual device manager 526 comprises sufficient software for controlling and managing virtual devices 528A through 528N. Both the virtual device manager and virtual bus tunnel will be discussed in more detail in following sections.

Target Device Interface

Target device interface 530 comprise a layer of functionality providing the logical blocks within module 500 access to target devices 540A through 540M. Contemplated target device interfaces include homogenous interfaces where all the target devices are of the same type or class, or heterogeneous interfaces where at least some of the target devices are different from the other target devices. Specifically contemplated homogenous interfaces include hard disk drive interfaces. Contemplated heterogeneous interfaces comprise various device level interfaces including DMA for low level devices, peripheral ports, APIs, or user interface ports.

Virtual Device Manager

Preferred versions of virtual device manager 526 comprise the capability for managing virtual devices 528A through 528N and functionality for mapping virtual devices to the capabilities or characteristics of target devices 540A through 540M. Within this context managing implies virtual device manger 526 has complete control over the existence or behavior of the virtual devices including the capability to create, destroy, modify, suspend, or resume operation of the virtual device.

Virtual Devices within a Module

Virtual devices 528A through 528N can be advantageously implemented through software constructs including programs, tasks, threads, processes, or monolithic coding systems. Each virtual device does not necessarily map to a target device in a one-to-one fashion. In addition, virtual devices do not necessarily have to correspond to any capability or characteristic of any of the target devices. In preferred embodiments at least one virtual device maps its functionality to a set of target devices including rotating or non-rotating media. Specifically contemplated non-rotating media includes all possible forms of memory. Specifically contemplated rotating media includes hard disk drives, optical drives, or other yet to be conceived rotating media. In the case of rotating media comprising hard disk drives, virtual devices 528A through 528N are contemplated to be logical partitions mapped to storage areas on the hard disk drive.

For modules to be deployed in existing computer systems including Windows® or Linux, preferred virtual devices are most advantageously implemented as drivers, threads, or processes. For self contained modules to be attached physically to target devices, virtual devices are advantageously implemented as tasks on embedded operating systems. In addition, it is contemplated that virtual devices can represent components other than target device capabilities including functions which can serve as remote event handlers for a decentralized computing system. Under such a contemplated virtual device, the host address associated with virtual devices becomes an event handler ID which can take the form of an IP address, port assignment, or memory address.

Module Implementation

FIG. 6 illustrates a possible physical embodiment of a module 600. Information packets from external to module 600 flow bi-directionally through network communication path 605 to network interface 610. Network interface 610 also provides a network packet path 615 from network interface 610 to CPU 620. CPU 620 executes any combination of software code or firmware core to support the logical structures within module 600. CPU 620 loads executable code instructions from flash 622. It is contemplated flash 622 serves many uses including code storage, module configuration storage, local file systems, or other uses pertaining to persistent data storage where CPU 620 uses code path 623 to write or read from flash 622. Although the word “flash” is used, one ordinarily skilled in the art of developing embedded systems will recognize any type of persistent or non-volatile memory can be used to serve the same purpose including a hard disk drive. For modules that need extreme high performance, it is contemplated CPU 620 can represent multiple actual processors. Contemplated modules with multiple processors include rack mount systems providing storage arrays. As CPU 620 executes module 600 code, system data is stored and retrieved from RAM 624 via data path 625. CPU 620 utilizes device data path 635 to interact bi-directionally with target device interface 630. Target device interface 630 also provides bi-directional paths 645A through 645M for device data for each target device 640A through 640M. Contemplated target device interfaces including backplanes for enclosures capable of supporting many target devices.

A preferred network interface 610 comprises the use of Ethernet to exchange packets. Because modules are contemplated to exist in a wide variety of networking environments, wired or wireless interfaces are contemplated. Preferred wireless interfaces include all forms of 802.11, Ultra Wide Band (UWB), WiMAX, HomeRF, HiperLAN, Bluetooth, IrDA, or other wireless networking technologies existing now or in the future. Contemplated wired interfaces include Ethernet supporting various speeds including 10 Mbps, 100 Mbps, 1 Gbps, 10 Gbps, or higher speeds. In addition, contemplated network interface 610 can comprise more than one physical channel for various reasons including to increase performance, to provide out-of-band management, to provide out-of-band module to module communication similar to an internal bus or backplane, or to provide logical separation of decentralized capabilities or services.

Preferred embodiments of target device interface 630 include any combination of homogenous or heterogeneous wired interfaces including serial or parallel interfaces. In addition, target device interface 630 comprises the capability of interfacing to one or more target devices. Yet more preferred embodiments of target device interface 630 include serial interfaces including RS-232, RS-485, RS-422, USB, Firewire, Ethernet, USB, or even programmable IO pins for low level devices. Yet even more preferred serial interfaces include SATA, or fiber channel interfaces. Contemplated parallel interfaces include SCSI, PATA, PCI, or PCI-like interfaces. Beyond wired interfaces, it is also contemplated that target device interface 630 can employ wireless interfaces including Bluetooth, IrDA, or even 802.11 wireless communications. In some embodiments, target device interface 630 can comprise API to access functionality of a larger system.

When implemented on a hardware platform for use with physical target devices 640A through 640M, module 600 includes a number of other components in order to support the overall functionality of the module. Those ordinarily skilled in the art of embedded systems or software development will recognize the standard components usually employed when creating such modules. The components include an embedded real-time operating system, TCP/IP stack, file system, web server for use as a user interface, or other middleware. Security for confidentiality, integrity, or authentication can be handled through the use of protocols including IPSec, VPNs, SSL, SSH, RADIUS, Kerberos, or other security systems as they gain currency. Additional hardware components might also be necessary, including clocks, internal buses, transceivers, or others. When module 600 comprises all software for computing systems including Windows® or Linux, the module will interface to elements within the computing system via a set of APIs, where the hardware associated with system is part of a generalized computing platform. Specifically contemplated implementations of modules include programmable gate arrays, ASICs, embedded boards that attach directly to target devices, back planes, rack-mount systems, or even software components.

Virtual Bus Tunnel

FIG. 7 illustrates a comparison of a traditional computing system in FIG. 7A and a decentralized computing system utilizing modules comprising virtual devices in FIG. 7B. Referring to FIG. 7A, computing system 700 comprises application 703 which requires a generalized operating system 705 including Windows® or Linux to function properly. When application 703 accesses system resources including hard drives, monitors, memory, network, or other localized devices the application makes API calls through operating system 705 to driver 710. Driver 710 then accesses a localized system bus to send to or receive information packets with device format from any of devices 750A through 750D.

FIG. 7B depicts a similar computing system 701 with the exception that module 712 replaces driver 710. Module 712 provides access to network 720 used for the virtual bus tunnel. As application 703 requests support from devices, application 703 again calls through operating system 705 which then makes device level requests through module 712. Module 712 transforms the packets with a device format from operating system 705 to a packet with a split-ID format or a contiguous-ID format destined for a remote device connected to module 740A or 740B. Modules 712, 740A, and 740B, utilize a preferred embodiment of a virtual bus tunnel which comprises the functionality that allows the objects within the modules to communicate with each other via information packets. Using module 712 allows the operating system to perceive devices 750A through 750D as if they are locally connected. Moreover, device 750A through 750D can appear as a single device if they have been aggregated. In addition, modules 740A and 740B allow devices 750A through 750D to perceive they connect directly to a real bus.

Network 720 comprises the virtual bus tunnel within modules 712, 740A, and 740B. Each virtual bus tunnel provides an addressing mechanism and a data transport mechanism. The addressing mechanism comprises a number of capabilities: the capability of addressing logical groups of target devices or virtual devices such that they can be addressed by a single point address, the capability of addressing individual target devices or virtual devices, or the ability to address specific characteristics associated with target devices or virtual devices. The virtual bus tunnel does not necessarily tunnel device data because device formats from one module to another can vary widely. Group level addressing allows multiple logically separate systems to utilize the same physical transport media as a virtual bus. For example, if a target device happens to be a CPU, multiple users access the CPU via different group IDs, where each group ID represents one user's decentralized computing system. The data transport mechanism comprises that ability to transfer and route information packets from system to system until the information packets reach their final destination. A preferred virtual bus tunnel comprises an internetworking protocol which provides addressing capabilities and data transport capabilities. Specific contemplated internetworking protocols include IPv4 or IPv6. Even more preferred virtual bus tunnels comprise logical groups of objects within modules and optionally objects from other modules addressed via multicast IP addresses.

IP Multicasting—Mirrors, Stripes, and Spans

FIG. 8 provides an illustration showing a possible embodiment where a user operating computer 800 could access data across a number of decentralized hard disk drives but have their capabilities or services aggregated via modules. Drives 832, 834, and 836 attach to module 830, drive 842 attaches to module 840, and drives 852 and 854 attach to module 850. Each of the drives further comprises logical partitions located on each of the drives: drive 832 comprises logical partition 833, drive 834 comprises logical partition 835, drive 836 comprises logical partition 837, drive 842 comprises logical partition 843, drive 852 comprises logical partition 853, and drive 854 comprises logical partition 855. Collectively, modules 830, 840, and 850 comprise virtual devices that aggregate the logical partitions to form multicast group 820. From the perspective of the user or application 802, the aggregated logical partitions represented by multicast group 820 form a single disk volume that appears as a single hard disk drive. As a user operates application 802 on computer 800, the application can request data from a hard disk drive through operating system 804. Operating system 804 perceives the data as residing on a single hard disk drive device and thereby requests data though module 806. Data requests in the form of information packets comprising split-ID or contiguous-ID format travel across a network 81 SA through 81 SD which can optionally including a number of routers such as router 810 or even can optionally include Internet 825. Data requests are received by modules 830, 840, and 850. The modules 830, 840, and 850 determine if the requests need action based on the target hard disk drives attached to the modules, the host address in the information packet, and LBA within the packet. Although FIG. 8 presents a possible configuration of modules and target devices, one ordinarily skilled in the art will recognize that any possible configuration of computers, network, modules, or target devices are possible while keeping within the scope of the presented subject matter.

Within the example presented in FIG. 8, modules 830, 840, and 850 provide host addresses for all objects associated with the modules including the target hard disk drives as well as logical partitions associated with each of the hard disk drives. Within this context, the logical partitions represent virtual devices. Multicast group 820 represents a possible virtual bus tunnel for the logical partitions such that the logical partitions, modules, or applications coordinate their behaviors. It is specifically contemplated that logical groups of logical partitions on storage devices form multicast mirrors, stripes, or spans. Therefore a logical group is one or more logical partitions. One or more logical groups can then combine to form a logical volume that appears as a single disk to a computer system.

Although a preferred embodiment of multicast group 820 utilizes IP multicasting to address logical groups of virtual devices, alternative embodiments are possible. For example, hosts can employ broadcasts on a subnet to address a logical group wherein the broadcast packets include a group ID representing the logical group. Members in the group use the group ID to determine if the packet should be accepted. Even though this document contemplates the use of IP multicasting to provide group addressing, all possible alternative group addressing based on a group IDs are also contemplated, specifically contemplated group addressing includes the ability to route across subnets. Therefore, “multicast” includes the concept of alternative logical group addressing.

Multicast Spans

A logical group of logical partitions form a multicast span similar to that represented by multicast group 820. Modules, virtual devices, or target devices join a multicast group represented by a single IP address, or group ID, where virtual devices are logical partitions on target hard disk drives. The single IP address represents a group ID that identifies a single logical volume. As an example, a single logical group can form a single spanned logical volume, as called a multicast stripe, identified by the group ID. Data spans across the logical partitions spread across plurality of drives such that the total available storage capacity can be extended beyond that provided by a single drive. In a multicast span, LBAs are distributed sequential from 0 to a maximum number. Each logical partition is responsible for a range of LBAs and each subsequent logical partition continues the LBA numbering sequence where the previous logical partitions left off. Logical partitions can cover an entire drive; however, logical partitions in a span can be of any size.

When a host wishes to send or retrieve data from the multicast stripe, the host sends a request to the multicast group via packets with a split-ID or contiguous-ID format containing the IP address of the group as well as the LBA associated with the data. All virtual devices, or logical partitions, within the group receive the information packet via their modules. The logical partitions determine if they are responsible for that particular data block as determined by the LBA. If a logical partition is not responsible for the data block, the information packet is silently discarded. If a logical partition is responsible for the data block, the information packet format will be transformed. On a write, the packet format will be changed to device format and the data will be written to a target disk. On a read, an information packet will be read from a target disk and transformed to split-ID or contiguous-ID format then sent back to the originating host. A response information packet can contain data, an error, or an acknowledgement. The host expects one response to its request to a multicast stripe.

Multicast Stripes

Multicast stripes are similar to multicast spans with the exception that a number of logical partitions spread across numerous drives to provide increased performance when reading data from the drives. In this sense, multicast stripes are a super-set of multicast spans where the partitions can be any size, not just the size of the hard disk drive.

Referring to FIG. 8, the logical group of logical partitions represented by multicast group 820 forms a striped logical volume, also called a multicast stripe. The logical partitions within the multicast stripe respond identically to host requests as the logical partitions within a multicast span. However, LBAs within a multicast stripe are distributed differently than in a multicast span. In a multicast stripe, LBAs are distributed in blocks laterally across the logical partitions such that the logical volume has a complete set of LBAs from 0 to a maximum value, but each logical partition does not have sequential LBAs. For example in a three partition multicast stripe comprising logical partition 833, 835, and 837 with a LBAs 0 through 29, logical partition 833 could be responsible for LBAs 0 through 4 and LBAs 15 through 19. Logical partition 835 could be responsible for LBAs 5 through 9 and LBAs 20 through 24. Finally, logical partition 837 could be responsible for LBAs 10 through 14 and LBAs 25 through 29. As data is written to the multicast stripe, the first blocks of data are written to partition 833, and then the second blocks of data are written to 835, and so on. Once blocks of data have reached logical partition 837, the pattern repeats back at logical partition 833. As in multicast spans, a virtual device representing a logical partition uses the LBA within the split-ID or contiguous-ID to determine where the data should reside. The previous example is to provide an illustration of how data can be written to multicast stripe and should not be considered restrictive to the presented subject matter. The depth of the data stripes, the number logical partitions, or placement of the physical disks are arbitrary.

Multicast Mirrors

Multicast mirrors can be constructed in a similar fashion as multicast spans and multicast stripes with the exception that there can be multiple logical groups that duplicate data. Consequently, a mirrored logical volume, also called a multicast mirror, comprises more than one set of LBAs numbered from 0 to a maximum value. Multicast group 820 can represent a multicast mirrored volume where some logical partitions mirror data from other logical partition. When a host sends information packets to the multicast mirror, more than one logical partition can be responsible for the data's LBA. Consequently, the host expects more than one response from the multicast mirror. Therefore, the main differences between a multicast mirror and multicast stripe or span is that the originating host can expect more than one response to a request.

It is contemplated that multicast mirrors, stripes, or spans can all coexist with each other where each can have their own multicast IP address allowing them to provide services to numerous remote hosts. It is also contemplated that the partitions that make up multicast mirrors, stripes, or span communicate with other partitions within other modules. Such communication allows for each type of volume to work together to provide enhanced abilities including data redundancy.

RAID Structures

In especially preferred embodiments, multicast spans, stripes, or mirror form RAID-like structures. Multicast stripes represent RAID level 0 where data is striped across numerous physical drives. Multicast mirrors represent RAID level 1 where data is mirrored from a primary source to a mirrored volume. By combining multicast mirrors, stripes, or spans along with other RAID concepts including parity, more complex RAID structures are possible, RAID-5 for example. One ordinarily skilled in the art of RAID system is able to build such systems.

Virtual Bus Tunnel Method

FIG. 9 depicts a possible method providing a virtual bus tunnel for virtual devices or target devices via modules.

Beginning at step 900, a module associates virtual devices with target device and their capabilities or characteristics. Virtual devices can be logical constructs that can be governed by software. Once the virtual devices are instantiated, whether in software or hardware, they are ready to begin performing their tasks and to communicate with other virtual devices. For example, in a distributed storage system logical partitions are associated with physical disk partitions. Once established, the logical partitions are able to participate with other virtual devices over a virtual bus tunnel.

At step 905 a module acquires multiple host addresses. The acquisition can be performed in a number of different possible ways. One contemplated method for acquiring host addresses includes using external module mechanisms including DHCP requests to gain a number of IP addresses from a DHCP server. Additional contemplated methods for acquiring IDs includes using internal module mechanisms including Auto-IP or having the host addresses pre-determined via initial configuration at manufacturing time or subsequent configuration during deployment. Host addresses combine with other ID information to eventually form split-IDs or contiguous-IDs.

At step 910 a module can assign the acquired host addresses to various objects associated with the module. Contemplated objects include real or virtual objects. Real objects include the module itself, interfaces, target devices, chassis, power supplies, or other physical devices that can be referenced by an external host. Virtual objects include virtual devices within the module, virtual bus tunnel, multicast mirrors, multicast stripes, multicast spans, logical volumes, logical groups, logical partitions, or other virtual systems providing capabilities or services to remote hosts.

At step 920 a module waits for information packets to arrive from any source including a target device interface, a network interface, virtual devices within the module, or other sources internal or external to the module. Contemplated information packets include packets with split-ID format, contiguous-ID format, or device format, where preferred embodiments use host addresses that comprise IP addresses.

At step 935 a module determines if an information packet has device format. If the packet does have device format, the module takes action on the packet at step 932 to transform the packet to use split-ID or contiguous-ID format. The transformation can be governed by a virtual device's functionality. Once the transform is complete, the packet is sent over a virtual bus tunnel at step 934. A preferred virtual bus tunnel in step 934 groups virtual device or target devices via multicasting groups to provide a single point contact for communication. Furthermore, a preferred media comprises Ethernet based media to transport packets over a network. Once the packet has been sent at step 934, the module again receives packets at 920.

If step 935 determines the information packet does not have device format, then at step 945 a module determines if the packet is an ID packet comprising a split-ID or contiguous-ID. If the packet is an ID packet, then step 955 determines if the packet is destined to travel to a target device. If so, the packet is transformed to a packet with device format at step 952 where all ID information and packet contents help determined the final transformation. At step 954, the final packet is sent to the target device. If at step 955 the packet is not destined for a target device, then the module determines if the packet is silently discarded or is transformed at step 957. If step 957 determines the packet must be silently discarded, as in the case of multicasting mirroring, striping, or spanning, then the packet is discarded and the module again waits to receive new packets back at step 920. Otherwise, if step 957 determines the packet must not be silently discarded, the packet is further transformed to a different ID format 956 and sent on via step 934. A contemplated use for this type of transform includes a virtual device representing a multicast stripe that can copy data directly to a multicast mirror without having a source host perform multiple operations to create data redundancy.

If at step 945 the received packet is not an information packet from the device or has a split-ID or contiguous-ID, the module processes 946 the packet in a manner that is determined by the programming of the module. Contemplated additional processing includes handling configuration information for target devices, managing virtual devices, updating firmware, or other support processing.

One ordinarily skilled in the art of developing software or embedded systems will recognize the above outlined method and module can be implemented in many different ways in order to optimize various aspects of a module. Implementations include providing a single monolithic processing image to ensure very fast processing or providing multiple threads or tasks that replicate the method presented above for each target device or for each virtual device to simplify the implementation.

Example Use—Data Storage

Specific data storage examples have already been presented; however, a more focused look at employing modules within data storage will help clarify a possible embodiment of the presented material.

Data storage products can be advantageously developed based on modules and hard disk drives. A possible goal is to create a network storage system. Users and applications can access the data store where data is spread over multiple hard disks located remotely from the user over a network in a manner that appears to be a single drive.

At least two types of modules are contemplated. One type attaches directly to the set of hard disk drives, and establishes a set of logical partitions all addressable via IP addresses. The logical partitions then can form a multicast group that presents the collection as a single disk volume to remote hosts. The second type is a software driver installed on a computer system. The computer's module presents its target device interface to the computer's file system. In addition, it is also is able to communicate to the multicast group. Under this case, split-IDs comprise IP addresses and LBAs. The virtual bus tunnel comprises the use UDP datagrams over an IP based Ethernet network. If the modules are constructed properly, then the system can process data at line-rates with respect to the network or the disk drive interfaces.

Consumer oriented systems can be advantageously developed where one or two drives have their capabilities aggregated. A consumer installs a module drive on their PC that can connect to the modules for the drives. Most consumer NAS products suffer from performance problems; however, a module based version allows for perform at line-rates. For enterprise level system, many drives can be combined within a rack-mount enclosure where all drives can be combined in numerous ways to enhance performance, reliability, or scalability. Because the module is virtualized, a user will have unprecedented control over their storage configuration.

The advantage this approach has over traditional NAS and SAN approaches are several fold. First, the module approach utilizes existing network infrastructure and does not need additional specialized controllers or equipment based on fiber channel or SCSI. Second, network storage systems can be developed on less expensive SATA drives reducing the final costs to an end user. Third, by applying various arrangements of multicast mirrors, stripes, and span data redundancy and reliability can be increased over existing RAID systems due to automatic data redundancy where stripes and mirrors can communicate directly with each other. Finally, because the system utilizes a virtual bus tunnel capable of transporting data anywhere in the network, a user has access to their complete computing experience.

Example Use—Medial Environments

In order to illustrate the broad applicability of virtual devices and virtual bus tunnels an additional example is provided beyond data storage. In the medical environment, a patient can be monitored through the use of virtual bus tunnels created via modules. Under such circumstances modules can take on many forms including adapters to specific equipment, embedded modules within medical devices, or even software database adapters.

Patient monitoring equipment can include blood analyzers, sphygmomanometers, fetal monitors, or other devices attached to patients. Modules can combine to form a multicast group whose address represents a patient ID. Consequently, when any application, doctor, or nurse requires patient data the request is sent to the multicast group and only the relevant device responds. As wireless sensors are deployed in a hospital environment, the need will grow for patient-centric monitoring rather than room-centric monitoring. A module based system has the advantage of allowing all sensors to move with the patient and remain accessible rather than being accessible only in a room. Additionally, modules can take the form of database adapters implemented completely in software where the database adapter module continually updates databases as patient telemetry is received.

As shown with the above examples, the decentralization and virtualization of computing resources offers the ability for users to maintain existing experiences no matter the physical locality of those resources. Although the examples show single applications, all possible extensions to the applicability of modules are contemplated including, but not limited to, the use of modules to provide fully decentralized computing systems where CPUs, drives, memory, or other components are separated from each other. Such a virtual computing system allows a user to access their computer experience from anywhere in the world.

Thus, specific compositions and methods virtual device communications, information packets, and virtual bus tunnels have been disclosed. It should be apparent, however, to those skilled in the art that many more modifications besides those already described are possible without departing from the inventive concepts herein. The inventive subject matter, therefore, is not to be restricted except in the spirit of the disclosure. Moreover, in interpreting the disclosure all terms should be interpreted in the broadest possible manner consistent with the context. In particular the terms “comprises” and “comprising” should be interpreted as referring to the elements, components, or steps in a non-exclusive manner, indicating that the referenced elements, components, or steps can be present, or utilized, or combined with other elements, components, or steps that are not expressly referenced.

Patent Citations
Cited PatentFiling datePublication dateApplicantTitle
US4890227Aug 25, 1987Dec 26, 1989Hitachi, Ltd.Autonomous resource management system with recorded evaluations of system performance with scheduler control including knowledge learning function
US5129088Jul 3, 1990Jul 7, 1992International Business Machines CorporationData processing method to create virtual disks from non-contiguous groups of logically contiguous addressable blocks of direct access storage device
US5193171Dec 10, 1990Mar 9, 1993Hitachi, Ltd.Method of managing space of peripheral storages and apparatus for the same
US5506969Nov 29, 1993Apr 9, 1996Sun Microsystems, Inc.Method and apparatus for bus bandwidth management
US5546541Feb 5, 1991Aug 13, 1996International Business Machines CorporationComputer implemented method
US5590124Jun 7, 1995Dec 31, 1996Apple Computer, Inc.Link and discovery protocol for a ring interconnect architecture
US5590276Mar 15, 1995Dec 31, 1996Emc CorporationMethod for synchronizing reserved areas in a redundant storage array
US5634111Mar 16, 1993May 27, 1997Hitachi, Ltd.Computer system including a device with a plurality of identifiers
US5742604Mar 28, 1996Apr 21, 1998Cisco Systems, Inc.Interswitch link mechanism for connecting high-performance network switches
US5758050Mar 12, 1996May 26, 1998International Business Machines CorporationReconfigurable data storage system
US5758188Nov 21, 1995May 26, 1998Quantum CorporationSynchronous DMA burst transfer protocol having the peripheral device toggle the strobe signal such that data is latched using both edges of the strobe signal
US5771354 *Nov 4, 1993Jun 23, 1998Crawford; Christopher M.Internet online backup system provides remote storage for customers using IDs and passwords which were interactively established when signing up for backup services
US5850449 *May 19, 1997Dec 15, 1998Sun Microsystems, Inc.Secure network protocol system and method
US5867686Nov 8, 1995Feb 2, 1999Conner; Kenneth H.High speed real-time information storage system
US5884038May 2, 1997Mar 16, 1999Whowhere? Inc.Method for providing an Internet protocol address with a domain name server
US5889935Mar 17, 1997Mar 30, 1999Emc CorporationDisaster control features for remote data mirroring
US5930786Oct 20, 1995Jul 27, 1999Ncr CorporationMethod and apparatus for providing shared data to a requesting client
US5937169Oct 29, 1997Aug 10, 19993Com CorporationOffload of TCP segmentation to a smart adapter
US5949977Oct 8, 1996Sep 7, 1999Aubeta Technology, LlcMethod and apparatus for requesting and processing services from a plurality of nodes connected via common communication links
US5991891Dec 23, 1996Nov 23, 1999Lsi Logic CorporationMethod and apparatus for providing loop coherency
US6018779Dec 15, 1997Jan 25, 2000Emc CorporationSystem for encapsulating a plurality of selected commands within a single command and transmitting the single command to a remote device over a communication link therewith
US6081879Nov 4, 1997Jun 27, 2000Adaptec, Inc.Data processing system and virtual partitioning method for creating logical multi-level units of online storage
US6101559Oct 22, 1997Aug 8, 2000Compaq Computer CorporationSystem for identifying the physical location of one or more peripheral devices by selecting icons on a display representing the one or more peripheral devices
US6105122Feb 6, 1998Aug 15, 2000Ncr CorporationI/O protocol for highly configurable multi-node processing system
US6128664Mar 24, 1998Oct 3, 2000Fujitsu LimitedAddress-translating connection device
US6157935Dec 17, 1996Dec 5, 2000Tran; Bao Q.Remote data access and management system
US6157955Jun 15, 1998Dec 5, 2000Intel CorporationPacket processing system including a policy engine having a classification unit
US6202060Oct 29, 1996Mar 13, 2001Bao Q. TranData management system
US6246683May 1, 1998Jun 12, 20013Com CorporationReceive processing with network protocol bypass
US6253273Feb 6, 1998Jun 26, 2001Emc CorporationLock mechanism
US6259448 *Jun 3, 1998Jul 10, 2001International Business Machines CorporationResource model configuration and deployment in a distributed computer network
US6275898May 13, 1999Aug 14, 2001Lsi Logic CorporationMethods and structure for RAID level migration within a logical unit
US6288716Jun 24, 1998Sep 11, 2001Samsung Electronics, Co., LtdBrowser based command and control home network
US6295584Aug 29, 1997Sep 25, 2001International Business Machines CorporationMultiprocessor computer system with memory map translation
US6330236Jul 22, 1998Dec 11, 2001Synchrodyne Networks, Inc.Packet switching method with time-based routing
US6330615Sep 14, 1998Dec 11, 2001International Business Machines CorporationMethod of using address resolution protocol for constructing data frame formats for multiple partitions host network interface communications
US6385638Sep 4, 1997May 7, 2002Equator Technologies, Inc.Processor resource distributor and method
US6396480Jul 17, 1995May 28, 2002Gateway, Inc.Context sensitive remote control groups
US6401183Apr 1, 1999Jun 4, 2002Flash Vos, Inc.System and method for operating system independent storage management
US6434683Nov 7, 2000Aug 13, 2002Storage Technology CorporationMethod and system for transferring delta difference data to a storage device
US6449607Sep 10, 1999Sep 10, 2002Hitachi, Ltd.Disk storage with modifiable data management function
US6466571Jan 19, 1999Oct 15, 20023Com CorporationRadius-based mobile internet protocol (IP) address-to-mobile identification number mapping for wireless communication
US6473774Sep 28, 1998Oct 29, 2002Compaq Computer CorporationMethod and apparatus for record addressing in partitioned files
US6480934Sep 28, 1999Nov 12, 2002Hitachi, Ltd.Storage control unit and method for handling data storage system using thereof
US6487555May 7, 1999Nov 26, 2002Alta Vista CompanyMethod and apparatus for finding mirrored hosts by analyzing connectivity and IP addresses
US6549983May 20, 1999Apr 15, 2003Samsung Electronics Co., Ltd.Cache memory system and method for managing the same
US6567863Dec 7, 1999May 20, 2003Schneider Electric Industries SaProgrammable controller coupler
US6601101Mar 15, 2000Jul 29, 20033Com CorporationTransparent access to network attached devices
US6601135Nov 16, 2000Jul 29, 2003International Business Machines CorporationNo-integrity logical volume management method and system
US6618743Oct 9, 1998Sep 9, 2003Oneworld Internetworking, Inc.Method and system for providing discrete user cells in a UNIX-based environment
US6629178 *Jun 15, 2000Sep 30, 2003Advanced Micro Devices, Inc.System and method for controlling bus access for bus agents having varying priorities
US6629264Mar 30, 2000Sep 30, 2003Hewlett-Packard Development Company, L.P.Controller-based remote copy system with logical unit grouping
US6678241 *Nov 30, 1999Jan 13, 2004Cisc Technology, Inc.Fast convergence with topology switching
US6681244Jun 9, 2000Jan 20, 20043Com CorporationSystem and method for operating a network adapter when an associated network computing system is in a low-power state
US6693912Apr 4, 2000Feb 17, 2004Oki Electric Industry Co., Ltd.Network interconnecting apparatus and active quality-of-service mapping method
US6701431Jan 29, 2001Mar 2, 2004Infineon Technologies AgMethod of generating a configuration for a configurable spread spectrum communication device
US6701432Apr 1, 1999Mar 2, 2004Netscreen Technologies, Inc.Firewall including local bus
US6711164Nov 5, 1999Mar 23, 2004Nokia CorporationMethod and apparatus for performing IP-ID regeneration to improve header compression efficiency
US6732171May 31, 2002May 4, 2004Lefthand Networks, Inc.Distributed network storage system with virtualization
US6732230Oct 20, 1999May 4, 2004Lsi Logic CorporationMethod of automatically migrating information from a source to an assemblage of structured data carriers and associated system and assemblage of data carriers
US6741554Aug 16, 2002May 25, 2004Motorola Inc.Method and apparatus for reliably communicating information packets in a wireless communication network
US6754662Dec 20, 2000Jun 22, 2004Nortel Networks LimitedMethod and apparatus for fast and consistent packet classification via efficient hash-caching
US6757845Nov 30, 2000Jun 29, 2004Bitmicro Networks, Inc.Method and apparatus for testing a storage device
US6772161Dec 19, 2001Aug 3, 2004Hewlett-Packard Development Company, L.P.Object-level migration in a partition-based distributed file system
US6775672Dec 19, 2001Aug 10, 2004Hewlett-Packard Development Company, L.P.Updating references to a migrated object in a partition-based distributed file system
US6775673Dec 19, 2001Aug 10, 2004Hewlett-Packard Development Company, L.P.Logical volume-level migration in a partition-based distributed file system
US6795534Sep 4, 2001Sep 21, 2004Nec CorporationData recording system for IP telephone communication
US6799244Dec 6, 2002Sep 28, 2004Hitachi, Ltd.Storage control unit with a volatile cache and a non-volatile backup cache for processing read and write requests
US6799255Feb 2, 2001Sep 28, 2004Emc CorporationStorage mapping and partitioning among multiple host processors
US6834326Mar 17, 2000Dec 21, 20043Com CorporationRAID method and device with network protocol between controller and storage devices
US6853382Oct 13, 2000Feb 8, 2005Nvidia CorporationController for a memory system having multiple partitions
US6854021Oct 2, 2000Feb 8, 2005International Business Machines CorporationCommunications between partitions within a logically partitioned computer
US6862606Jun 7, 2001Mar 1, 2005Novell, Inc.System and method for partitioning address space in a proxy cache server cluster
US6876657Dec 14, 2000Apr 5, 2005Chiaro Networks, Ltd.System and method for router packet control and ordering
US6886035May 13, 2002Apr 26, 2005Hewlett-Packard Development Company, L.P.Dynamic load balancing of a network of client and server computer
US6894976Jun 15, 2000May 17, 2005Network Appliance, Inc.Prevention and detection of IP identification wraparound errors
US6895461Apr 22, 2002May 17, 2005Cisco Technology, Inc.Method and apparatus for accessing remote storage using SCSI and an IP network
US6895511May 19, 1999May 17, 2005Nortel Networks LimitedMethod and apparatus providing for internet protocol address authentication
US6901497Oct 26, 2001May 31, 2005Sony Computer Entertainment Inc.Partition creating method and deleting method
US6904470Mar 26, 2003Jun 7, 2005Emc CorporationDevice selection by a disk adapter scheduler
US6907473Mar 31, 2003Jun 14, 2005Science Applications International Corp.Agile network protocol for secure communications with assured system availability
US6912622Apr 15, 2002Jun 28, 2005Microsoft CorporationMulti-level cache architecture and cache management method for peer-to-peer name resolution protocol
US6917616Sep 17, 1999Jul 12, 2005Alcatel Canada Inc.Method and apparatus for reduction and restoration of data elements pertaining to transmitted data packets in a communications network
US6922688Mar 3, 1999Jul 26, 2005Adaptec, Inc.Computer system storage
US6928473Sep 26, 2000Aug 9, 2005Microsoft CorporationMeasuring network jitter on application packet flows
US6941555 *May 1, 2003Sep 6, 2005Bea Systems, Inc.Clustered enterprise Java™ in a secure distributed processing system
US6947430Mar 16, 2001Sep 20, 2005International Business Machines CorporationNetwork adapter with embedded deep packet processing
US6985956Nov 2, 2001Jan 10, 2006Sun Microsystems, Inc.Switching system
US7039934Dec 7, 2000May 2, 2006Sony CorporationRecording system
US7051087Jun 5, 2000May 23, 2006Microsoft CorporationSystem and method for automatic detection and configuration of network parameters
US7065579Jan 22, 2002Jun 20, 2006Sun Microsystems, Inc.System using peer discovery and peer membership protocols for accessing peer-to-peer platform resources on a network
US7069295Feb 14, 2002Jun 27, 2006The Escher Group, Ltd.Peer-to-peer enterprise storage
US7073090Aug 20, 2002Jul 4, 2006Emc CorporationRemote data mirroring system having a remote link adapter
US7111303 *Jul 16, 2002Sep 19, 2006International Business Machines CorporationVirtual machine operating system LAN
US7120666Oct 30, 2002Oct 10, 2006Riverbed Technology, Inc.Transaction accelerator for client-server communication systems
US7145866 *Jun 1, 2001Dec 5, 2006Emc CorporationVirtual network devices
US7149769Mar 26, 2002Dec 12, 2006Hewlett-Packard Development Company, L.P.System and method for multi-destination merge in a storage area network
US7152069Oct 15, 2002Dec 19, 2006Network Appliance, Inc.Zero copy writes through use of mbufs
US7184424Oct 4, 2005Feb 27, 2007Zetera CorporationMultiplexing storage element interface
US7188194Apr 22, 2002Mar 6, 2007Cisco Technology, Inc.Session-based target/LUN mapping for a storage area network and associated method
US7203730Feb 13, 2002Apr 10, 2007Network Appliance, Inc.Method and apparatus for identifying storage devices
US7206805Nov 17, 2000Apr 17, 2007Oracle International CorporationAsynchronous transcription object management system
US7237036Sep 27, 2002Jun 26, 2007Alacritech, Inc.Fast-path apparatus for receiving data corresponding a TCP connection
US7260638Jul 23, 2001Aug 21, 2007Bluesocket, Inc.Method and system for enabling seamless roaming in a wireless network
US7263108Aug 6, 2002Aug 28, 2007Netxen, Inc.Dual-mode network storage systems and methods
US7278142 *May 17, 2001Oct 2, 2007Veritas Operating CorporationDynamic computing environment using remotely allocable resources
US7296050Jan 17, 2003Nov 13, 2007Hewlett-Packard Development Company L.P.Distributed computing system and method
US7353266Oct 24, 2005Apr 1, 2008Microsoft CorporationSystem and method for managing states and user context over stateless protocols
US20050198371 *Feb 19, 2004Sep 8, 2005Smith Michael R.Interface bundles in virtual network devices
WO2004025477A2 *Sep 15, 2003Mar 25, 2004David Julian ClarkeNetwork interface and protocol supporting mappings of virtual memory locations at different processing devices
Non-Patent Citations
Reference
1"Computer Networking Essentials" Copyright 2001, Cisco Systems, Inc., 2001.
2"Limited distributed DASD Checksum, a RAID Hybrid" IBM Technical Disclosure Bulletin, vol. 35, No. 4a, Sep. 1992, pp. 404-405, XP000314813 Armonk, NY, USA.
3B. Quinn et al. IP Multicast Applications: Challenges and Solutions. Sep. 2001. Network Working Group, RFC 3170.
4Bruschi and Rosti, "Secure multicast in wireless networks of mobile hosts: protocols and issues", Mobile Networks and Applications, vol. 7, issue 6 (Dec. 2002), pp. 503-511.
5Chavez, A Multi-Agent System for Distributed Resource Allocation, MIT Media Lab, XP-002092534.
6International Search Report for Application No. PCT/US02/40205 dated May 27, 2004.
7Ki-Il Kim et al., "Internet Multicast Provisioning Issues for Hierarchical Architecture", 2001, IEEE.
8Kim et al., "Internet multicast provisioning issues for hierarchical architecture", Networks, 2001, Proceedings, Ninth IEEE International Conference, published Oct. 12, 2001, pp. 401- 404.
9Lee and Thekkath, "Petal: Distributed Virtual Disks", Systems Research Center.
10Lee et al. "A Comparison of Two Distributed Disk Systems" Digital Systems Research Center-Research Report SRC-155, Apr. 30, 1998, XP002368118.
11Lee et al. "Petal: Distributed Virtual Disks", 7th International Conference on Architectural Support for Programming Languages and Operation Systems. Cambridge, MA., Oct. 1-5, 1996. International Conference on Architectural Support for Programming Languages and Operation Systems (ASPLOS), New, vol. Conf. 7, Oct. 1, 1996, pp. 84-92, XP000681711, ISBN: 0-89791-767-7.
12Lin JC and Paul S, "Rmtp: a reliable multicast transport protocol," Proceedings of IEEE Infocom '96, vol. 3, pp. 1414-1424, 1996.
13PCT International Search Report for PCT App. No. PCTUS05/01542 dated Aug. 25, 2008.
14Satran et al. "Internet Small Computer Systems Interface (iSCSI)" IETF Standard, Internet Engineering Task Force, IETF, Ch, Apr. 2004, XP015009500, ISSN: 000-0003.
15Satran et al., iSCSI, Internet Draft draft-ietf-ips-iscsi-19.txt.
16Thomas E. Anderson, Michael D. Dahlin, Jeanna M. Neefe, David A. Patterson, Drew S. Roselli, and Randolph Y. Wang, Serverless network file systems. Dec. 1995. In Proceedings of the 15th Symposium on Operating Systems Principles.
17VMWare Workstations User's Manual, Version 3.2, VMWare, Inc., Copyright 1998-2002.
Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US7752294 *Oct 28, 2003Jul 6, 2010Netapp, Inc.Method and system for dynamic expansion and contraction of nodes in a storage area network
US8554956Feb 24, 2010Oct 8, 2013Open Invention Network LlcMethod for creation of a device driver for a peripheral device
US8554957Feb 24, 2010Oct 8, 2013Open Invention Network, LlcMethod for creation of device drivers and device objects for peripheral devices
US20100318579 *Aug 18, 2010Dec 16, 2010Hitachi, Ltd.Control Device for Storage System Capable of Acting as a Constituent Element of Virtualization Storage System
Classifications
U.S. Classification726/15, 719/324, 707/999.202
International ClassificationG06F17/00, G06F15/16, G06F9/00
Cooperative ClassificationY10S707/99953, H04L67/1097, H04L61/00, H04L29/12009
European ClassificationH04L61/00, H04L29/12A
Legal Events
DateCodeEventDescription
Mar 18, 2013FPAYFee payment
Year of fee payment: 4
Oct 26, 2010CCCertificate of correction
Feb 23, 2010ASAssignment
Owner name: ZETERA CORPORATION, DELAWARE CORPORATION,CALIFORNI
Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE INVENTOR S NAME FROM WILLIAM R. BABBIT TO WILLIAM R. BABBITT, JR. PREVIOUSLY RECORDED ON REEL 020735 FRAME 0326. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNOR HEREBY CONFIRMS THE CORRECT SPELLING OF INVENTOR NAME IS WILLIAM R. BABBITT, JR;ASSIGNOR:BABBITT, WILLIAM R., JR.;US-ASSIGNMENT DATABASE UPDATED:20100224;REEL/FRAME:23973/443
Effective date: 20080222
Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE INVENTOR S NAME FROM WILLIAM R. BABBIT TO WILLIAM R. BABBITT, JR. PREVIOUSLY RECORDED ON REEL 020735 FRAME 0326. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNOR HEREBY CONFIRMS THE CORRECT SPELLING OF INVENTOR NAME IS WILLIAM R. BABBITT, JR;ASSIGNOR:BABBITT, WILLIAM R., JR.;REEL/FRAME:023973/0443
Apr 29, 2008ASAssignment
Owner name: RATEZE REMOTE MGMT. L.L.C., DELAWARE
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZETERA CORPORATION;REEL/FRAME:020866/0888
Effective date: 20080415
Owner name: RATEZE REMOTE MGMT. L.L.C.,DELAWARE
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100330;REEL/FRAME:20866/888
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100413;REEL/FRAME:20866/888
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100420;REEL/FRAME:20866/888
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100518;REEL/FRAME:20866/888
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZETERA CORPORATION;REEL/FRAME:20866/888
Apr 18, 2008ASAssignment
Owner name: ZETERA CORPORATION, CALIFORNIA
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;REEL/FRAME:020824/0376
Effective date: 20080418
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;REEL/FRAME:020824/0215
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:THE FRANK REVOCABLE LIVING TRUST OF CHARLES W. FRANK AND KAREN L.FRANK;REEL/FRAME:020823/0949
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WARBURG PINCUS PRIVATE EQUITY VIII, L.P.;REEL/FRAME:020824/0074
Owner name: ZETERA CORPORATION,CALIFORNIA
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;US-ASSIGNMENT DATABASEUPDATED:20100413;REEL/FRAME:20824/215
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;US-ASSIGNMENT DATABASEUPDATED:20100330;REEL/FRAME:20824/215
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;US-ASSIGNMENT DATABASEUPDATED:20100330;REEL/FRAME:20824/376
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;US-ASSIGNMENT DATABASEUPDATED:20100413;REEL/FRAME:20824/376
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:THE FRANK REVOCABLE LIVING TRUST OF CHARLES W. FRANK AND KAREN L.FRANK;US-ASSIGNMENT DATABASE UPDATED:20100330;REEL/FRAME:20823/949
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:THE FRANK REVOCABLE LIVING TRUST OF CHARLES W. FRANK AND KAREN L.FRANK;US-ASSIGNMENT DATABASE UPDATED:20100413;REEL/FRAME:20823/949
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WARBURG PINCUS PRIVATE EQUITY VIII, L.P.;US-ASSIGNMENT DATABASE UPDATED:20100330;REEL/FRAME:20824/74
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WARBURG PINCUS PRIVATE EQUITY VIII, L.P.;US-ASSIGNMENT DATABASE UPDATED:20100413;REEL/FRAME:20824/74
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;US-ASSIGNMENT DATABASEUPDATED:20100420;REEL/FRAME:20824/215
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;US-ASSIGNMENT DATABASEUPDATED:20100420;REEL/FRAME:20824/376
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:THE FRANK REVOCABLE LIVING TRUST OF CHARLES W. FRANK AND KAREN L.FRANK;US-ASSIGNMENT DATABASE UPDATED:20100420;REEL/FRAME:20823/949
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WARBURG PINCUS PRIVATE EQUITY VIII, L.P.;US-ASSIGNMENT DATABASE UPDATED:20100420;REEL/FRAME:20824/74
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;US-ASSIGNMENT DATABASEUPDATED:20100518;REEL/FRAME:20824/215
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;US-ASSIGNMENT DATABASEUPDATED:20100518;REEL/FRAME:20824/376
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:THE FRANK REVOCABLE LIVING TRUST OF CHARLES W. FRANK AND KAREN L.FRANK;US-ASSIGNMENT DATABASE UPDATED:20100518;REEL/FRAME:20823/949
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WARBURG PINCUS PRIVATE EQUITY VIII, L.P.;US-ASSIGNMENT DATABASE UPDATED:20100518;REEL/FRAME:20824/74
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;REEL/FRAME:20824/215
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:THE FRANK REVOCABLE LIVING TRUST OF CHARLES W. FRANK AND KAREN L.FRANK;REEL/FRAME:20823/949
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998;REEL/FRAME:20824/376
Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WARBURG PINCUS PRIVATE EQUITY VIII, L.P.;REEL/FRAME:20824/74
Apr 1, 2008ASAssignment
Owner name: ZETERA CORPORATION, DELAWARE CORPORATION, CALIFORN
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:FRANK, CHARLES WILLIAM;LUDWIG, THOMAS EARL;BABBIT, WILLIAM R.;AND OTHERS;REEL/FRAME:020735/0326;SIGNING DATES FROM 20080222 TO 20080312
Oct 5, 2007ASAssignment
Owner name: WARBURG PINCUS PRIVATE EQUITY VIII, L.P., NEW YORK
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;REEL/FRAME:019927/0793
Effective date: 20071001
Owner name: WARBURG PINCUS PRIVATE EQUITY VIII, L.P.,NEW YORK
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100413;REEL/FRAME:19927/793
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100330;REEL/FRAME:19927/793
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100420;REEL/FRAME:19927/793
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100518;REEL/FRAME:19927/793
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;REEL/FRAME:19927/793
Jun 20, 2007ASAssignment
Owner name: CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998, CALIFO
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;REEL/FRAME:019453/0845
Effective date: 20070615
Owner name: CORTRIGHT FAMILY TRUST, DATED MAY 13, 1998,CALIFOR
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100413;REEL/FRAME:19453/845
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100330;REEL/FRAME:19453/845
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100420;REEL/FRAME:19453/845
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;US-ASSIGNMENT DATABASE UPDATED:20100518;REEL/FRAME:19453/845
Free format text: SECURITY AGREEMENT;ASSIGNOR:ZETERA CORPORATION;REEL/FRAME:19453/845