Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS20050243611 A1
Publication typeApplication
Application numberUS 11/151,499
Publication dateNov 3, 2005
Filing dateJun 13, 2005
Priority dateMar 26, 2002
Also published asUS6947981, US20030187945
Publication number11151499, 151499, US 2005/0243611 A1, US 2005/243611 A1, US 20050243611 A1, US 20050243611A1, US 2005243611 A1, US 2005243611A1, US-A1-20050243611, US-A1-2005243611, US2005/0243611A1, US2005/243611A1, US20050243611 A1, US20050243611A1, US2005243611 A1, US2005243611A1
InventorsClark Lubbers, Susan Elkington, Randy Hess, Stephen Sicola, James McCarty, Anuja Korgaonkar, Jason Leveille
Original AssigneeClark Lubbers, Susan Elkington, Randy Hess, Sicola Stephen J, Mccarty James, Anuja Korgaonkar, Jason Leveille
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Flexible data replication mechanism
US 20050243611 A1
Abstract
A data replication management (DRM) architecture comprising a plurality of storage cells interconnected by a fabric. Flexibility in connectivity is provided by configuring each storage cell port to the fabric to handle both host data access requests and DRM traffic. Each storage cell comprises one or more storage controllers that can be connected to the fabric in any combination. Processes executing in the storage controller find a path to a desired destination storage cell. The discovery algorithm implements a link service that exchanges information related to DRM between the storage controllers. The DRM architecture is symmetric and peer cooperative such that each controller and storage cell can function as a source and a destination of replicated data. The DRM architecture supports parallel and serial “fan-out” to multiple destinations, whereby the multiple storage cells may implement data replicas.
Images(6)
Previous page
Next page
Claims(19)
1-17. (canceled)
18. A SAN storage controller comprising:
a data processor system configured to implement software processes;
at least one host port;
a pool of storage accessible through the data processor system;
processes executing in the processor to enable the at least one host port to couple to an external communication network;
processes executing in the processor for communicating data access requests with the at least one host port; and
processes executing in the processor for communicating data replication transactions with external storage controllers through the at least one host port.
19. The storage controller of claim 18 wherein the at least one port handles both data access requests and data replication transactions.
20. The storage controller of claim 18 further comprising at least two host ports.
21. The storage controller of claim 18 further comprising more than two host ports.
22. The storage controller of claim 18 further comprising processes within the controller causing the controller to register with a name server in the external communication network.
23. The storage controller of claim 22 wherein the name server stores a storage cell universally unique identification (UUID) value.
24. The storage controller of claim 23 further comprising processes for discovering external storage controllers coupled to the external communications network using the UUID registration information in the name server.
25. The storage controller of claim 18 wherein the at least one port comprises a fibre channel port.
26. The storage controller of claim 18 further comprising processes executing in the processor to virtualize the storage capacity.
27. A storage controller, comprising:
a first port adapted to provide a connection to a switch and configured to handle data transactions and data replication traffic;
a second port adapted to provide a connection to a switch and configured to handle data transactions and data replication traffic, wherein:
the first port serves as a primary port for a data replication operation with a first destination and a redundant port for a data replication operation with a second destination;
the second port serves as a primary port for a data replication operation with a second destination and a redundant port for a data replication operation with a first destination;
data transactions and data replication traffic managed by the first port are routed to the second port in when a data replication operation fails on the first controller; and
data transactions and data replication traffic managed by the second port are routed to the first port in when a data replication operation fails on the first controller.
28. The storage controller of claim 27, wherein at least one port is adapted to maintain a Fibre Channel Arbitrated Loop connection.
29. The storage controller of claim 27, wherein the storage controller comprises a unique identification number.
30. The storage controller of claim 27, wherein in response to a connection failure, the storage controller migrates data replication operations to a separate storage controller.
31. A data replication controller, comprising:
a data processor system configured to implement a software process;
a first port adapted to provide a connection to a switch and configured to handle data transactions and data replication traffic;
a second port adapted to provide a connection to a switch and configured to handle data transactions and data replication traffic;
a pool of storage accessible through the data replication controller;
a first software process executing in the data processor to enable the first host port to couple to an external communication network;
a second software process executing in the data processor to designate a source virtual disk in the pool of storage; and
a third process processes executing in the processor to create a destination virtual disk in the pool of storage.
32. The data replication controller of claim 31, wherein the first port and the second port communicate with a plurality of host systems to handle host data operations and with each other to handle data operations.
33. The data replication controller of claim 31, wherein at least one port is adapted to maintain a Fibre Channel Arbitrated Loop connection.
34. The data replication controller of claim 31, wherein the controller comprises a unique identification number.
35. The data replication controller of claim 31, wherein in response to a connection failure, the data replication controller migrates data replication operations to a separate controller.
Description
    FIELD OF THE INVENTION
  • [0001]
    The present invention relates generally to error recovery in data storage systems, and more specifically, to a system for providing controller-based remote data replication using a redundantly configured Fibre Channel Storage Area Network to support data recovery after an error event and enhance data distribution and migration.
  • BACKGROUND OF THE INVENTION AND PROBLEM
  • [0002]
    Recent years have seen a proliferation of computers and storage subsystems. Demand for storage capacity grows by over seventy-five percent each year. Early computer systems relied heavily on direct-attached storage (DAS) consisting of one or more disk drives coupled to a system bus. More recently, network-attached storage (NAS) and storage area network (SAN) technology are used to provide storage with greater capacity, higher reliability, and higher availability. The present invention is directed primarily SAN systems that are designed to provide shared data storage that is beyond the ability of a single host computer to efficiently manage.
  • [0003]
    Mass data storage systems are implemented in networks or fabrics that provide means for communicating data between systems that use data, and the storage systems that implement the physical storage. In many cases, host computers act as storage servers and are coupled to the network and configured with several disk drives that cumulatively provide more storage capacity or different storage functions (e.g., data protection) than could be implemented by a DAS system. For example, a server dedicated to data storage can provide various degrees of redundancy and mirroring to improve access performance, availability and reliability of stored data. Collecting storage sub-systems, where a separate server manages each sub-system, can form a large storage system. More recently, virtualized storage systems such as the StorageWorks® Enterprise Virtual Array announced by Compaq Corporation in October, 2001 provide storage controllers within a fabric or network that present virtualized storage to hosts that require data storage in a manner that enables the host to be uninvolved in the physical configuration, allocation and management of the storage devices. StorageWorks is a registered trademark of Compaq Computer Corporation. In this system, hosts simply access logical units of storage that appear to the host as a range of logical address space. Virtualization improves performance and utilization of storage.
  • [0004]
    SAN systems enable the possibility of storing multiple copies or “replicas” of data at various physical locations throughout the system. Data replication across multiple sites is desirable for a variety of reasons. To provide disaster tolerance, copies of data stored at different physical locations is desired. When one copy becomes unavailable due to equipment failure, a local network outage, natural disaster or the like, a replica located at an alternate site can allow access to the data. Replicated data can also theoretically improve access in normal operation in that replicas can be accessed in parallel, avoiding bottlenecks associated with accessing a single copy of data from multiple systems. However, prior systems were organized such that one site had a primary role and another site was a replica. Access requests were handled by the primary site until failure, at which time the replica became active. In such architecture, the replica provided little benefit until failure. Similarly, the resources allocated to creating and managing replicas provided minimal load balancing benefit that would enable data access requests to be directed intelligently to replicas such that resources were used more efficiently. Moreover, when multiple replicas are distributed throughout a network topology, it would be beneficial if network delays associated with accessing a topologically remote storage subsystem could be lessened.
  • [0005]
    In the past, managing a data replication system required significant time and expense. This time and expense was often related to tasks involved in setting up and configuring data replication on a SAN. Physical storage devices between original and replica locations had to be closely matched which could require knowledge at the spindle level to set up a storage site to hold a replica. Similarly detailed knowledge of the physical devices at a storage site were required to set up logging of replication operations. Moreover, the logical structures used to represent, access and manage the stored data had to be substantially identically reproduced at each storage site. Many of these operations required significant manual intervention, as prior data replication architectures were difficult to automate. This complexity made it difficult if not impossible to expand the size of a replicated volume of storage, as the changes on one site needed to be precisely replicated to the other site. A need exists to provide data replication systems in a SAN that enable functions involved in setup and configuration of a replication system to be automated, and allow the configuration to be readily expanded.
  • [0006]
    It is desirable to provide the ability for rapid recovery of user data from a disaster or significant error event at a data processing facility. This type of capability is often termed ‘disaster tolerance’. In a data storage environment, disaster tolerance requirements include providing for replicated data and redundant storage to support recovery after the event. In order to provide a safe physical distance between the original data and the data to be backed up, the data is migrated from one storage subsystem or physical site to another subsystem or site. It is also desirable for user applications to continue to run while data replication proceeds in the background. Data warehousing, ‘continuous computing’, and enterprise applications all benefit from remote copy capabilities.
  • [0007]
    Compaq Corporation introduced a data replication management product in its Array Controller Software (ACS) operating on an HSG80 storage controller and described in U.S. patent application Ser. No. 09/539,745 assigned to the assignee of the present application and incorporated herein by reference. This system implemented architecture with redundant storage controllers at each site. Two sites could be paired to enable data replication. While effective, the HSG80 architecture defined relatively constrained roles for the components, which resulted in inflexibility.
  • [0008]
    For example, each of the controllers comprised one port that was dedicated to user data, and a separate port that was dedicated to data replication functions. Even where redundant fabrics were implemented, for a given controller both of these ports were coupled to a common fabric switch. Despite the fact that each controller had two ports for communicating with other controllers, one of the ports was constrained in the role of handling user data, and the other port was constrained in the role of handling data replication. Failure of either port would be, in effect, a failure of the entire controller and force migration of storage managed by the failed controller to the redundant controller. Similarly, failure of a communication link or fabric coupled to one port or the other would render the controller unable to perform its tasks and force migration to the redundant controller. Such migration was disruptive and typically required manual intervention and time in which data was unavailable.
  • [0009]
    As another example, prior data replication management solutions simplified the implementation issues by assigning fixed roles to storage locations. A particular storage site would be designated as a primary when it handled operational data traffic, and another site would be designated as a secondary or backup site. Such architectures were unidirectional in that the backup site was not available for operational data transactions until the failure of the primary site. Such rigidly assigned roles limited the ability to share storage resources across multiple topologically distributed hosts. Moreover, configuration of such systems was complex as it was necessary to access and program storage controllers at both the primary and secondary sites specifically for their designated roles. This complexity made it impractical to expand data replication to more than two sites.
  • [0010]
    This lack of flexible configuration results in constraints imposed on the configuration and functionality of DRM implementations. Most existing data replication solutions have specific constraints around the number of places that the data may be copied, the simultaneity of multi-directional copies. Further, the specific nature of the synchronicity of the data transmission between sites was per controller, not per volume. As a result, all of the copy sets managed by a particular controller had the exact same initiator and target role designations. Also, all copy sets had to go in one direction such that if a controller was an initiator for one copy set, it was an initiator for all copy sets managed by that controller. Further, the replicas were not allowed to vary from the original in any material respect. The target disks were required to have the same size, data protection scheme, and the like as the original. Prior systems could not readily support dynamic changes in the size of storage volumes.
  • [0011]
    The lack of flexible configuration constrains the number of replicas that can be effectively created. Current systems allow an original data set to be replicated at a single location. While a single replica is beneficial for disaster tolerance, it is of limited benefit to improving performance benefits from migrating or distributing data to locations closer to where the data is used. A need exists for a data replication system that improves the ability to fan-out a larger number of replicas to improve geographic and topological diversity.
  • [0012]
    Therefore, there remains a need in the art for a data storage system capable of providing flexible data replication services without the direct involvement of the host computer. Moreover, a data storage system is needed that is readily extensible to provide multiple replication, load balancing, and disaster tolerance without limitations imposed by designating rigid roles for the system components.
  • SUMMARY OF THE INVENTION
  • [0013]
    Briefly stated, the present invention involves architecture for flexible data replication management comprising first and second storage cells at separate topological locations on a storage area network. Each storage location comprises a storage controller having first and second ports coupled to a communication network, and a pool of storage devices accessible through the controller. Each of the first and second ports communicates both user data and replication data with the communication network. Preferably, each storage location is enabled to operate as a source that initiates replication of data stored within its pool of storage devices, and as a destination for storing replicas of data stored within storage devices of another storage location. Preferably, the communication network comprises redundant sub-networks, and each controller is coupled such that the first port couples to one of the sub-networks and the second port couples to another of the sub-networks.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • [0014]
    FIG. 1 shows a storage area network (SAN) environment in which the present invention is implemented;
  • [0015]
    FIG. 2 illustrates an alternative SAN environment embodying the present invention;
  • [0016]
    FIG. 3 shows an implementation emphasizing redundant connections to improve disaster tolerance;
  • [0017]
    FIG. 4 illustrates an implementation with less redundancy than the implementation of FIG. 3;
  • [0018]
    FIG. 5 illustrates a basic implementation of the present invention;
  • [0019]
    FIG. 6 shows data flow relationships in a data replication management operation in accordance with the present invention; and
  • [0020]
    FIG. 7 illustrates logical relationships between data structures of the present invention.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • [0021]
    In general, the present invention describes a data replication management (DRM) architecture comprising a plurality of storage cells interconnected by a data communication network such as a fibre channel fabric. The present invention emphasizes symmetry and peer-cooperative relationships amongst the elements to provide greater flexibility in configuration and operation. Storage cells are for the most part autonomous units of virtualized storage that implement hundreds of gigabytes to terabytes of storage capacity. In accordance with the present invention, each storage cell can act as a source or primary location for data storage, and at the same time act as a destination or secondary location holding a replica data from a primary location of another storage cell. Similarly, the architecture of the present invention seeks to minimize rigid roles placed on interface ports and connections such that all available connections between components can support both host data traffic (e.g., traffic involving data that is to be stored and retrieved), and DRM traffic (e.g., traffic related to managing replicas between storage cells). In this manner, resources are efficiently deployed and operational performance is improved.
  • [0022]
    The present invention is described with various levels of specificity to ease description and understanding. However, unless specified otherwise, the specific implementations are examples only, and not limitations of the invention. For example, network connectivity is illustrated by fibre channel mechanisms, however, other network mechanisms provide suitable functionality in particular environments. In particular implementations, storage capacity is presented as SCSI (small computer system interface) logical units, although the particular protocol is readily varied to meet the needs of a particular application.
  • [0023]
    FIG. 1 illustrates a multi-site storage area network (SAN) implementing features in accordance with present invention. The example shown in FIG. 1 includes only three sites (Site 1, Site 2 and Site 3) for to ease illustration and understanding, however, it should be understood that any number of sites may be provided to meet the needs of a particular application. The various sites may be physically proximate, or separated by an arbitrary distance. A particular implementation may well limit the number of possible sites, and may well limit the maximum or minimum physical separation of sites. Each site includes one or more storage cells 101, such as cells 101A, 101B, 101C and 101D. Any number of storage cells 101 maybe included in any site, although the number implemented in any particular application may be constrained to meet the needs of that application.
  • [0024]
    The storage implemented at various sites is accessed by host computer 102, such as host computers 102A, 102B and 102C. Host computers are generally machines that consume or require data storage. Typical host computers 102 demand large quantities of storage such as mainframes, web servers, transaction processors, and the like. However, a host computer 102 may comprise a computer of any processing capacity that requires or benefits from network storage either to reduce cost of the host 102, implement more storage capacity than practical in a host 102, share data amongst multiple hosts 102, or the like. A host computer 102 may couple to the storage cells 101 via a connection to network 103 such as illustrated by host computers 102A and 102B. In many cases, a host computer 102 will be located in a site such as host computer 102C located in site 2. The topological location of host computers 102 is a matter of design choice selected to meet the needs of the particular application. In many cases, one or more host computers 102 will be located at a site. It should be appreciated, however, that host computers 102 have little to do with the management and configuration of the DRM system of the present invention, other than the reality that as consumers of data, the DRM system is preferably able to satisfy storage needs of host computers 102 at various physical and topological locations.
  • [0025]
    Storage cells 101 and hosts 102 couple to data communication network 103. Storage cells 101 implement a quantity of data storage capacity that is accessible through storage controllers 105 that implement one or more connections to network 101. Storage cells 101 typically implement hundreds of gigabytes to terabytes of physical storage capacity. Preferably, controllers 105 virtualize the physical storage capacity such that it is configurable into logical units (LUNs) of storage capacity. The LUNs implement an arbitrary quantity of logical address block storage, where each LUN has a specified level of data protection such as RAID 0-5 data protection.
  • [0026]
    The logical construction or context of a storage cell 101 comprises the logical and data structures that are used to represent, manage and configure the raw physical storage capacity into devices that can be accessed by hosts 102 in a desired manner. To do this, controllers 105 implement and manage various logical objects that implement data structures and behavior to represent various components of each storage cell 101, preferably without involvement if hosts 102. All objects in the system are identified with a universally unique identification (UUID). The UUIDs are stored persistently in metadata structures within the physical storage which enables all or part of a storage cell, and all the LUNs implemented therein, to be mapped and presented by any storage controller 105. In a data replication application, the ability to present the LUNs and the context in which they exist is useful, as described in greater detail below, in event of a controller failure at one of the replica sites.
  • [0027]
    Of particular significance to the present invention is a logical object representing each LUN that is presented to a host system 102. In addition to the UUID, LUNs are also identified by a worldwide LUN ID (WWLID) which identifies the LUN to hosts 102. Hosts 102 access physical storage capacity by addressing read and write operations to specified LUNs using the WWLID, and can be otherwise unaware of the physical storage architecture or data protection strategy for a particular LUN that is being accessed. Storage controllers 105 manage the tasks of allocating physical storage capacity to specified LUNs, monitoring and maintaining integrity of the LUNs, moving data between physical storage devices, and other functions that maintain integrity and availability of the data stored therein.
  • [0028]
    Network 103 comprises any of a variety of available networks, and may comprises a plurality of interconnected networks. In particular examples, network 103 comprises at least two independent fibre channel fabrics to provide redundancy. These fibre channel fabrics may comprise long-distance connection mechanism 201 such as asynchronous transfer mode (ATM) and Internet protocol (IP) connections that enable sites to be separated by arbitrary distances.
  • [0029]
    At least one SAN management appliance (SMA) 107 is coupled to network 103 to enable connections to storage cells 101. In practice, a number of SMAs 107 are provided, and typically an SMA 107 is located at each site to provide management operation for the storage cells 101 at that site. However, because each SMA 107 communicates through network 103, the physical location of SMAs 107 is arbitrary. SMAs 107 are preferably implemented at topological locations that provide redundant connections to the storage cells 101.
  • [0030]
    Network 103 is accessible directly or indirectly to all components at Site 1, Site 2, and Site 3 including hosts 102 and controllers 105. In preferred implementations, each component has redundant links to network 103, and network 103 is preferably implemented as having redundant sub-networks as shown in FIG. 2. Redundancy provides connectivity in event of failure or degradation of some portions of network 103. Redundancy also enables connectivity in event of failure or degradation of controllers 105 and/or interface components of hosts 102. In prior data replication systems, it was known to provide multiple channels between components, however, these channels were not interchangeable. For example, each controller might have two host ports to a network, however, one of the ports was designed to support operational data traffic while the other port was dedicated to data replication operations. While this division of responsibility was simpler to implement, it increased the likelihood that one or the other of the ports would be used at less than full capacity. Moreover, special purpose ports are not completely redundant, hence the system would be vulnerable to failure of one or the other of the channels. Even where two controllers were provided, this asymmetry between ports affected performance.
  • [0031]
    In the systems of FIG. 1 and FIG. 2, it is important to realize that storage cell 101 is able to function as a primary storage location for any of hosts 102, and at the same time function as a secondary or alternative storage location for a replica of data from another storage cell 101. This feature is referred to as “bi-directionality”. Unlike prior DRM systems where sites were rigidly configured as either primary or secondary sites, the present invention provides for flexible assignment at a much finer level of granularity, preferably such that individual virtual disks can be designated as either primary or alternate. One effect of such flexibility is that it becomes practical to implement enormous quantities of storage in each storage cell 101 as the capacity at each site can be readily applied to various applications with disparate storage needs.
  • [0032]
    FIG. 3 illustrates various features of the present invention embodied in a disaster tolerant configuration. The configuration of FIG. 3 emphasizes redundancy in physical components. As a result, multiple paths exist between any host and any storage. Storage management appliances 107 include a Bridge and a data replication management console (DRMC) which are cooperative software processes that each present a user interface for configuration and monitoring of DRM processes. The Bridge functions as an element manager for controllers 105, and handles basic functions of connecting to the communication network 103, issuing commands to controllers 105, and implementing security policies. All customer configuration and monitoring of DRM activities goes through the Bridge. The DRMC is provides increased functionality over the Bridge interface, by managing solely DRM storage and providing assists, such as with site recovery.
  • [0033]
    One goal of a disaster tolerant configuration in FIG. 3 is to provide hardware redundancy for multiple levels of fault tolerance in order to keep the primary and alternate sites running, almost at ‘non-stop’ metrics (seconds per year downtime, barring disaster or planned downtime). In the face of single component failures at a site, DRM will failover to a redundant component at that site to allow continued operation. If a significant failure happens to the primary site, then data processing can be resumed at the alternate site where the data is intact. Database or application recovery is achieved with little or no downtime.
  • [0034]
    In this regard, wherever redundant components are provided, they are preferably interchangeable in accordance with the present invention such that functionality of one component can be taken over by a redundant component. For example, node ports in the SMA 107 are redundant such that failure of one node port can be handled by moving functions handled by that port over to the other, non-failed node-port.
  • [0035]
    Physical storage devices 302 are coupled to controllers 105 by a high-speed redundant connection such as fibre channel arbitrated loop (FCAL) connections 303 in FIG. 3. An FCAL connection allows each of tens or hundreds of individual hard disk drives to be accessed not only at high speed, but by either controller 105 in a storage cell. Further, by providing redundant FCALs, one FCAL can fail while each storage device 302 remains accessible by either controller. The present invention is largely independent of the specific implementation of physical storage devices 302, and it is contemplated that other physical storage architectures will provide suitable equivalents for purposes of the present invention.
  • [0036]
    FIG. 3's disaster tolerance configuration provides for high availability with a dual fabric implemented by switches 301, dual hosts 102, and dual storage topology, where a single switch, host, or storage can fail and the system can still continue access on the SAN. Each fabric is composed of two switches 301, with the switches connected to each other over what is called an E-port, or expansion port. The E-port can be a connection with distances that vary depending on technology. Fibre Channel allows for 10km and with extenders up to 100 km. ATM provides for ‘round the world’ distances. New technology, called FC-BB (Fibre Channel Backbone), provides the opportunity to extend Fibre Channel over leased Telco lines (also called WAN tunneling). There is no theoretical constraint imposed on the number of switches 301, although a particular implementation may choose to impose such a constraint. With more switches 301, the fabric 103 becomes more complex and port-to-port communications may require more “hops”. Hence, both the latency and the variability in latency increase with more complex fabrics.
  • [0037]
    Hosts 102 preferably run multi-pathing software that dynamically allows failover between storage paths as well as static load balancing of storage volumes (LUNs) between the paths of controllers 102. Multi-pathing software enables a host 102 to identify multiple connection options that lead to desired storage, and select amongst the available paths based on selected criteria. These criteria may vary widely, but typically include path availability, path performance, path load, controller load, and the like. This allows for applications to continue given any failure of fabric or the controllers. A lower cost variant of this configuration uses a fabric comprising only one switch, such that there is no E-port. Such a configuration may be appropriate for shorter distances.
  • [0038]
    FIG. 4 illustrates a configuration of the present invention that emphasizes data movement rather than redundancy of the example in FIG. 3. The configuration of FIG. 4 provides a lower cost solution ideal for customers only interested in data backup, data migration, data distribution, and data mining. FIG. 4 has a single fabric formed by two interconnected switches. With shorter distances, an even lower cost variant could use only one switch, such that there is no E-port. FIG. 5 illustrates an even more basic configuration using a fabric having a single switch 301, one host 102, and non-redundant controllers 105.
  • [0039]
    Because ports of controllers 105 are configured to handle both operational data transactions and DRM traffic, a single port connection between each controller 105 and switch 301 is all that is necessary to implement data replication features of the present invention. However, the basic configuration shown in FIG. 5 sacrifices much redundancy, and would not be recommended for most applications. It is important to understand that a great deal of flexibility exists in configuring the present invention for specific applications. Redundant connections and components can be added and removed flexibly to meet the needs of a particular application. Hence, various features of the configurations shown in FIG. 1 through FIG. 5 can be mixed and matched to configure a specific implementation.
  • [0040]
    A SAN, such as shown in FIG. 1 through FIG. 5, is established by coupling controllers 105 and hosts 102 to switch 301, or a fabric formed from multiple switches 301. Each of the devices performs a log in operation according to applicable fibre channel standards. The log in operation involves an exchange of information in which each device is registered with the fabric 103, and devices can discover other devices that are registered with fabric 103. In the preferred implementations, each controller 105 is identified by a storage cell unique identification number (UUID) which is stored by a name server in fabric 103. By way of this registration and discovery process, each controller 105 can obtain knowledge necessary to connect to any other controller 105.
  • [0041]
    As noted before, metadata needed to reconstruct a storage cell 101 including the UUIDs of the various objects in that storage cell is stored persistently. Because of this, in the event of a destination controller failure, another controller 105, typically a redundant or partner controller 105 in the same storage cell 101 as the failed controller 105, can reconstruct the storage cell 101. Other controllers 10 involved in a copy set can continue to locate the now reconstructed storage cell 101, and continue data replication operations without loss of data or requiring manual processes. The ability to implement this level of failure recovery substantially or wholly automatically is unprecedented.
  • [0042]
    In the event of a connection failure between a source controller 105 and a destination controller 105, the source controller 105 can readily identify each available alternate controller 105 and continue operation without loss of data or requiring manual intervention. A connection failure or link failure is distinct from a controller failure in that the controllers 105 remain operational, but unable to maintain a data communication link with one or more other controllers 105. In response to a connection failure, the source controller has several options, depending on the nature of the failure. When the source controller 105 is unable to link to fabric 103, suggesting a failure in a port of switch 301, the controller 105 can attempt to use an alternate link to fabric 103, or a redundant fabric 103 if available. When the link between fabric 103 and a destination controller 105 has failed, the source controller 105 attempts to establish a new link with a redundant port of the destination controller 105. In both of these cases, there will be no need to reconstruct the storage cell 101, and data replication operations will proceed with substantially no interruption, and in most cases without any need to notify the host 102 of the aberrant condition. In contrast, prior systems had only one port on each controller available for DRM operations and so a link failure would require reconstruction of either the source or destination LUN in an alternate controller.
  • [0043]
    In some cases a link failure may require migration of a LUN from one controller 105 to another. For example, if all ports on either the destination or source 5 controllers 105 failed (which would typically be considered a controller failure), or several ports to fabric 103 failed simultaneously in particular combination, recovery would require migrating presentation of one or more LUNs from one controller 105 to a redundant or partner controller 105 in a manner similar to a controller failure event. This migration can be performed efficiently using the mechanisms for mapping and presenting a storage cell 101 described above. Notably, all LUNs do not need to be presented before pending data replication operations can be completed. In implementations where the LUNs are virtualized, as described in co-pending patent application Ser. No. 10/040,194 entitled “SYSTEM AND METHOD FOR ATOMIZING STORAGE” filed on Oct. 22, 2001 and which is assigned to the assignee of the present invention and incorporated herein by reference, it is possible to implement only so much of the data structures needed to represent the portion of the LUN that is the subject of a pending data transaction.
  • [0044]
    FIG. 6 shows hierarchical relationships defined in accordance with the present invention to model data replication management. Prior to performing DRM operations, a storage cell 101 is implemented at each site, and a virtual disk 601 is allocated within a storage cell 101 that is handing operational data with one or more hosts 102. For any particular copy set, one virtual disk 601 is designated as a source, although it should be understood that a source virtual disk 601 might be allocated at any site implementing the DRM features of the present invention. Virtual disk 601 may comprise a plurality of physical storage resources that span multiple physical drives within the storage cell 101, and may implement any desired capacity and data protection type.
  • [0045]
    A destination virtual disk 602 is allocated within a storage cell 101 at a designated alternative or destination site. In normal operation, write operations are directed to source virtual disk 601, and copied in a background process to one or more destination virtual disks 602. Destination virtual disk has the same logical storage capacity as the source virtual disk, but may provide a different data protection configuration. Controllers 105 of the destination storage cell handle the allocation of destination virtual disk 602 autonomously. This allocation involves creating data structures that map logical addresses to physical storage capacity, and in a particular implementation involve processed described in greater detail in U.S. patent application Ser. No. 10/040,194.
  • [0046]
    The actual allocation and copying of data may take up to several minutes to several hours in the case of storage of many gigabytes or terabytes. These processes can be streamlined by allocating and copying only logical storage blocks that are used in source virtual disk 601. For example, a virtual disk 601 may implement 1 terabyte of storage, but a sparsely populated example may use only a few gigabytes of capacity to store actual data. In accordance with the present invention, destination virtual disk 602 will reserve resources needed to implement the entire 1 terabyte, but will only allocate and copy the locations that are actually used to store data and may allocate the other locations. This greatly reduces the time required to create replicates.
  • [0047]
    A “copy set” is a set of member virtual disks where each member virtual disk is a replica of the others, and may comprise any number of replica virtual disks. While it may be easier to describe and understand the present invention by designating one virtual disk as a source or original virtual disk and the others as destinations or replicas, it is important to understand that once the copy set is created each member virtual disk is essentially a peer of all others. To create a copy set, the user designates a virtual disk as the source, and an alternate site storage cell (not a virtual disk) for the destination. The destination virtual disk does not exist prior to the copy set creation operation. Instead, it is created specifically to be a member of the copy set. Because each destination or alternative virtual disk is created on demand, it is created specifically and automatically to be compatible with the copy set. Hence, there is little or no manual procedures that must be performed to create new members for a copy set. This allows the copy set to increase (or decrease) its membership readily.
  • [0048]
    As noted before, a LUN is identified to a host 102 by a WWLID. One feature of a particular embodiment of the present invention is that each LUN in a copy set is presented to hosts 102 using the same WWLID. In this manner, the LUN appears to the host 102 as a single LUN that can be accessed by multiple paths. Preferably, each controller 105 that manages one of the LUNs in a copy set can be queried by a host 102 to determine a particular LUN or controller 105 (identified by the UUID) that is preferred (but not mandatory) for use by that host for access to the LUN. This feature supports large fan out in that any number of LUNs in a copy set can be presented as a single LUN with replication between the LUNs handled automatically by controllers 105 without involving hosts 102.
  • [0049]
    Initially, when a controller 105 at the destination site is requested to create a destination virtual disk, it determines whether its storage cell has sufficient resources (e.g., storage capacity) to service the request, reserves the necessary resources if available, and responds back to the controller 105 at the source site that the destination virtual disk is created. An existing virtual disk cannot be used for the destination. This process is very quick as the destination virtual disk is not actually allocated, and the data is not actually copied from source to destination at this time. Once the source virtual disk 601 receives this confirmation, it can continue normal operations while the actual allocation of virtual disk 602 and copying of data can occur as background processes. Hence, the actual time that source virtual disk 602 is quiesced during the data replication initialization is a few milliseconds. It is also possible that accesses to source virtual disk 601 can be cached or buffered during the initialization process so that no downtime is observed from the hosts 102. A controller may support tens, hundreds, or thousands of copy sets, limited only by the practical constraint of the computing resources (e.g., memory and processing power) needed to manage the number of copy sets.
  • [0050]
    To implement a copy transaction between a source and destination, a path to the alternate site is found and a connection is established. This connection need not be a persistent connection, although for data that frequently changes, a persistent connection may be efficient. Preferably, heartbeat is initiated over the connection. Heartbeats are in both directions, meaning the source site generates a heartbeat on each connection, and the alternate site(s) generates a heartbeat on each connection. Heartbeat timeout intervals are adaptive based on distance (computed round trip delay).
  • [0051]
    An attempt is made to choose paths by not moving the units between controllers. In other words, a controller that is used to create the source or destination virtual disks is initially preferred, which constrains the number of available paths initially. The chosen path/connection is used until it breaks, then an attempt is made to find another path. The port/path selection algorithm may include other default designations such as a preferred fabric or switch, or preferred ports on controllers, or similar default preferences that may operate to select among redundant resources.
  • [0052]
    In the preferred implementation, objects do not span sites or storage controllers. Instead, objects exist at each site and are associated with each other as suggested in FIG. 7. For instance, the controller views a “DRM Group” 705 as a source group associated with a destination group (represented by a dashed-line oval in FIG. 7). The controller does not have a copy set object per se, rather group members (virtual disks) that are associated with a remote member's (virtual disks). The copy set is represented by this association. In FIG. 7, virtual disks 701 are associated with virtual disks 702 to form a copy set and virtual disks 711 are associated with virtual disks 712 to form another copy set.
  • [0053]
    DRM groups 705 comprise a set of related virtual disks or LUNs that belong to copy sets all of which have the same source and destination. Multiple groups 705 can exist for a given source and destination combination. While in one sense a group 705 spans sites, in practice a separate logical object that represents a group is implemented in each site (i.e., each controller 105 managing a group). Hence, in another sense each group 705 comprises a set of source halves of copy sets (i.e., the source LUNs 701 and 711), or destination halves of copy sets (i.e., destination LUNs 702 and 712). A group 705 is considered a “local group” by the controller 105 that implements the group logical object, and is considered a “remote group” when implemented by another controller 105. In other words, whether a group 705 is local or remote is determined from the perspective of a particular controller 105. A local group 105 is associated with a remote group 105 when the LUNs contained in each group are in the same copy set. In a multidirectional implementation, each local group 705 will be associated with a plurality of remote groups 105.
  • [0054]
    In essence, a group contains a collection of LUNS for which a combined I/O stream must be replicated in the LUNs of each remote group to provide certain properties with regard to its interruption (e.g. a copy set holding a database and a copy set holding a journal related to the database). DRM groups 705 are used for maintaining crash consistency and preserving WRITE ordering. Crash consistency refers operation in the event that a host 102 or link to a host 102 crashes during a data transaction. In such event, it is important that the data transaction be committed to all members of the group or fail in all members of the group. In a database/journal implementation, for example, it is important to prevent a situation in which the database is modified, but the journal does not reflect the modification, or vice versa. The consistency property applies when the group has more than one member. In general, more than one virtual disk should be added to a group only if the application requires it. In most cases, a group should comprise only one member for optimal performance, in which cases the crash consistency behaviors are less important.
  • [0055]
    A group maintains write ordering among the members for asynchronous operation and logging/merging. Asynchronous operation refers to an operation mode in which a modification to one member of a copy set can be propagated to other members of the copy set after a time delay. During this time delay, the various replicas are inexact. When asynchronous operation is allowed, it is important that all replicas eventually implement the modification. Since multiple modification operations may be pending but uncommitted against a particular replica, it is necessary that the original order in which the modifications were presented be preserved when the pending modifications are applied to each replica. Even where asynchronous operation is not explicitly allowed, a destination LUN may become unavailable for a variety of reasons, in which case a copy set is implicitly operating in an asynchronous mode.
  • [0056]
    To ensure write order preservation, a record is maintained in a non-volatile storage device such as a cache for each group 705 that records the history of write commands and data from a host. The record is sized to store all write transactions until the transaction is committed to each member of a copy set. When required, the record can be written to a log on media-based storage and then later be replayed to merge the pending writes, in order, to each remote group 705. The ordering algorithm uses a “group sequence number” and the remote groups 705 ensure that the data is written in order sequence. Group members enter and exit logging at the same time, to assure order across the volumes.
  • [0057]
    Virtual disks in a group should have the same alternate site. A group is limited to 32 virtual disks in a particular implementation. A virtual disk can belong to at most one group. Virtual disks in the same DRM group 705 may belong to different disk groups. When a group object is created on the primary site controller, the controller automatically creates a symmetric group object on the alternate site controller. A group is created during copy set creation, if the user chooses not to use an existing group.
  • [0058]
    At a site, all members of a group are accessed on the same controller in a dual pair to enforce cache coherency (i.e., not split between dual controllers). The preferred storage controller should be the same on all members. When members are added to a group, they are automatically moved to reside on the same controller, and thereafter will failover together. If the preferred storage controller is changed on one member, it will change the setting for all members. These constraints create a configuration in which all members in a group share the same connection/path between source and destination sites. Different groups may share the same connection/path to the same alternate site or a different one. If more than one member in a group needs to copy, then they are copied one at a time with a queue depth selected to keep the pipe full for optimal performance.
  • [0059]
    One benefit of the flexible architecture described above is that the number of replicas in a copy set can be increased. As noted hereinbefore, conventional systems allow only one replica in a given copy set. In accordance with the present invention, any number of replicas can be included in a copy set. This is because roles can be switched dynamically between source and destination which eases propagation of data across multiple members. Also, because the storage at each site is virtualized, LUNs of any desired size and configuration can be automatically configured from physical storage capacity available at that site. Accordingly, there is no need to specially configure hardware or software at each site in order to implement a LUN that is compatible with a given copy set. Because each controller 105 can automatically increase the size or change the configuration of the LUN(s) it manages, the manual tasks involved to increase the size of a LUN do not grow geometrically or exponentially with the number of replicas in the copy set. Instead, the user simply manages the LUN as a single entity, and the implementation of changes in each replica is handled automatically.
  • [0060]
    It is apparent that the system of the present invention provides efficiency in that redundant components can operate simultaneously and in parallel to provide improved throughput as well as serving during failover to replace functionality of failed systems and connections. Because system components have few constraints on their functionality, the present invention allows great flexibility in implementing a particular system to emphasize any balance of, for example, low latency, high availability, disaster tolerance, and cost.
  • [0061]
    Although the invention has been described and illustrated with a certain degree of particularity, it is understood that the present disclosure has been made only by way of example, and that numerous changes in the combination and arrangement of parts can be resorted to by those skilled in the art without departing from the spirit and scope of the invention, as hereinafter claimed.
Patent Citations
Cited PatentFiling datePublication dateApplicantTitle
US5682518 *Mar 22, 1995Oct 28, 1997Nec CorporationSystem for updating inactive system memory using dual port memory
US5692155 *Apr 19, 1995Nov 25, 1997International Business Machines CorporationMethod and apparatus for suspending multiple duplex pairs during back up processing to insure storage devices remain synchronized in a sequence consistent order
US5870537 *Mar 13, 1996Feb 9, 1999International Business Machines CorporationConcurrent switch to shadowed device for storage controller and device errors
US5923876 *Aug 24, 1995Jul 13, 1999Compaq Computer Corp.Disk fault prediction system
US6148414 *Mar 3, 1999Nov 14, 2000Seek Systems, Inc.Methods and systems for implementing shared disk array management functions
US6161192 *Oct 13, 1995Dec 12, 2000Compaq Computer CorporationRaid array data storage system with storage device consistency bits and raidset consistency bits
US6170063 *Mar 7, 1998Jan 2, 2001Hewlett-Packard CompanyMethod for performing atomic, concurrent read and write operations on multiple storage devices
US6295578 *Apr 9, 1999Sep 25, 2001Compaq Computer CorporationCascaded removable media data storage system
US6397293 *Jan 19, 2001May 28, 2002Hewlett-Packard CompanyStorage management system and auto-RAID transaction manager for coherent memory map across hot plug interface
US6487636 *Apr 24, 2000Nov 26, 2002Hewlett-Packard Co.Method and apparatus for mapping data in a heterogeneous disk array storage system
US6490122 *Aug 28, 1998Dec 3, 2002Hewlett-Packard CompanySystem and method for providing power and control signals to a cartridge access device in a cartridge storage system
US6493656 *Feb 26, 1999Dec 10, 2002Compaq Computer Corporation, Inc.Drive error logging
US6505268 *Apr 8, 1999Jan 7, 2003Compaq Computer CorporationData distribution in a disk array
US6523749 *Mar 6, 2001Feb 25, 2003Hewlett-Packard CompanyApparatus and method for retrieving data cartridge information external to a media storage system
US6546459 *Mar 15, 2001Apr 8, 2003Hewlett Packard Development Company, L. P.Redundant data storage systems and methods of operating a redundant data storage system
US6560673 *Jan 31, 2001May 6, 2003Hewlett Packard Development Company, L.P.Fibre channel upgrade path
US6587962 *Oct 20, 1999Jul 1, 2003Hewlett-Packard Development Company, L.P.Write request protection upon failure in a multi-computer system
US6594745 *Jan 31, 2001Jul 15, 2003Hewlett-Packard Development Company, L.P.Mirroring agent accessible to remote host computers, and accessing remote data-storage devices, via a communcations medium
US6601187 *Mar 31, 2000Jul 29, 2003Hewlett-Packard Development Company, L. P.System for data replication using redundant pairs of storage controllers, fibre channel fabrics and links therebetween
US6606690 *Feb 20, 2001Aug 12, 2003Hewlett-Packard Development Company, L.P.System and method for accessing a storage area network as network attached storage
US6609145 *Sep 22, 1998Aug 19, 2003Hewlett-Packard Development Company, L.P.User selectable priority for disk array background operations
US6629108 *Mar 9, 2001Sep 30, 2003Hewlett-Packard Development Company, Lp.Method for insuring data integrity for mirrored independently accessible memory devices
US6629273 *Jan 24, 2000Sep 30, 2003Hewlett-Packard Development Company, L.P.Detection of silent data corruption in a storage system
US6643795 *Mar 30, 2000Nov 4, 2003Hewlett-Packard Development Company, L.P.Controller-based bi-directional remote copy system with storage site failover capability
US6647514 *Mar 23, 2000Nov 11, 2003Hewlett-Packard Development Company, L.P.Host I/O performance and availability of a storage array during rebuild by prioritizing I/O request
US6681308 *Nov 16, 2001Jan 20, 2004Hewlett-Packard Development Company, L.P.Method for automatically converting block size and formatting backend fiber channel discs in an auto inclusive storage array environment
US6708285 *Mar 15, 2001Mar 16, 2004Hewlett-Packard Development Company, L.P.Redundant controller data storage system having system and method for handling controller resets
US6715101 *Mar 15, 2001Mar 30, 2004Hewlett-Packard Development Company, L.P.Redundant controller data storage system having an on-line controller removal system and method
US6718404 *Jun 1, 2001Apr 6, 2004Hewlett-Packard Development Company, L.P.Data migration using parallel, distributed table driven I/O mapping
US6718434 *May 31, 2001Apr 6, 2004Hewlett-Packard Development Company, L.P.Method and apparatus for assigning raid levels
US6721902 *Oct 12, 2000Apr 13, 2004Hewlett-Packard Development Company, L.P.Method and system for providing LUN-based backup reliability via LUN-based locking
US6725393 *Nov 6, 2000Apr 20, 2004Hewlett-Packard Development Company, L.P.System, machine, and method for maintenance of mirrored datasets through surrogate writes during storage-area network transients
US6742020 *Jun 8, 2000May 25, 2004Hewlett-Packard Development Company, L.P.System and method for managing data flow and measuring service in a storage network
US6745207 *Jun 1, 2001Jun 1, 2004Hewlett-Packard Development Company, L.P.System and method for managing virtual storage
US6763409 *Jan 31, 2001Jul 13, 2004Hewlett-Packard Development Company, L.P.Switch-on-the-fly GBIC disk channel adapter and disk channel system
US6772231 *Jun 1, 2001Aug 3, 2004Hewlett-Packard Development Company, L.P.Structure and process for distributing SCSI LUN semantics across parallel distributed components
US6775790 *Jun 1, 2001Aug 10, 2004Hewlett-Packard Development Company, L.P.Distributed fine-grained enhancements for distributed table driven I/O mapping
US6795904 *Mar 28, 2002Sep 21, 2004Hewlett-Packard Development Company, L.P.System and method for improving performance of a data backup operation
US6802023 *Mar 15, 2001Oct 5, 2004Hewlett-Packard Development Company, L.P.Redundant controller data storage system having hot insertion system and method
US6807605 *Oct 3, 2002Oct 19, 2004Hewlett-Packard Development Company, L.P.Managing a data storage array, a data storage system, and a raid controller
US6817522 *Jan 24, 2003Nov 16, 2004Hewlett-Packard Development Company, L.P.System and method for distributed storage management
US6823453 *Oct 6, 2000Nov 23, 2004Hewlett-Packard Development Company, L.P.Apparatus and method for implementing spoofing-and replay-attack-resistant virtual zones on storage area networks
US6839824 *Dec 28, 2001Jan 4, 2005Hewlett-Packard Development Company, L.P.System and method for partitioning a storage area network associated data library employing element addresses
US6842833 *Jun 30, 1998Jan 11, 2005Hewlett-Packard Development Company, L.P.Computer system and method for transferring data between multiple peer-level storage units
US6845403 *Oct 31, 2001Jan 18, 2005Hewlett-Packard Development Company, L.P.System and method for storage virtualization
US6947981 *Mar 26, 2002Sep 20, 2005Hewlett-Packard Development Company, L.P.Flexible data replication mechanism
US20010019920 *Jan 24, 2001Sep 6, 2001Framatome Connectors InternationalControl unit of a valve mechanism
US20020019863 *Jun 1, 2001Feb 14, 2002Reuter James M.Structure and process for distributing SCSI LUN semantics across parallel distributed components
US20020019908 *Jun 1, 2001Feb 14, 2002Reuter James M.System and method for managing virtual storage
US20020019922 *Jun 1, 2001Feb 14, 2002Reuter James M.Data migration using parallel, distributed table driven I/O mapping
US20020019923 *Jun 1, 2001Feb 14, 2002Reuter James M.Distributed fine-grained enhancements for distributed table driven I/O mapping
US20020048284 *Feb 13, 2001Apr 25, 2002Moulton Gregory HaganSystem and method for data protection with multidimensional parity
US20020188800 *May 15, 2001Dec 12, 2002Tomaszewski Richard J.Self-mirroring high performance disk drive
US20030051109 *Jun 28, 2001Mar 13, 2003Cochran Robert A.Method and system for providing logically consistent logical unit backup snapshots within one or more data storage devices
US20030056038 *Jun 28, 2001Mar 20, 2003Cochran Robert A.Method and system for providing advanced warning to a data stage device in order to decrease the time for a mirror split operation without starving host I/O request processsing
US20030063134 *Jan 5, 2001Apr 3, 2003Bob LordSystem for displaying a hierarchical directory
US20030074492 *May 29, 2001Apr 17, 2003Cochran Robert A.Method and system for efficient format, read, write, and initial copy processing involving sparse logical units
US20030079014 *Oct 22, 2001Apr 24, 2003Lubbers Clark E.System and method for interfacing with virtual storage
US20030079082 *Oct 19, 2001Apr 24, 2003Sicola Stephen J.Unified management system and method for multi-cabinet data storage complexes
US20030079083 *Oct 22, 2001Apr 24, 2003Lubbers Clark E.High performance multi-controller processing
US20030079102 *Oct 22, 2001Apr 24, 2003Lubbers Clark E.System and method for generating point in time storage copy
US20030079156 *Oct 19, 2001Apr 24, 2003Sicola Stephen J.System and method for locating a failed storage device in a data storage system
US20030084241 *Oct 22, 2001May 1, 2003Lubbers Clark E.System and method for atomizing storage
US20030101318 *Nov 20, 2002May 29, 2003Hitachi, Ltd.Data copy method
US20030110237 *Sep 25, 2002Jun 12, 2003Hitachi, Ltd.Methods of migrating data between storage apparatuses
US20030126315 *Dec 28, 2001Jul 3, 2003Choon-Seng TanData storage network with host transparent failover controlled by host bus adapter
US20030126347 *Dec 27, 2001Jul 3, 2003Choon-Seng TanData array having redundancy messaging between array controllers over the host bus
US20030140191 *Jan 24, 2002Jul 24, 2003Mcgowen Michael E.System, method, and computer program product for on-line replacement of a host bus adapter
US20030145045 *Jan 31, 2002Jul 31, 2003Greg PellegrinoStorage aggregator for enhancing virtualization in data storage networks
US20030145130 *Jan 31, 2002Jul 31, 2003Schultz Stephen M.Array controller ROM cloning in redundant controllers
US20030170012 *Mar 6, 2002Sep 11, 2003Robert A. CochranMethod and system for reliable remote-mirror resynchronization in disk arrays and other mass storage devices
US20030177323 *Jan 11, 2002Sep 18, 2003Mathias PoppRemote mirrored disk pair resynchronization monitor
US20030187847 *Mar 26, 2002Oct 2, 2003Clark LubbersSystem and method for ensuring merge completion in a storage area network
US20030187947 *Mar 26, 2002Oct 2, 2003Clark LubbersSystem and method for multi-destination merge in a storage area network
US20030188085 *Nov 20, 2002Oct 2, 2003Hitachi, Ltd.Clustered storage system and its control method
US20030188114 *Mar 26, 2002Oct 2, 2003Clark LubbersData replication with virtualized volumes
US20030188119 *Mar 26, 2002Oct 2, 2003Clark LubbersSystem and method for dynamically managing memory allocated to logging in a storage area network
US20030188153 *Apr 2, 2002Oct 2, 2003Demoff Jeff S.System and method for mirroring data using a server
US20030188218 *Mar 26, 2002Oct 2, 2003Clark LubbersSystem and method for active-active data replication
US20030188229 *Mar 26, 2002Oct 2, 2003Clark LubbersSystem and method for managing data logging memory in a storage area network
US20030188233 *Mar 28, 2002Oct 2, 2003Clark LubbersSystem and method for automatic site failover in a storage area network
US20030191909 *Nov 21, 2002Oct 9, 2003Hitachi, Ltd.Computer system, storage and storage utilization and monitoring method
US20030191919 *Nov 21, 2002Oct 9, 2003Hitachi, Ltd.Volume management method and apparatus
US20030196023 *Feb 20, 2003Oct 16, 2003Inostor CorporationData redundancy methods and apparatus
US20030212781 *Nov 20, 2002Nov 13, 2003Hitachi, Ltd.Network topology management system, management apparatus, management method, management program, and storage media that records management program
US20030229651 *Oct 24, 2002Dec 11, 2003Hitachi, Ltd.Snapshot acquisition method, storage system
US20040019740 *Jun 6, 2003Jan 29, 2004Hitachi, Ltd.Destaging method for storage apparatus system, and disk control apparatus, storage apparatus system and program
US20040022546 *Jul 31, 2002Feb 5, 2004Cochran Robert A.Method and apparatus for compacting data in a communication network
US20040024838 *Jul 31, 2002Feb 5, 2004Cochran Robert A.Intelligent data tunnels multiplexed within communications media directly interconnecting two or more multi-logical-unit-mass-storage devices
US20040024961 *Jul 31, 2002Feb 5, 2004Cochran Robert A.Immediately available, statically allocated, full-logical-unit copy with a transient, snapshot-copy-like intermediate stage
US20040030727 *Aug 6, 2002Feb 12, 2004Philippe ArmangauOrganization of multiple snapshot copies in a data storage system
US20040030846 *Aug 6, 2002Feb 12, 2004Philippe ArmangauData storage system having meta bit maps for indicating whether data blocks are invalid in snapshot copies
US20040049634 *Sep 11, 2003Mar 11, 2004Cochran Robert A.Unified data sets distributed over multiple I/O-device arrays
US20040078638 *Jul 31, 2002Apr 22, 2004Cochran Robert A.Method and system for preventing data loss within disk-array pairs supporting mirrored logical units
US20040078641 *Sep 23, 2002Apr 22, 2004Hewlett-Packard CompanyOperating system-independent file restore from disk image
US20040128404 *Dec 12, 2003Jul 1, 2004Cochran Robert A.Method and system for providing advanced warning to a data stage device in order to decrease the time for a mirror split operation without starving host I/O request processing
US20040168034 *Jul 25, 2003Aug 26, 2004Hitachi, Ltd.Storage apparatus and its management method
US20040215602 *Apr 23, 2003Oct 28, 2004Hewlett-Packard Development Company, L.P.Method and system for distributed remote resources
US20040230859 *May 15, 2003Nov 18, 2004Hewlett-Packard Development Company, L.P.Disaster recovery system with cascaded resynchronization
Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US7464232 *Nov 18, 2005Dec 9, 2008Hitachi, Ltd.Data migration and copying in a storage system with dynamically expansible volumes
US7697554Apr 13, 2010Emc CorporationOn-line data migration of a logical/virtual storage array by replacing virtual names
US7757059Jul 13, 2010Emc CorporationVirtual array non-disruptive management data migration
US7945748May 17, 2011Hitachi, Ltd.Data migration and copying in a storage system with dynamically expansible volumes
US8060792Mar 31, 2009Nov 15, 2011Amazon Technologies, Inc.Monitoring and automated recovery of data instances
US8072987Dec 6, 2011Emc CorporationFull array non-disruptive data migration
US8074107Oct 26, 2009Dec 6, 2011Amazon Technologies, Inc.Failover and recovery for replicated data instances
US8107467Dec 27, 2005Jan 31, 2012Emc CorporationFull array non-disruptive failover
US8127094Apr 28, 2011Feb 28, 2012Hitachi, Ltd.Data migration and copying in a storage system with dynamically expansible volumes
US8307003Nov 6, 2012Amazon Technologies, Inc.Self-service control environment
US8312235Nov 13, 2012Hitachi, Ltd.Data migration in a storage system with dynamically expansible volumes
US8332365 *Mar 31, 2009Dec 11, 2012Amazon Technologies, Inc.Cloning and recovery of data volumes
US8335765Oct 26, 2009Dec 18, 2012Amazon Technologies, Inc.Provisioning and managing replicated data instances
US8452928May 28, 2013Emc CorporationVirtual array non-disruptive migration of extended storage functionality
US8533408Jun 29, 2006Sep 10, 2013Emc CorporationConsolidating N-storage arrays into one storage array using virtual array non-disruptive data migration
US8539177Jun 29, 2006Sep 17, 2013Emc CorporationPartitioning of a storage array into N-storage arrays using virtual array non-disruptive data migration
US8583861Jun 29, 2006Nov 12, 2013Emc CorporationPresentation of management functionality of virtual arrays
US8589504Jun 29, 2006Nov 19, 2013Emc CorporationFull array non-disruptive management data migration
US8595547Nov 18, 2011Nov 26, 2013Amazon Technologies, Inc.Failover and recovery for replicated data instances
US8612396Sep 15, 2012Dec 17, 2013Amazon Technologies, Inc.Cloning and recovery of data volumes
US8620724 *Jun 22, 2010Dec 31, 2013Accenture Global Services LimitedIntegration framework for enterprise content management systems
US8631283Nov 10, 2011Jan 14, 2014Amazon Technologies, Inc.Monitoring and automated recovery of data instances
US8676753Oct 26, 2009Mar 18, 2014Amazon Technologies, Inc.Monitoring of replicated data instances
US8706764Sep 15, 2012Apr 22, 2014Amazon Technologies, Inc.Control service for relational data management
US8713060Mar 31, 2009Apr 29, 2014Amazon Technologies, Inc.Control service for relational data management
US8713061Sep 15, 2012Apr 29, 2014Amazon Technologies, Inc.Self-service administration of a database
US9063895 *Jun 29, 2007Jun 23, 2015Emc CorporationSystem and method of non-disruptive data migration between heterogeneous storage arrays
US9063896 *Jun 29, 2007Jun 23, 2015Emc CorporationSystem and method of non-disruptive data migration between virtual arrays of heterogeneous storage arrays
US9098211Jun 29, 2007Aug 4, 2015Emc CorporationSystem and method of non-disruptive data migration between a full storage array and one or more virtual arrays
US9111220Mar 14, 2014Aug 18, 2015Commvault Systems, Inc.Systems and methods for storage modeling and costing
US9135283Oct 7, 2009Sep 15, 2015Amazon Technologies, Inc.Self-service configuration for data environment
US9152685 *Feb 24, 2014Oct 6, 2015Commvault Systems, Inc.Systems and methods for migrating components in a hierarchical storage network
US9164692May 5, 2014Oct 20, 2015Commvault Systems, Inc.System and method for allocation of organizational resources
US9207984Mar 31, 2009Dec 8, 2015Amazon Technologies, Inc.Monitoring and automatic scaling of data volumes
US9218245 *Oct 25, 2013Dec 22, 2015Amazon Technologies, Inc.Cloning and recovery of data volumes
US9298728Nov 25, 2013Mar 29, 2016Amazon Technologies, Inc.Failover and recovery for replicated data instances
US9336292Sep 15, 2012May 10, 2016Amazon Technologies, Inc.Provisioning and managing replicated data instances
US9348530Dec 27, 2005May 24, 2016Emc CorporationPresentation of virtual arrays using n-port ID virtualization
US20070073988 *Nov 18, 2005Mar 29, 2007Hitachi, Ltd.Data processing system, data management method and storage system
US20070208836 *Dec 27, 2005Sep 6, 2007Emc CorporationPresentation of virtual arrays using n-port ID virtualization
US20070263637 *Dec 27, 2005Nov 15, 2007Emc CorporationOn-line data migration of a logical/virtual storage array
US20090077331 *Nov 6, 2008Mar 19, 2009Hitachi, Ltd.Data migration and copying in a storage system with dynamically expansible volumes
US20100250499 *Sep 30, 2010Mcalister Grant Alexander MacdonaldCloning and Recovery of Data Volumes
US20100250748 *Mar 31, 2009Sep 30, 2010Swaminathan SivasubramanianMonitoring and Automatic Scaling of Data Volumes
US20100251002 *Sep 30, 2010Swaminathan SivasubramanianMonitoring and Automated Recovery of Data Instances
US20100251242 *Sep 30, 2010Swaminathan SivasubramanianControl Service for Relational Data Management
US20110202742 *Aug 18, 2011Hitachi, Ltd.Data migration and copying in a storage system with dynamically expansible volumes
US20110258136 *Jun 22, 2010Oct 20, 2011Accenture Global Services GmbhIntegration framework for enterprise content management systems
US20140172796 *Feb 24, 2014Jun 19, 2014Commvault Systems, Inc.Systems and methods for migrating components in a hierarchical storage network
US20150339197 *Jul 31, 2015Nov 26, 2015Commvault Systems, Inc.Systems and methods for migrating components in a hierarchical storage network
Classifications
U.S. Classification365/189.05, 714/E11.095
International ClassificationG06F11/20, H04L29/14, H04L29/06, H04L29/08
Cooperative ClassificationY10S707/99953, Y10S707/99955, H04L29/06, G06F11/201, H04L2029/06054, H04L69/40, H04L67/1097, H04L67/1034, H04L67/1008, H04L67/101, H04L67/1002
European ClassificationH04L29/08N9A11, H04L29/08N9A1B, H04L29/08N9A1C, H04L29/06, H04L29/08N9S, G06F11/20C4S