Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS8204859 B2
Publication typeGrant
Application numberUS 12/424,115
Publication dateJun 19, 2012
Filing dateApr 15, 2009
Priority dateDec 10, 2008
Also published asCA2746416A1, EP2366151A1, EP2366151B1, US20100145909, US20120317074, WO2010068570A1
Publication number12424115, 424115, US 8204859 B2, US 8204859B2, US-B2-8204859, US8204859 B2, US8204859B2
InventorsDavid Ngo
Original AssigneeCommvault Systems, Inc.
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Systems and methods for managing replicated database data
US 8204859 B2
Abstract
Systems and methods for replicating database data and generating read-only copies of the replicated data in a clean shutdown state. For example, systems can include a tracking module (e.g., a filter driver) that monitors transactions from a database application to a source storage device to generate log entries having at least one marker indicating a known good state of the application. The systems further include a computer coupled to a target storage device comprising a database and log files. The computer processes the transactions, based on the log entries, to replicate data to the target storage device; performs a first snapshot on data stored in the database and log files; replays into the database data stored in the log files; performs another snapshot on the database; and reverts the database back to a state in which the database existed at the time of the first snapshot.
Images(6)
Previous page
Next page
Claims(16)
1. A method for managing replicated data in a database system, the method comprising:
monitoring a plurality of data transactions associated with a database application, the plurality of data transactions operative to write data to at least one source storage device;
copying the data to at least one target storage device based at least in part on the plurality of data transactions, wherein said at least one target storage device comprises a target database and target transaction log files, wherein said monitoring and copying is performed without shutting down the database application;
generating a first snapshot of at least a portion of the data stored in the target database and the target transaction log files, the first snapshot being indicative of stored data from the database application in a dirty shutdown state that prevents the first snapshot from being read or otherwise accessed by applications other than the database application, a stored indication of the dirty shutdown state being associated with the first snapshot;
replaying into the target database, following the first snapshot, data stored in the target transaction log files as one or more transaction logs such that the data stored in the target database following said replaying becomes out of sync with the data stored on the source storage device;
generating, following said replaying, a second snapshot of at least a portion of the target database, the second snapshot being indicative of stored data from the database application in a recoverable and clean shutdown state; and
reverting, following said generating the second snapshot, the target database back to a state in which the target database existed at the time of said generating the first snapshot, wherein the data stored in the target database following said reverting is re-synchronized with the data stored on the source storage device, wherein the second snapshot can be read or otherwise accessed without the use of the database application.
2. The method of claim 1, wherein said generating the first snapshot is performed based on a marker in the copied data indicating a known good state of the database application.
3. The method of claim 2, further comprising:
quiescing the database application at a time of the known good state of the database application; and
inserting the marker in the copied data.
4. The method of claim 1, wherein said copying the data further comprises transmitting between the at least one source storage device and the at least one target storage device a plurality of log entries indicative of the plurality of data transactions associated with the database application.
5. The method of claim 1, further comprising:
moving to a first memory location data overwritten in the target database during said replaying;
moving to a second memory location data added to the target database during said replaying; and
wherein said reverting comprises moving the overwritten data back to the target database following said generating the second snapshot.
6. The method of claim 1, further comprising suspending said copying prior to said generating the first snapshot.
7. The method of claim 6, wherein said suspending is performed at a time at which the data stored on the target database is in a known good state.
8. The method of claim 1, further comprising suspending said copying according to one or more storage policies defined by a user.
9. The method of claim 1, further comprising synchronizing the source database and the target database following said reverting.
10. The method of claim 1, further comprising reading the second snapshot with one or more application programming interfaces while the database application is offline with respect to the data in the second snapshot.
11. A system for performing data management operations in a computer network environment, the system comprising:
a database application configured to execute on a source computer;
a first storage device coupled to the source computer to receive data transactions from the database application;
at least one module configured to monitor the data transactions and to generate a plurality of log entries based on the data transactions, at least one of the plurality of log entries having a marker indicative of a time of a known good state of the database application;
a second storage device comprising a target database and target transaction log files; and
a target computer coupled to the second storage device and configured to:
process, based on the plurality of log entries, the data transactions to replicate data to the second storage device,
perform a first snapshot operation on data stored in both the target database and the target transaction log files to create a first snapshot, the first snapshot being indicative of stored data from the database application in a dirty shutdown state that prevents the first snapshot from being read or otherwise accessed by applications other than the database application, a stored indication of the dirty shutdown state being associated with the first snapshot,
following the first snapshot operation, replay into the target database data stored in the target transaction log files such that the data stored in the target database following the replay becomes out of sync with the data stored on the first storage device,
following said replaying, perform a second snapshot operation on at least a portion of the target database to create a second snapshot, the second snapshot being indicative of stored data from the database application in a recoverable and clean shutdown state, and
revert, following the performance of the second snapshot, the target database back to a state in which the target database existed at the time of said performing the first snapshot wherein the data stored in the target database following said reverting is re-synchronized with the data stored on the first storage device, wherein the second snapshot can be read or otherwise accessed without the use of the database application.
12. The system of claim 11, wherein the target computer further comprises a replication module having a plurality of threads configured to replay the data transactions to replicate the data to the second storage device.
13. The system of claim 11, wherein the target computer comprises at least one application programming interface configured to commit to the target database data stored in the transaction log files.
14. The system of claim 11, wherein the at least one module comprises a filter driver.
15. The system of claim 11, wherein performing the second snapshot operation produces a read-only copy of the target database data, the read-only copy being indicative of database data that can be read without the use of the database application.
16. The system of claim 11, wherein the target computer is further configured to:
store data overwritten in the target database during the replaying of the data stored in the transaction log files into the target database; and
store data added to the target database during the replaying of the data stored in the transaction log files into the target database.
Description
RELATED APPLICATION

This application claims the benefit of priority under 35 U.S.C. §119(e) of U.S. Provisional Patent Application No. 61/121,418, entitled “Systems and Methods for Managing Replicated Database Data,” filed Dec. 10, 2008, which is hereby incorporated herein by reference in its entirety to be considered part of this specification.

BACKGROUND OF THE INVENTION

1. Field of the Invention

The present disclosure relates to systems and methods for replicating data of one or more database applications.

2. Description of the Related Art

Electronic information has become an integral part of business operations such that many banks, insurance companies, brokerage firms, financial service providers, and a variety of other businesses rely on computer networks to store, manipulate, and display information that is constantly subject to change. Oftentimes, the success or failure of an important transaction may turn on the availability of electronic information that is both accurate and current. Accordingly, businesses seek reliable, cost-effective ways to protect and later access the information stored on their computer networks.

Many approaches to protecting data involve creating a copy of the data, such as backing up and/or replicating a database to one or more storage devices. Data shadowing and mirroring, or duplexing, provide for copying but can require lengthy amounts of time, consume valuable processing power and/or occupy large amounts of storage space for large databases. Moreover, such storage management systems can have a significant impact on the source or primary system.

To address these drawbacks, certain storage systems utilize snapshot techniques to preserve a read-only copy of database data. In general, a snapshot records the state of a storage device, database file system, or volume at a certain point in time. That is, the snapshot may be used to provide a point-in-time image of a live storage volume. Additional operations can then be performed using the snapshot copy without affecting the performance of the live volume.

In certain circumstances, however, a snapshot of a database does not necessarily provide easily accessible data. For instance, snapshots of data from a MICROSOFT EXCHANGE database generated while EXCHANGE is online results in a copy of the data that is in a “dirty shutdown” state, which prevents the data from being read or otherwise accessed by standard application programming interfaces (APIs). Although, one option is to shut down EXCHANGE each time a snapshot is to be performed on the database, such an option is not practical as the shutdowns can be time-consuming and costly.

SUMMARY OF THE INVENTION

In view of the foregoing, a need exists for improved systems and methods for database replication. For instance, there is a need for systems and methods for generating copies of database data in a clean shutdown state without taking the native database application offline. Moreover, a need exists for systems and methods that provide snapshots of database data in a useful condition without requiring the native application, or its accompanying APIs, for access to the database data.

In certain embodiments of the invention, systems and methods are disclosed for performing substantially continuous replication of a database, such as a MICROSOFT EXCHANGE database, and for providing usable read-only copies of the replicated database. For instance, snapshot systems and methods are disclosed that provide a snapshot of a replicated MICROSOFT EXCHANGE database that reflects data in a recoverable and clean shutdown state and that can be accessed or otherwise read using standard APIs without the need for the MICROSOFT EXCHANGE application program.

Such systems and methods allow known good replication copies to be viewed as copies of production volume data. For example, this technology, in certain embodiments, further allows a management component in the computing system to directly access, copy, restore, backup or otherwise manipulate the replication copies of production data as if the data was the production data of the source system, thereby improving various system performance characteristics such as access time, reducing memory requirements and reducing impact on source, or client, applications.

In certain embodiments, a method is disclosed for managing replicated data in a database system. The method comprises monitoring data transactions associated with a database application, the data transactions operative to write data to at least one source storage device. The method further includes copying the data to a target storage device based at least in part on the data transactions, wherein the target storage device comprises a target database and target transaction log files, and wherein said monitoring and copying is performed without shutting down the database application. The method also includes: (i) generating a first snapshot of at least a portion of the data stored in the target database and transaction log files; (ii) replaying, into the target database, data stored in the target transaction log files as one or more transaction logs; (iii) generating a second snapshot of at least a portion of the target database that is indicative of stored data from the database application in a recoverable state; and then (iv) reverting the target database back to a state in which the target database existed at the time of the first snapshot.

In certain embodiments, a system is disclosed for performing data management operations in a computer network environment. The system includes a database application configured to execute on a source computer and a first storage device coupled to the source computer to receive data transactions from the database application. The system also includes a module, such as a filter driver, configured to monitor the data transactions and to generate log entries based on the data transactions, at least one of the log entries having a marker indicative of a time of a known good state of the database application. The system further includes a second storage device comprising a target database and target transaction log files and a target computer coupled to the second storage device. The target computer is configured to: (i) process, based on the log entries, the data transactions to replicate data to the second storage device; (ii) perform a first snapshot operation on data stored in both the target database and the target transaction log files; (iii) replay into the target database data stored in the target transaction log files; (iv) perform a second snapshot operation on at least a portion of the target database; and (v) revert the target database back to a state in which the target database existed at the time of the first snapshot.

In certain embodiments, a method is disclosed for copying data generated on a source database system in a computer network. The method comprises processing replication data indicative of data transactions generated by a database application executing on a source system, the database transactions being directed to operations on a source database of a source storage device; replaying the replication data on a target storage device to copy the data transactions to the target storage device, the target storage device comprising a target database and target transaction log files; creating a first read-only copy of at least a portion of data stored in the target database and the target transaction log files; committing to the target database transaction logs stored in the target transaction log files, said committing following said creating the first read-only copy; creating, following said committing, a second read-only copy of at least a portion of the target database, the second read-only copy being indicative of database data from the database application in a recoverable state; and reverting the target database to a state in which the target database existed at the time of said creating the first read-only copy.

In certain embodiments, a system is disclosed for performing data management operations in a computer network environment. The system includes means for processing replication data indicative of database operations generated by a database application executing on a source computer, the database operations being directed to data on a source storage device; means for replaying the replication data on a target storage device to copy data to the target storage device, the target storage device comprising a target database and transaction log files; means for performing a first snapshot operation on data stored in both the target database and the target transaction log files; means for replaying into the target database data stored in the transaction log files; said means for performing being further configured to perform a second snapshot operation on at least a portion of the target database following replaying of the data stored in the transaction log files, the second snapshot being indicative of database data from the database application in a clean shutdown state; and means for reverting the target database back to a state in which the target database existed at the time of performing the first snapshot, wherein at least one of said means is executed on one or more computing devices.

For purposes of summarizing the disclosure, certain aspects, advantages and novel features of the inventions have been described herein. It is to be understood that not necessarily all such advantages may be achieved in accordance with any particular embodiment of the invention. Thus, the invention may be embodied or carried out in a manner that achieves or optimizes one advantage or group of advantages as taught herein without necessarily achieving other advantages as may be taught or suggested herein.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 illustrates a block diagram of a database replication system according to certain embodiments of the invention.

FIG. 2 illustrates a block diagram of an exemplary embodiment of a database replication system configured to provide a snapshot of data in a clean shutdown state.

FIG. 3 illustrates a flowchart of an exemplary embodiment of a database replication process usable by the database replication systems of FIGS. 1 and 2.

FIG. 4 illustrates a flowchart of an exemplary embodiment of a double snapshot process of the data replication process of FIG. 3.

FIG. 5 illustrates a timeline of states of data during the database replication process of FIG. 3, according to certain embodiments of the invention.

DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

Systems and methods disclosed herein provide for the offline access of a replicated database without needing to run the native database application. Moreover, disclosed systems and methods provide for the replication of database data without shutting down the native database application to put the replicated data in a “clean shutdown” state.

In certain embodiments of the invention, systems and methods are disclosed for performing substantially continuous replication of a database, such as a MICROSOFT EXCHANGE database. For instance, certain embodiments provide for the generation of one or more snapshots of a replicated MICROSOFT EXCHANGE database, wherein the snapshot data allows for offline access to the replicated database. For instance, in certain embodiments, the replicated database can be read using JET or like APIs in place of slower APIs (e.g., MAPI) associated with MICROSOFT EXCHANGE. As a result, the direct reading of the database allows for much faster access to the replicated data.

In general, replication of database data can pose particular drawbacks. For example, standard replication of data from certain database applications can result in data that is in a recoverable state, wherein committed transactions in the database can be recovered in case of a crash. In such a state, particular computing operations of the application are complete to a point such that further operation, recovery and/or rolling back of the application can occur, based on the committed transaction data This point of referential integrity is generally referred to herein as a known good state of the application data.

However, without shutting down the database application, the data resulting therefrom is not in a clean shutdown state (i.e., transaction logs exist that are not committed to the database). Moreover, with MICROSOFT EXCHANGE databases, for instance, data in a dirty shutdown state is evidenced by a particular bit value and prevents standard APIs from accessing the data.

The features of the systems and methods will now be described with reference to the drawings summarized above. Throughout the drawings, reference numbers are re-used to indicate correspondence between referenced elements. The drawings, associated descriptions, and specific implementation are provided to illustrate embodiments of the invention and not to limit the scope of the disclosure.

In addition, methods and processes described herein are not limited to any particular sequence, and the blocks or states relating thereto can be performed in other sequences that are appropriate. For example, described blocks or states may be performed in an order other than that specifically disclosed, or multiple blocks or states may be combined in a single block or state.

Moreover, certain embodiments of the invention described herein can utilize one or more intelligent data replication processes and/or components that replicate application-specific data from a source system to a destination system. Certain examples of such processes and components are described in U.S. patent application Ser. No. 11/640,826, filed Dec. 18, 2006, now U.S. Pat. No. 7,651,593, which is hereby incorporated herein by reference in its entirety to be considered part of this specification.

FIG. 1 illustrates a block diagram of a database replication system 100 according to certain embodiments of the invention. As shown, the replication system 100 comprises a source system 102 capable of communicating with a target system 104 by sending and/or receiving data over a network 106. For instance, in certain embodiments, the target system 104 receives and/or stores a replicated copy of at least a portion of data, such as application-specific data, associated with the source system 102.

In certain embodiments, the source system 102 comprises one or more computing devices capable of processing data and can include, for example, a server computer, a workstation, a personal computer, a cell phone, a portable computing device, a handheld computing device, a personal digital assistant (PDA) or the like.

The illustrated network 106 advantageously comprises any means for communicating data between two or more systems or components. It certain embodiments, the network 106 comprises a computer network. For example, the network 106 may comprise a public network such as the Internet, virtual private network (VPN), token ring or TCP/IP based network, wide area network (WAN), local area network (LAN), an intranet network, point-to-point link, a wireless network, cellular network, wireless data transmission system, two-way cable system, interactive kiosk network, satellite network, broadband network, baseband network, combinations of the same or the like. In embodiments wherein the source system 102 and target system 104 are part of the same computing device, the network 106 may represent a communications socket or other suitable internal data transfer path or mechanism.

As shown, the source system 102 comprises one or more applications 108 residing on and/or being executed by a computing device. For instance, the applications 108 may comprise software applications that interact with a user to process data and may include, for example, database applications (e.g., SQL applications), word processors, spreadsheets, financial applications, management applications, e-commerce applications, browsers, combinations of the same or the like.

The source system 102 further comprises one or more processes, such as filter drivers 110, that interact with data (e.g., production data) associated with the applications 108. For instance, the filter driver 110 can comprise a file system filter driver, an operating system driver, a filtering program, a data trapping program, an application, a module of the application(s) 108, an API, or other like software module or process that, among other things, monitors and/or intercepts particular application requests targeted at a database, a file system, another file system filter driver, a network attached storage (“NAS”), a storage area network (“SAN”), mass storage and/or other memory or raw data. In some embodiments, the filter driver 110 may reside in the I/O stack of the application 108 and may intercept, analyze and/or copy certain data traveling from the application 108 to storage.

In certain embodiments, the filter driver 110 can intercept data modification operations that include changes, updates and new information (e.g., data writes) with respect to the application(s) 108 of interest. For example, the filter driver 110 may locate, monitor and/or process one or more of the following with respect to a particular application 108, application type or group of applications: data management operations (e.g., data write operations, file attribute modifications), logs or journals (e.g., NTFS change journal), configuration files, file settings, control files, other files used by the application 108, combinations of the same or the like. In certain embodiments, such data may also be gathered from files across multiple storage systems within the source system 102. Furthermore, the filter driver 110 may be configured to monitor changes to particular files, such as files identified as being associated with data of the applications 108.

In certain embodiments, multiple filter drivers 110 may be deployed on the source system 102, each filter driver 110 being dedicated to data of a particular application 108. In such embodiments, not all information associated with the source system 102 may be captured by the filter drivers 110 and, thus, the impact on system performance may be reduced. In other embodiments, the filter driver 110 may be suitable for use with multiple application types and/or may be adaptable or configurable for use with multiple applications 108.

The illustrated source system 102 further comprises one or more source storage devices 120. In certain embodiments, the source storage device(s) 120 are configured to store production data associated with one or more of the applications 108. The source storage device 120 may include any type of media capable of storing data. For example, the source storage device 120 may comprise magnetic storage (such as a disk or a tape drive) or other type of mass storage. In certain embodiments, the source storage device 120 may be internal and/or external to (e.g., remote to) the computing device(s) having the applications 108 and the filter drivers 110.

As illustrated in FIG. 1, the source storage devices 120 further comprise at least one database 122 and transaction log files 124. In particular, the transaction log files 124 can comprise a set of changes, such as, for example, insertions, deletions and updates received from the application(s) 108 that are to be applied to the data in the database 122.

In certain embodiments, the transaction log files 124 are stored on at least one dedicated disk so that the logs are not affected by any disk failures that can potentially corrupt the database 122. For instance, the transaction log files 124 can be stored on a high-performance disk while the database 122 is stored on one or more slower disks. In other embodiments, the transaction log files 124 and the database 122 can be maintained on the same storage medium.

As further illustrated in FIG. 1, the target system 104 comprises a replication module 128 that communicates with one or more target storage devices 130. In certain embodiments, the target system 104 comprises any computing device capable of processing data and includes, for example, a server computer, a workstation, a personal computer or the like.

In certain embodiments, the replication module 128 is configured to monitor and/or manage the copying of data from the source system 102 to the target system 104, such as data obtained by the filter drivers 110. For example, the replication module can receive one or more log files or entries, data transactions, or other like replication data indicative of the data transactions or operations being generated by the application 108 to modify data stored on the source storage device(s) 120. In yet other embodiments, the replication module 128 is a “dumb” server or terminal that receives and executes instructions from the source system 102.

The target storage device(s) 130 may include any type of media capable of storing data, such as replication data sent from the source system 102. For example, the target storage device 130 may comprise magnetic storage (such as a disk or a tape drive) or other type of mass storage. In certain embodiments, the target storage device 130 may be internal and/or external to the computing device(s) having the replication module 128.

In certain embodiments, the source storage device 120 and/or the target storage device 130 may be implemented as one or more storage “volumes” that include physical storage disks defining an overall logical arrangement of storage space. For instance, disks within a particular volume may be organized as one or more groups of redundant array of independent (or inexpensive) disks (RAID). In certain embodiments, either or both of the storage devices 120, 130 can include multiple storage devices of the same or different media.

As shown in FIG. 1, the target storage device(s) 130 at least comprise a database 132 and transaction log files 134. In particular, the target database 132 is generally synchronized with the source database 122. Likewise, the target transaction log files 134 can comprise a replication of those log files present in the transaction log files 124 of the source storage device 120. As discussed above, in certain embodiments, the database 132 and the transaction log files 134 are advantageously maintained on separate media such that the transaction log files 134 can be stored on a relatively high-performance storage medium. In yet other embodiments, the database 132 and transaction log files 134 are stored in different volumes of the same disk.

As further shown, the replication system 100 is structured to generate read-only copies 136 of the replicated data stored in the target storage device(s) 130. For instance, such copies 136 can be generated according to one or more schedules, storage policies (e.g., user- or system-defined), or the like. In certain embodiments, the replication module 128 coordinates the generation of the copies 136. For example, the replication module 128 can comprise a data agent module executing on the replication system 100 that manages read-only copies (e.g., snapshots) of the replicated data. In yet other embodiments, a storage manager, a stand-alone application, or the like, manages the generation of the read-only copies 136.

In certain embodiments, the read-only copies 136 advantageously comprise a plurality of snapshots. For instance, the snapshots can reflect point-in-time copies of the database data. In certain embodiments, the snapshots allow for access to and/or manipulation of database data without affecting the production data on the source storage device(s) 120. Moreover, in certain embodiments, the snapshots record database data in a clean shutdown mode and allow for offline access of the data without the need for the native database application 108.

In certain embodiments, the source system 102 communicates with the associated target system 104 to verify that the two systems are synchronized. For instance, the source system 102 may receive from the target system 104 an identification (e.g., unique serial number) of a transaction log last committed by the target system 104. The source system 102 may then compare the received identification with state of the data in the source storage device(s) 120.

FIG. 2 illustrates further details of a replication system 200 in accordance with certain embodiments of the invention. In particular, FIG. 2 illustrates additional details of an embodiment of the replication system 100 of FIG. 1. Thus, for ease of reference and description, elements of the replication system 200 of FIG. 2 will not be redescribed in detail if they were described above. Rather, the elements in the embodiment of FIG. 2 will be given a reference numeral that retains the same last two digits as the reference numeral used in the embodiment of FIG. 1, and the last two digits will be preceded with a numeral “2.” Thus, the replication system 200 generally corresponds to the replication system 100 with certain differences that will be illuminated in the following discussion.

As shown, the replication system 200 comprises a source system 202 that communicates with a target system 204 via a network 206. In particular, the target system 204 is structured to replicate database data originating from and/or stored by the source system 202.

The illustrated source system 202 comprises a database application 208 that interacts with one or more users or programs to generate production data. For instance, the database application 208 can comprise one or more applications that require transactions to be played or committed in order for the data of the application 208 to be at consistent point or state.

For exemplary purposes, the database application 208 will often be described hereinafter with reference to a MICROSOFT EXCHANGE environment. In yet other embodiments, other types of database applications and/or collaboration applications can be used, such as, for example, ACTIVE DIRECTORY, ORACLE, other JET-based applications or the like.

With continued reference to the source system 202, a filter driver 210 tracks file changes being generated by the application 208. For example, the filter driver 210 can be configured to populate data relating to transactions of the database application 208 in source logs to be replicated to the target system 204.

In certain embodiments, the filter driver 210 may be deployed in the stack as an I/O buffer and/or process in the data path between the database application 208 and a memory 212. In such embodiments, the filter driver 210 may intercept, snoop, supervise, trap, process or otherwise be cognizant of some or all operations (e.g., data modification operations, file modification operations, read operations and the like) from the database application 208 to its associated location(s) on the source storage device(s) 220.

In embodiments of the invention having multiple database applications 208, multiple filter drivers 210 can be used such that each filter driver 210 corresponds to a single database application 208. In such embodiments, data relating to each database application 208 of interest can be written to a particular log file established for that application. In yet other embodiments, a single filter driver 210 can communicate with multiple database applications 208.

In certain embodiments, the transactions from the database application 208 are preferably stored in the memory 212 after or while being logged by the filter driver 210. A database engine 214 then commits the transactions stored in the memory 210 to disk (source storage device(s) 220).

The source storage device(s) 220 comprise one or more media for storing production data relating to the database application 208. As shown, the source storage device 220 includes a database or information store 222 for storing the committed transaction data from the database application 208. The source storage device(s) 220 further include transaction log files 224 that represent a sequence of files to keep a secure copy on disk of the data (e.g., uncommitted transaction logs) in the memory 212.

As described above with reference to FIG. 1, the replication system 200 is configured to replicate data, such as through a continuous data replication process, to target storage device(s) 230 of the target system 204. To facilitate and/or coordinate such replication, the target system 204 includes a replication module 228.

In certain embodiments, the replication module 228 comprises and/or communicates with one or more replay threads, processes or routines that populate data to the target storage device(s) 230 during the replication process. For instance, the replication module 228 can replay data from log files that are received from the source system 202. In certain embodiments, the target storage device(s) 230 comprise one or more databases 232 and transaction log files 234 that store a copy of the data residing in, respectively, the source database(s) 222 and the transaction log files 224.

The target system 204 is further configured to generate one or more snapshots 236 of data stored in the target storage device(s) 230. In certain embodiments, the snapshot 236 comprises a read-only copy of the data stored in the database 232 and the transaction log files 234, as discussed in more detail above.

In certain embodiments, the replication system 200 can optionally include one or more additional components for further coordinating and/or facilitating replication between the source system 202 and the target system 204. For instance, FIG. 2 further illustrates the source system 202 comprising an optional data agent 216.

In certain embodiments, the data agent 216 comprises a module responsible for performing data and/or storage tasks related to the source system 202. For example, the data agent 216 may manage and/or coordinate the compilation of and/or transferring of replication data from the source system 202. In other embodiments, the data agent 216 may provide archiving, migrating, and/or recovery of system data.

In certain embodiments, the source system 202 comprises a plurality of data agents 216, each of which performs data management operations related to data associated with each database application 208. In such embodiments, the data agent 216 may be aware of the various logs, files, folders, registry files and/or system resources that are impacted by a particular database application 208.

In certain embodiments, the data agent 216 is configured to perform data management operations in accordance with one or more “storage policies” or other preferences. In certain embodiments, a storage policy includes a data structure or other information having a set of preferences and/or other storage criteria for performing a storage operation. The preferences and storage criteria may include, but are not limited to, information regarding storage locations and/or timing, relationships between system components, network pathways, retention policies, data characteristics, compression or encryption requirements, preferred system components, combinations of the same or the like.

In certain further embodiments, the replication system 200 can optionally include a storage manager 238 that communicates with the source system 202 and the target system 204. In certain embodiments, the storage manager 238 directs the performance of one or more storage operations and, in particular, the replication of data from the source system 202 to the target system 204. In further embodiments, the storage manager 238 may perform one or more of the operations or functions described above with respect to the data agent 216 and/or the replication module 228. For instance, the storage manager 238 may direct and/or coordinate the performance of one or more storage operations on the replicated data (e.g., snapshots of the replicated data) of the target storage device(s) 230.

In certain embodiments, the storage manager 238 maintains an index (not shown), such as a cache, for storing information relating to: logical relationships and associations between components of the replication system 200, user preferences, management tasks, and/or other useful data. For example, the storage manager 238 may use its index to track the location and timestamps of one or more snapshots of the replicated data.

The storage manager 238 may also use its index to track the status of data management operations to be performed, storage patterns associated with the system components such as media use, storage growth, network bandwidth, Service Level Agreement (“SLA”) compliance levels, data protection levels, storage policy information, storage criteria associated with user preferences, retention criteria, storage operation preferences, and/or other storage-related information. The index may typically reside on the storage manager's hard disk and/or other database.

In certain embodiments, the storage manager 238 further communicates with a database (not shown) for storing system management information relating to the replication of data. For instance, the storage manager database may be configured to store storage and/or restore policies, user preferences, the status or location of system components or data, combinations of the same and the like. In yet other embodiments, the storage manager database may be configured to store information described above with respect to the storage manager index. In yet other embodiments, at least a portion of the storage manager index may be stored on the storage manager database.

In other embodiments, the storage manager 238 may alert the user or system when a particular resource of the replication system 200 is unavailable or congested or when components are unavailable due to hardware failure, software problems, or other reasons. In certain embodiments, the storage manager 238 may utilize replication system data to suggest solutions to such problems when they occur or even before they occur. For example, the storage manager 238 might alert the user that a storage device in the target system 204 is full or otherwise congested, and then suggest, based on job and data storage information contained in its index cache, an alternate storage device. In yet further embodiments, the storage manager 238 or other system component may take action to remedy the problem at issue. For example, the storage manager 238 may perform load balancing, error correction, or the like, based on information received regarding the target system 204.

In certain embodiments, the storage manager 238 may include other components and/or modules. For example, the storage manager 238 may include a jobs agent module (not shown) that monitors the status of storage operations that have been performed, that are being performed, or that are scheduled to be performed in the replication system 200.

Moreover, the storage manager 238 may include an interface agent module (not shown). In certain embodiments, the interface agent module may provide presentation logic, such as a graphical user interface (“GUI”), an API, or other interface by which users and system processes may be able to retrieve information about the status of storage operations and issue instructions to the replication system 200 regarding the performance of storage operations. For example, a user may modify the schedule of a number of pending snapshot copies or other types of copies. As another example, a user may use the GUI to view the status of all storage operations currently pending in the replication system 200 or the status of particular components in the replication system 200.

Additional details of storage manager modules useful with embodiments of the replication systems 100, 200 described herein are disclosed in U.S. patent application Ser. No. 09/354,063, filed Jul. 15, 1999, now U.S. Pat. No. 7,389,311, which is hereby incorporated herein by reference in its entirety.

FIG. 3 illustrates a flowchart of an exemplary embodiment of a database replication process 300. In certain embodiments, the process 300 creates a crash consistent copy of database data in a clean shutdown state such that the data can be read or otherwise utilized without needing the native database application or its accompanying APIs. In certain embodiments, the replication process 300 can be performed by the database replication systems 100, 200 of FIGS. 1 and 2. For exemplary purposes, the replication process 300 will be described hereinafter with respect to the components of the replication system 200 of FIG. 2.

The replication process 300 begins at Block 305, wherein the database application 208 is quiesced. For example, in certain embodiments, the Volume Shadow Copy Service (VSS) offered by MICROSOFT can be used to temporarily quiesce write requests from the database application 208, such as MICROSOFT EXCHANGE, to the source storage device 220. For instance, in such embodiments, the data agent 216 can insert a marker in the application data (e.g., transaction logs) that indicates the data is in a recoverable (e.g., crash consistent, stable) state.

At Block 310, the replication module 228 suspends the replay threads on the target system 204 that are replicating the data to the target storage device 230. In certain embodiments, such suspension occurs based on the threads' detection of the markers inserted by the VSS service. At this point, the target storage device(s) 230 and the source storage device(s) 220 are preferably synchronized and/or include data that is in a crash recoverable state, although the data is not in a “clean shutdown” state.

With the replay threads suspended, the replication module 228 initiates a first snapshot of the target volume, including the database 232 and the transaction log files 234 (Block 315). In certain embodiments, the first snapshot includes a read-only copy of the database application data in a recoverable state. However, as discussed above, this data is generally not in a clean shutdown state and does not allow for access of the data without the use of the native application (e.g., MICROSOFT EXCHANGE) to interpret the data.

At Block 320, the copy of the uncommitted transaction logs stored in the target transaction log files 234 is replayed into the target database 232. In certain embodiments, such replaying is performed through the execution of one or more JET APIs. After this point, the copy of the database application 208 data in the target database 232 is advantageously in a clean shutdown state since the pending transaction logs have been committed. That is, the copy of the data in the database 232 is in a state that can be read offline without the use of the native database application 208.

At Block 325, the replication module 228 initiates a second snapshot of the target volume. Because the target transaction logs have been replayed into the target database 232, the second snapshot advantageously comprises a read-only copy of application data in both a recoverable and clean shutdown state.

However, with the replaying of the copy of the transaction logs at Block 320, the data on the target storage device(s) 230 becomes out of sync with the production data stored on the source storage device(s) 220. Thus, prior to resuming replication between the source storage device(s) 220 and the target storage device(s) 230, the replication process 300 reverts the target volume back to the state of the first snapshot (Block 330). In certain embodiments, this revert process is performed by taking a difference (e.g., file changes that have been cached) of the first and second snapshots, as is discussed in more detail with respect to FIG. 4.

Following Block 330, the replication module 228 thaws the replay threads and allows the replication between the source and target storage devices 220, 230 to resume (Block 335).

Although the database replication process 300 has been described with reference to a particular arrangement, other embodiments of the invention can include more, fewer or different blocks or states. For instance, in other embodiments, instead of snapshots, other types of read-only copies can be performed on the data in the target storage device(s) 230.

In yet other embodiments, at Block 305, the data agent 216 and/or the filter driver 210 can be advantageously configured to pause, or quiesce, the database application 208 during data replication. For instance, the data agent 216 may cause the application 208 to temporarily suspend data management operations to the source storage device 220 once the application 208 reaches a known “good,” “stable” or “recoverable” state.

In certain embodiments, the data agent 216 instructs the quiescing of the application 208 through an application programming interface (API). When the application 208 has placed itself in a known good state, the application 208 may send an acknowledgment to the data agent 216.

In certain embodiments, once the data management operations are suspended, the filter driver 210 and/or data agent 216 then inserts a logical marker or tag in the source log file denoting that a “consistency point” or “consistency recovery point” has been reached. In some embodiments, the consistency point indicates the time at which the application 208 is at a known good state.

Moreover, in certain embodiments, the target system 204 is further capable of performing one or more data management operations, such as, for example, storage operations (e.g., backup), search operations, data classification, combinations of the same or the like on the replicated data at certain consistency points. Performing data management operations on the replicated data allows for the processing of copies of application data without significantly impacting the resources of the source system. Furthermore, when copying the replicated data at consistency points, the copied data presumably represents a known good state of the application.

In certain embodiments of the invention, at Block 305 the application 208 is periodically quiesced based on particular criteria. For instance, the quiescing of the application 208 may be based on one or more system- or user-defined preferences (e.g., every five minutes). The periodic quiescing of the application 208 may be based on the desired frequency of performing replication, backup or other data modification operations on the subject data. For instance, applications 208 dealing with data-sensitive information may necessitate more frequent quiescing (and creation of consistency points) than other types of applications.

In yet other embodiments, quiescing of the application 108 may be performed based on an automatic reporting procedure. For instance, a module of the replication system 200 may be configured to gather, receive and/or analyze information associated with a failure rate and/or health of applicable servers. Additional details of such status monitoring are provided in U.S. patent application Ser. No. 11/120,619, filed May 2, 2005, now U.S. Pat. No. 7,343,453, which is hereby incorporated herein by reference in its entirety.

FIG. 4 illustrates a flowchart of an exemplary embodiment of a double snapshot process 400 that may be used when performing certain blocks (e.g., Blocks 315 to 330) of the data replication process 300 of FIG. 3. For exemplary purposes, the double snapshot process 400 will be described hereinafter with respect to the components of the replication system 200 of FIG. 2.

At Block 405, the process 400 performs a first snapshot of the target volume, including the target database 232 and the target transaction log files 234. Following the first snapshot, the transaction logs are replayed into the target database 232 to put the copy of the data in a clean shutdown state. During the replay of the transaction logs (e.g., by a JET API), the overwritten bits in the target database 232 can be moved to a first cache location (Block 410).

The process 400 then performs a second snapshot of the target volume, wherein the target volume represents the application data in a clean shutdown state (Block 415). Following the second snapshot, the process 400 reverts the target volume back to the state in which it existed at the time of the first snapshot. To do so, at Block 420, the new bits added between the first and second snapshots (i.e., during the replaying of the transaction logs) are moved to a second cache location (Block 425). The overwritten bits stored in the first cache location are then moved back to the target volume (Block 425).

At Block 430, the second snapshot is preserved as a consistent recovery point in a clean shutdown state. In certain embodiments, a customized driver is used to preserve the snapshot (e.g., the snapshot is not deleted).

In certain embodiments, the first and second cache locations can be on different storage devices or volumes. In yet other embodiments, the first and second cache locations can be located on the same storage device.

FIG. 5 illustrates an exemplary timeline 500 of states of data during the database replication process 300 of FIG. 3, according to certain embodiments of the invention. In particular, the timeline 500 includes a simplified representation of the states of files on a source system (e.g., source storage device 220) and a target system (e.g., target storage device 230) during replication.

As shown, at State A, the source database includes File A and File B and the source log files include a transaction relating to the creation of File C and the modification of File B (i.e., File B′). With reference to the target system, the target database and the target log files are synchronized with the source system.

At State B, a first snapshot (i.e., Snapshot A) is taken of the target system. As discussed above with reference to Block 315 of the replication process 300, the application data from the target database is in a dirty shutdown state.

At State C, an additional transaction log for the creation of a File D has been added to the source log files. Moreover, the target transaction logs (Files B′ and C) have been replayed from the target log files to the target database (see, e.g., Block 320). Thus, the target database includes File A, File B′ and File C. At this point, the data in the target database is in a clean shutdown state, and no outstanding transaction logs are left in the target log files. As further illustrated, the system then takes a second snapshot (Snapshot B) of at least the data in the target database to create a consistent, recoverable copy of the database application data that can be accessed and utilized without the native database application which created the data (see, e.g., Block 325).

At State D, the target database is reverted back to the state in which it existed at the time Snapshot A was taken (see, e.g., Block 330), and the target log files are repopulated with transactions relating to Files B′ and C. Moreover, at State E, a transaction log relating to File D is replicated from the source system to the target system.

At State F, the transaction logs in both the source and target log files relating to File B′ and File C are replayed, respectively, into the source and target databases. Moreover, an additional transaction log relating to File E is added to the source log files and replicated to the target system such that the source and target systems are re-synchronized.

In certain embodiments of the invention, data replication systems and methods disclosed herein may be used in a modular storage management system, embodiments of which are described in more detail in U.S. Pat. No. 7,035,880, issued Apr. 25, 2006, which is hereby incorporated herein by reference in its entirety. For example, the data replication system may be part of a storage operation cell that includes combinations of hardware and/or software components directed to performing storage operations on electronic data. Exemplary storage operation cells usable with embodiments of the invention include CommCells as embodied in the QNet storage management system and the QiNetix storage management system by CommVault Systems, Inc. (Oceanport, N.J.), and as further described in U.S. patent application Ser. No. 10/877,831, filed Jun. 25, 2004, now U.S. Pat. No. 7,454,569, which is hereby incorporated herein by reference in its entirety.

Systems and modules described herein may comprise software, firmware, hardware, or any combination(s) of software, firmware, or hardware suitable for the purposes described herein. Software and other modules may reside on servers, workstations, personal computers, computerized tablets, PDAs, and other devices suitable for the purposes described herein. Software and other modules may be accessible via local memory, via a network, via a browser, or via other means suitable for the purposes described herein. Data structures described herein may comprise computer files, variables, programming arrays, programming structures, or any electronic information storage schemes or methods, or any combinations thereof, suitable for the purposes described herein. User interface elements described herein may comprise elements from graphical user interfaces, command line interfaces, and other interfaces suitable for the purposes described herein.

Embodiments of the invention are also described above with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, may be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the acts specified in the flowchart and/or block diagram block or blocks.

These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to operate in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the acts specified in the flowchart and/or block diagram block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operations to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the acts specified in the flowchart and/or block diagram block or blocks.

While certain embodiments of the inventions have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the disclosure. Indeed, the novel methods and systems described herein may be embodied in a variety of other forms; furthermore, various omissions, substitutions and changes in the form of the methods and systems described herein may be made without departing from the spirit of the disclosure. The accompanying claims and their equivalents are intended to cover such forms or modifications as would fall within the scope and spirit of the disclosure.

Patent Citations
Cited PatentFiling datePublication dateApplicantTitle
US4296465Nov 3, 1977Oct 20, 1981Honeywell Information Systems Inc.Data mover
US4686620Jul 26, 1984Aug 11, 1987American Telephone And Telegraph Company, At&T Bell LaboratoriesDatabase backup method
US4995035Oct 31, 1988Feb 19, 1991International Business Machines CorporationCentralized management in a computer network
US5005122Sep 8, 1987Apr 2, 1991Digital Equipment CorporationArrangement with cooperating management server node and network service node
US5093912Jun 26, 1989Mar 3, 1992International Business Machines CorporationDynamic resource pool expansion and contraction in multiprocessing environments
US5133065Jul 27, 1989Jul 21, 1992Personal Computer Peripherals CorporationBackup computer program for networks
US5193154Oct 25, 1991Mar 9, 1993Hitachi, Ltd.Buffered peripheral system and method for backing up and retrieving data to and from backup memory device
US5212772Feb 11, 1991May 18, 1993Gigatrend IncorporatedSystem for storing data in backup tape device
US5226157Mar 2, 1989Jul 6, 1993Hitachi, Ltd.Backup control method and system in data processing system using identifiers for controlling block data transfer
US5239647Sep 7, 1990Aug 24, 1993International Business Machines CorporationData storage hierarchy with shared storage level
US5241668Apr 20, 1992Aug 31, 1993International Business Machines CorporationMethod and system for automated termination and resumption in a time zero backup copy process
US5241670Apr 20, 1992Aug 31, 1993International Business Machines CorporationMethod and system for automated backup copy ordering in a time zero backup copy session
US5263154Apr 20, 1992Nov 16, 1993International Business Machines CorporationMethod and system for incremental time zero backup copying of data
US5265159Jun 23, 1992Nov 23, 1993Hughes Aircraft CompanySecure file erasure
US5276860Dec 19, 1989Jan 4, 1994Epoch Systems, Inc.Digital data processor with improved backup storage
US5276867Dec 19, 1989Jan 4, 1994Epoch Systems, Inc.Digital data storage system with improved data migration
US5287500Jun 3, 1991Feb 15, 1994Digital Equipment CorporationSystem for allocating storage spaces based upon required and optional service attributes having assigned piorities
US5301351Jan 13, 1993Apr 5, 1994Nec CorporationData transfer control system between high speed main memory and input/output processor with a data mover
US5311509Feb 6, 1992May 10, 1994International Business Machines CorporationConfigurable gigabits switch adapter
US5317731Feb 25, 1991May 31, 1994International Business Machines CorporationIntelligent page store for concurrent and consistent access to a database by a transaction processor and a query processor
US5321816Dec 9, 1992Jun 14, 1994Unisys CorporationLocal-remote apparatus with specialized image storage modules
US5333315Jun 27, 1991Jul 26, 1994Digital Equipment CorporationSystem of device independent file directories using a tag between the directories and file descriptors that migrate with the files
US5347653Jun 28, 1991Sep 13, 1994Digital Equipment CorporationSystem for reconstructing prior versions of indexes using records indicating changes between successive versions of the indexes
US5369757Jun 18, 1991Nov 29, 1994Digital Equipment CorporationRecovery logging in the presence of snapshot files by ordering of buffer pool flushing
US5403639Sep 2, 1992Apr 4, 1995Storage Technology CorporationFile server having snapshot application data groups
US5410700Sep 4, 1991Apr 25, 1995International Business Machines CorporationComputer system which supports asynchronous commitment of data
US5448724Mar 17, 1994Sep 5, 1995Fujitsu LimitedData processing system having double supervising functions
US5455926Jun 20, 1994Oct 3, 1995Data/Ware Development, Inc.Virtual addressing of optical storage media as magnetic tape equivalents
US5487072Jun 30, 1994Jan 23, 1996Bell Communications Research Inc.Error monitoring algorithm for broadband signaling
US5491810Mar 1, 1994Feb 13, 1996International Business Machines CorporationMethod and system for automated data storage system space allocation utilizing prioritized data set parameters
US5495607Nov 15, 1993Feb 27, 1996Conner Peripherals, Inc.Network management system having virtual catalog overview of files distributively stored across network domain
US5504873Nov 12, 1993Apr 2, 1996E-Systems, Inc.Mass data storage and retrieval system
US5544345Nov 8, 1993Aug 6, 1996International Business Machines CorporationCoherence controls for store-multiple shared data coordinated by cache directory entries in a shared electronic storage
US5544347Apr 23, 1993Aug 6, 1996Emc CorporationData storage system controlled remote data mirroring with respectively maintained data indices
US5546536Mar 23, 1995Aug 13, 1996Digital Equipment CorporationLog for selective management of specific address in a shadow storage system
US5555404May 26, 1995Sep 10, 1996Telenor AsContinuously available database server having multiple groups of nodes with minimum intersecting sets of database fragment replicas
US5559957May 31, 1995Sep 24, 1996Lucent Technologies Inc.File system for a data storage device having a power fail recovery mechanism for write/replace operations
US5559991Jun 27, 1995Sep 24, 1996Lucent Technologies Inc.Incremental computer file backup using check words
US5598546Aug 31, 1994Jan 28, 1997Exponential Technology, Inc.Dual-architecture super-scalar pipeline
US5604862Mar 14, 1995Feb 18, 1997Network Integrity, Inc.Continuously-snapshotted protection of computer files
US5615392May 5, 1995Mar 25, 1997Apple Computer, Inc.Method and apparatus for consolidated buffer handling for computer device input/output
US5619644Sep 18, 1995Apr 8, 1997International Business Machines CorporationSoftware directed microcode state save for distributed storage controller
US5638509Jun 13, 1996Jun 10, 1997Exabyte CorporationData storage and protection system
US5642496Sep 23, 1993Jun 24, 1997Kanfi; ArnonMethod of making a backup copy of a memory over a plurality of copying sessions
US5673381Jan 25, 1996Sep 30, 1997Cheyenne Software International Sales Corp.System and parallel streaming and data stripping to back-up a network
US5675511Dec 21, 1995Oct 7, 1997Intel CorporationApparatus and method for event tagging for multiple audio, video, and data streams
US5677900Apr 15, 1991Oct 14, 1997Sharp Kabushiki KaishaMethod and apparatus for replacing a selected file with another longer or shorter file with no portion of the selected file remaining
US5682513Mar 31, 1995Oct 28, 1997International Business Machines CorporationCache queue entry linking for DASD record updates
US5687343Jun 6, 1995Nov 11, 1997International Business Machines CorporationProduct for global updating modified data object represented in concatenated multiple virtual space by segment mapping
US5689706Oct 10, 1996Nov 18, 1997Lucent Technologies Inc.Distributed systems with replicated files
US5699361Jul 18, 1995Dec 16, 1997Industrial Technology Research InstituteMultimedia channel formulation mechanism
US5719786Feb 3, 1993Feb 17, 1998Novell, Inc.Digital media data stream network management system
US5720026May 7, 1996Feb 17, 1998Mitsubishi Denki Kabushiki KaishaIncremental backup system
US5729743Nov 30, 1995Mar 17, 1998Deltatech Research, Inc.Computer apparatus and method for merging system deltas
US5737747Jun 10, 1996Apr 7, 1998Emc CorporationPrefetching to service multiple video streams from an integrated cached disk array
US5742792May 28, 1996Apr 21, 1998Emc CorporationRemote data mirroring
US5751997Jan 19, 1996May 12, 1998Apple Computer, Inc.Method and apparatus for transferring archival data among an arbitrarily large number of computer devices in a networked computer environment
US5758359Oct 24, 1996May 26, 1998Digital Equipment CorporationMethod and apparatus for performing retroactive backups in a computer system
US5761677Jan 3, 1996Jun 2, 1998Sun Microsystems, Inc.Computer system method and apparatus providing for various versions of a file without requiring data copy or log operations
US5761734Aug 13, 1996Jun 2, 1998International Business Machines CorporationToken-based serialisation of instructions in a multiprocessor system
US5764972Jun 7, 1995Jun 9, 1998Lsc, Inc.Archiving file system for data servers in a distributed network environment
US5765173Jan 11, 1996Jun 9, 1998Connected CorporationHigh performance backup via selective file saving which can perform incremental backups and exclude files and uses a changed block signature list
US5778395Oct 23, 1995Jul 7, 1998Stac, Inc.System for backing up files from disk volumes on multiple nodes of a computer network
US5790114Oct 4, 1996Aug 4, 1998Microtouch Systems, Inc.Electronic whiteboard with multi-functional user interface
US5790828Sep 30, 1996Aug 4, 1998Southwestern Bell Technology Resources, Inc.Disk meshing and flexible storage mapping with enhanced flexible caching
US5805920Nov 13, 1995Sep 8, 1998Tandem Computers IncorporatedDirect bulk data transfers
US5812398Jun 10, 1996Sep 22, 1998Sun Microsystems, Inc.Method and system for escrowed backup of hotelled world wide web sites
US5813009Jul 28, 1995Sep 22, 1998Univirtual Corp.Computer based records management system method
US5813017Sep 3, 1996Sep 22, 1998International Business Machines CorporationSystem and method for reducing storage requirement in backup subsystems utilizing segmented compression and differencing
US5829046Jun 10, 1996Oct 27, 1998Emc CorporationOn-line tape backup using an integrated cached disk array
US5860104Aug 31, 1995Jan 12, 1999Advanced Micro Devices, Inc.Data cache which speculatively updates a predicted data cache storage location with store data and subsequently corrects mispredicted updates
US5875478Dec 3, 1996Feb 23, 1999Emc CorporationComputer backup using a file system, network, disk, tape and remote archiving repository media system
US5875481Jan 30, 1997Feb 23, 1999International Business Machines CorporationDynamic reconfiguration of data storage devices to balance recycle throughput
US5878408Dec 6, 1996Mar 2, 1999International Business Machines CorporationData management system and process
US5887134Jun 30, 1997Mar 23, 1999Sun MicrosystemsSystem and method for preserving message order while employing both programmed I/O and DMA operations
US5901327Mar 17, 1997May 4, 1999Emc CorporationBundling of write data from channel commands in a command chain for transmission over a data link between data storage systems for remote data mirroring
US5907621Nov 15, 1996May 25, 1999International Business Machines CorporationSystem and method for session management
US5907672Oct 4, 1995May 25, 1999Stac, Inc.System for backing up computer disk volumes with error remapping of flawed memory addresses
US5924102May 7, 1997Jul 13, 1999International Business Machines CorporationSystem and method for managing critical files
US5926836Dec 3, 1996Jul 20, 1999Emc CorporationComputer and associated method for restoring data backed up on archive media
US5933104Nov 22, 1995Aug 3, 1999Microsoft CorporationMethod and system for compression and decompression using variable-sized offset and length fields
US5933601Sep 30, 1996Aug 3, 1999Ncr CorporationMethod for systems management of object-based computer networks
US5950205Sep 25, 1997Sep 7, 1999Cisco Technology, Inc.Data transmission over the internet using a cache memory file system
US5956519May 1, 1997Sep 21, 1999Discovision AssociatesPicture end token in a system comprising a plurality of pipeline stages
US5958005Jul 17, 1997Sep 28, 1999Bell Atlantic Network Services, Inc.Electronic mail security
US5970233May 3, 1996Oct 19, 1999Intel CorporationMultiple codecs for video encoding format compatibility
US5970255Oct 16, 1995Oct 19, 1999Altera CorporationSystem for coupling programmable logic device to external circuitry which selects a logic standard and uses buffers to modify output and input signals accordingly
US5974563Oct 2, 1998Oct 26, 1999Network Specialists, Inc.Real time backup system
US5987478Oct 31, 1995Nov 16, 1999Intel CorporationVirtual small block file manager for flash memory array
US5991779Sep 14, 1998Nov 23, 1999Electric CommunitiesProcess for distributed garbage collection
US5995091May 10, 1996Nov 30, 1999Learn2.Com, Inc.System and method for streaming multimedia data
US6003089Mar 31, 1997Dec 14, 1999Siemens Information And Communication Networks, Inc.Method for constructing adaptive packet lengths in a congested network
US6009274Jun 24, 1997Dec 28, 19993Com CorporationMethod and apparatus for automatically updating software components on end systems over a network
US6012090Mar 14, 1997Jan 4, 2000At&T Corp.Client-side parallel requests for network services using group name association
US6021415Oct 29, 1997Feb 1, 2000International Business Machines CorporationStorage management system with file aggregation and space reclamation within aggregated files
US6021475Feb 19, 1998Feb 1, 2000International Business Machines CorporationMethod and apparatus for polling and selecting any paired device in any drawer
US6023710Dec 23, 1997Feb 8, 2000Microsoft CorporationSystem and method for long-term administration of archival storage
US6026414Mar 5, 1998Feb 15, 2000International Business Machines CorporationSystem including a proxy client to backup files in a distributed computing environment
US6049889Jan 13, 1998Apr 11, 2000Digital Equipment CorporationHigh performance recoverable communication method and apparatus for write-only networks
US20040267835 *Apr 28, 2004Dec 30, 2004Microsoft CorporationDatabase data recovery system and method
Non-Patent Citations
Reference
1Armstead et al., "Implementation of a Campus-wide Distributed Mass Storage Service: The Dream vs. Reality," IEEE, 1995, pp. 190-199.
2Arneson, "Mass Storage Archiving in Network Environments" IEEE, 1998, pp. 45-50.
3Arneson, David A., Control Data Corporation, Development of Omniserver; Mass Storage Systems, 1990, pp. 88-93.
4Ashton, et al., "Two Decades of policy-based storage management for the IBM mainframe computer", www.research.ibm.com, 19 pages, published Apr. 10, 2003, printed Jan. 3, 2009-cited in U.S. Appl. No. 12/276,868., www.research.ibm.com, Apr. 10, 2003, pp. 19.
5Ashton, et al., "Two Decades of policy-based storage management for the IBM mainframe computer", www.research.ibm.com, 19 pages, published Apr. 10, 2003, printed Jan. 3, 2009—cited in U.S. Appl. No. 12/276,868., www.research.ibm.com, Apr. 10, 2003, pp. 19.
6Canadian Office Action dated Dec. 29, 2010, Application No. CA2546304.
7Eitel, "Backup and Storage Management in Distributed Heterogeneous Environments," IEEE, 1994, pp. 124-126.
8European Examination Report; Application No. 06848901.2, Apr. 1, 2009, pp. 7.
9Examiner's First Report; Application No. 2006331932 May 11, 2011 in 2 pages.
10Examiner's Report for Australian Application No. 2003279847, Dated Dec. 9, 2008, 4 pages.
11Final Office Action for Japanese Application No. 2003531581, Mail Date Mar. 24, 2009, 6 pages.
12First Office Action for Japanese Application No. 2003531581, Mail Date Jul. 8, 2008, 8 pages.
13Gray (#2 of 2, pp. 604-609), Jim; Reuter Andreas, Transaction Processing Concepts and Techniques, Morgan Kaufmann Publisher, USA 1994, pp. 604-609.
14Gray, Jim, Reuter, Andreas: "Transaction Processing: Concepts and Techniques" 1994, Morgan Kaufmann Publisher, USA; pp. 646-655.
15Harrington, "The RFP Process: How To Hire a Third Party", Transportation & Distribution, Sep. 1988, vol. 39, Issue 9, in 5 pages.
16http://en.wikipedia.org/wiki/Naive-Bayes-classifier.
17http://en.wikipedia.org/wiki/Naive—Bayes—classifier.
18International Preliminary Report on Patentability, PCT Application No. PCT/US2009/066880, mailed Jun. 23, 2011, in 9 pages.
19International Search Report and Written Opinion issued in PCT Application No. PCT/US2011/030396, mailed Jul. 18, 2011, in 20 pages.
20International Search Report and Written Opinion issued in PCT Application No. PCT/US2011/38436, mailed Sep. 21, 2011, in 18 pages.
21International Search Report and Written Opinion; dated Mar. 25, 2010; International Application No. PCT/US2009/066880; 10 pages.
22International Search Report from International Application No. PCT/US2006/048273, dated May 15, 2007.
23International Search Report, PCT/US2004/038324, dated Dec. 28, 2009.
24International Search Report, PCT/US2004/038455 dated Jan. 11, 2006.
25Jander, "Launching Storage-Area Net," Data Communications, US, McGraw Hill, NY, vol. 27, No. 4(Mar. 21, 1998), pp. 64-72.
26Jason Gait, "The Optical File Cabinet: A Random-Access File system for Write-Once Optical Disks," IEEE Computer, vol. 21, No. 6, pp. 11-22 (1988) (see in particular figure 5 in p. 15 and the recitation in claim 5).
27Kashyap, et al., "Professional Services Automation: A knowledge Management approach using LSI and Domain specific Ontologies", FLAIRS-01 Proceedings, 2001, pp. 300-302.
28Luis-Felipe Cabrera et al.; ADSM: A Multi-Platform, Scalable, Backup and Archive Mass Storage System; Mar. 5, 1995; pp. 420-427, IBM Research Division.
29Lyon J., Design considerations in replicated database systems for disaster protection, COMPCON 1988, Feb. 29, 1988, pp. 428-430.
30Microsoft Corporation, "Microsoft Exchange Server: Best Practices for Exchange Database Management," 1998.
31PCT International Search Report and Written Opinion for International Application No. PCT/US07/81681, Mail Date Nov. 13, 2009, 8 pages.
32Rosenblum et al., "The Design and Implementation of a Log-Structure File System," Operating Systems Review SIGOPS, vol. 25, No. 5, New York, US, pp. 1-15 (May 1991).
33Second Examination Report in EU Appl. No. 06 848 901.2-2201 dated Dec. 3, 2010.
34The Oracle8 Replication Manual, Part No. A58245-01; Chapters 1-2; Dec. 1, 1997; obtained from website: http://download-west.oracle.com/docs/cd/A64702-01/doc/server.805/a58245/toc.htm on May 20, 2009.
35The Oracle8 Replication Manual, Part No. A58245-01; Chapters 1-2; Dec. 1, 1997; obtained from website: http://download-west.oracle.com/docs/cd/A64702—01/doc/server.805/a58245/toc.htm on May 20, 2009.
36U.S. Appl. No. 12/433,238, filed Apr. 30, 2009, Prahlad et al.
37Veritas Software Corporation, "Veritas Volume Manager 3.2, Administrator's Guide," Aug. 2001, 360 pages.
38Wiesmann M, Database replication techniques: a three parameter classification, Oct. 16, 2000, pp. 206-215.
Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US8504526 *Dec 30, 2010Aug 6, 2013Commvault Systems, Inc.Failover systems and methods for performing backup operations
US20110302140 *Dec 30, 2010Dec 8, 2011Commvault Systems, Inc.Failover systems and methods for performing backup operations
Classifications
U.S. Classification707/639
International ClassificationG06F17/30
Cooperative ClassificationG06F11/2097, G06F2201/84, G06F11/2094, G06F11/1471
European ClassificationG06F11/20U, G06F11/14A12, G06F11/20S6, G06F17/30S
Legal Events
DateCodeEventDescription
Jun 26, 2009ASAssignment
Owner name: COMMVAULT SYSTEMS, INC.,NEW JERSEY
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NGO, DAVID;REEL/FRAME:22882/627
Effective date: 20090527
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NGO, DAVID;REEL/FRAME:022882/0627
Owner name: COMMVAULT SYSTEMS, INC., NEW JERSEY