WO1997035270A1 - Configurable conflict resolution in a computer implemented distributed database - Google Patents

Configurable conflict resolution in a computer implemented distributed database Download PDF

Info

Publication number
WO1997035270A1
WO1997035270A1 PCT/US1997/004399 US9704399W WO9735270A1 WO 1997035270 A1 WO1997035270 A1 WO 1997035270A1 US 9704399 W US9704399 W US 9704399W WO 9735270 A1 WO9735270 A1 WO 9735270A1
Authority
WO
WIPO (PCT)
Prior art keywords
conflict resolution
conflict
configurable
methods
site
Prior art date
Application number
PCT/US1997/004399
Other languages
French (fr)
Other versions
WO1997035270A9 (en
Inventor
Benny Souder
Lip Boon Doop
Alan Downing
Original Assignee
Oracle Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oracle Corporation filed Critical Oracle Corporation
Priority to AU23353/97A priority Critical patent/AU2335397A/en
Priority to CA002249422A priority patent/CA2249422C/en
Priority to GB9819943A priority patent/GB2326495B/en
Publication of WO1997035270A1 publication Critical patent/WO1997035270A1/en
Publication of WO1997035270A9 publication Critical patent/WO1997035270A9/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/23Updating
    • G06F16/2308Concurrency control
    • G06F16/2315Optimistic concurrency control
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/202Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant
    • G06F11/2035Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant without idle spare hardware
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/202Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant
    • G06F11/2048Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant where the redundant components share neither address space nor persistent storage
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/2097Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements maintaining the standby controller/processing unit updated
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/23Updating
    • G06F16/2308Concurrency control
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/23Updating
    • G06F16/2308Concurrency control
    • G06F16/2315Optimistic concurrency control
    • G06F16/2322Optimistic concurrency control using timestamps
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/27Replication, distribution or synchronisation of data between databases or within a distributed database system; Distributed database system architectures therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/80Database-specific techniques
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users
    • Y10S707/99953Recoverability

Definitions

  • the present invention relates to computer implemented database systems. Specifically, the present invention relates to conflict resolution within a database system.
  • Modern data processing systems once the domain of large centralized mainframe computers, have evolved into collections of dispersed independent processing systems interconnected by networked links. That is, modern systems are distributed over many interconnected processing nodes. The efficiency of these distributed systems depends not only upon the processing power of each independent node, but also upon the ability of the system to efficiently move information between processing nodes across the network. It is common for these distributed data processing systems to support the execution of complex application programs which typically access large databases of information. These application programs and distributed systems must be capable of operating across multiple processing platforms which can be geographically separated by great distances. For example, many commercial business operations are geographically dispersed. Some locations may perform manufacturing or product development while other locations perform sales, marketing, inventory control, billing, or various administrative functions.
  • business operations require the use of various types of business data including, for example, customer data, order data, shipping data, billing data, etc. It will be apparent to those of ordinary skill in the art that many other types of information may be required to allow a particular business operation to run efficiently.
  • These various types of business information are stored in distributed systems for access by application programs executed by the data processing systems at the local and remote business locations.
  • These data processing systems called nodes herein, typically include computers, processors, or terminals physically resident at the same proximate location.
  • the distributed systems being accessed by these complex application programs are typically comprised of many tables, arrays, files, and other complex interdependent data structures and related programs.
  • a conventional distributed database technology is replication.
  • multiple replicas or instantiations of data structures or programs can exist in more than one database in a distributed system.
  • a remote database would be used to access a local node in which a desired distributed data structure is resident.
  • the remote node would then transfer a copy of the data structure to itself for local processing without the need for distributed transactions.
  • replication increases the speed of accessing a particular data structure, this methodology produces a problem of maintaining coherency between multiple replicas of the same data structure or program. For example, two different nodes may replicate a distributed data structure and concurrently modify the data structure differently. It then becomes very difficult to reconcile the two modified replicas or to merge the updates into a composite version of the data structure.
  • synchronous replication may be used.
  • synchronous (not asynchronous) replication each update or modification of a data structure is immediately replicated to all other replicas of the data structure existing on any other processing node of the distributed system as part of their local transaction.
  • the data structure modification is not allowed to complete until all other replicas of the data structure have been similarly updated. In this manner, all replicas of the data structure across the distributed system are guaranteed to be the same.
  • synchronous replication methodology provides a simple means for maintaining distributed system coherency, this method is sensitive to network latencies and intermittent network failures and does not work at all for dormant clients.
  • Dormant clients are those that cannot perform an update to a data structure within a predetermined time period. Because each data structure modification is stalled until all nodes have been updated, network delays impact each such modification.
  • synchronous replication does not provide a means for differentiating particular data access transactions as higher or lower priority. Thus, low priority accesses can still produce significant system delays when this result may be unnecessary.
  • the second conventional method for maintaining data structure coherency in a replication methodology is asynchronous replication.
  • asynchronous replication Using this method, local replicas of a particular data structure can be slightly different for a time until an asynchronous update is performed.
  • a distributed node can modify its local copy of a data structure without forcing a network access as in the synchronous replication methodology.
  • all previously un-replicated distributed data structure modifications are combined into a minimal set of updates which are transferred across the network at one time and applied to all other replicas of the data structure.
  • the asynchronous replication method provides a means for balancing network traffic over time.
  • the asynchronous method also provides an effective means for tuning the performance of a particular distributed system; however, a problem of data update conflicts is introduced.
  • Data update conflicts occur when two distributed nodes modify the same data object in a distributed data structure before the asynchronous update process has executed.
  • two data objects may be validly updated locally at two different distributed processing nodes; however, an invalid combination of the two data objects may be produced when the updates to the data objects are propagated to other nodes.
  • these conflicts can usually be reduced, it is important in designing any distributed system to understand and minimize these conflicts as much as possible.
  • remote accesses between nodes are performed using a conventional data manipulation language such as SQL or other conventional protocol.
  • the remote database can use a remote procedure call (RPC) to activate a data access procedure on the local database in a synchronous RPC context.
  • RPC remote procedure call
  • the remote database waits for the RPC to finish before completing the transaction.
  • An RPC can also be used to queue a request on the local system in an asynchronous RPC context.
  • the remote database does not wait for the RPC to finish before completing the transaction.
  • RPC remote procedure call
  • Another form of replication is procedural replication.
  • procedural replication a procedure invocation is replicated to other nodes in a distributed system. This is different from an RPC in that the RPC only produces a procedure execution in the remote node.
  • procedural replication a procedure execution is produced in both the local node and the remote node.
  • both synchronous and asynchronous forms of procedural replication can be implemented. The synchronous form stalls the transaction until the procedure execution is completed, while the asynchronous form allows the transaction to continue before the procedure execution has completed.
  • a second conventional shared database model used multiple copies of a database, each resident at the distributed computing sites. As long as each distributed site operates in a read-only mode (i.e., no data modifications to the shared database), multiple copies of the same data item can be located at multiple sites without raising any data integrity issues.
  • a third technique for distributed data sharing is described in co-pending U.S. patent application serial number 08/126,586, titled “Method and Apparatus for Data Replication", filed September 24, 1993 and assigned to the assignee of the present application.
  • data replication in a networked computing environment is described as allowing duplicate copies of the same data to be resident at more than one site on the network. Any data modifications or updates are replicated to all other sites in an asynchronous manner. In this way, modifications to the shared data can be duplicated at other sites thereby preserving data integrity at each of the distributed sites.
  • Various details of the replication technique as employed with a database having multiple rows, each row having multiple columns or attributes, is described in the above- referenced co-pending patent application.
  • Updatable snapshots are defined to contain a full copy of a master table or a defined subset of the rows in the master table that satisfy a value-based selection criterion. Snapshots are refreshed from the master at time-based intervals or on demand. Any changes to the snapshot are propagated and applied to masters using asynchronous RPCs as the underlying mechanism. Any changes to the master table since the last refresh are then propagated and applied to the snapshot.
  • a snapshot table can only have one master table, but a master table can be the master for many snapshots.
  • update conflicts may occur if two sites concurrently modify the same data item before the data modification can be propagated to other sites. If update conflicts are not handled in some convergent manner, the data integrity of the replicated copies will begin to diverge. It is therefore necessary to first detect update conflicts and secondly to resolve these conflicts in a manner that allows sustained data integrity.
  • Prior art data replication systems either do not handle update conflicts at all or handle conflicts in a rigid and fixed manner. Those conventional systems that do not handle update conflicts at all require that a particular shared database environment be configured in a way to prevent update conflicts such as by partitioning data appropriately. Other conventional data replication systems detect some update conflicts and respond in a predetermined and fixed manner. This rigid conflict resolution technique of the prior art, however, limits significantly the adaptability of a replication system to a particular application or user environment. Conflict resolution is usually highly application specific.
  • a desired conflict resolution method may vary significantly from one application to the next.
  • the present invention is a means and method for providing adaptable and configurable conflict resolution within a replicated data environment.
  • a configurable conflict resolution apparatus comprising; 1) a conflict detection module for detecting a conflicting modification to corresponding portions of the first and the second data structures; 2) a plurality of conflict resolution methods, one or more of the plurality of conflict resolution methods being configurably associated with the corresponding portions of the first and the second data structures; and 3) a conflict resolution module for activating a first conflict resolution method of the one or more of the plurality of conflict resolution methods when the conflict detection module detects the conflicting modification to the corresponding portions of the first and the second data structures.
  • configurable conflict resolution allows complete flexibility of conflict resolution specification. It is a further advantage of the present invention that configurable conflict resolution includes detection and resolution of update, uniqueness, and delete conflicts. It is a further advantage of the present invention that system-provided conflict resolution methods are declarative and user-extensible. It is a further advantage of the present invention that various system-provided or user-extended conflict resolution methods can be applied to resolve a particular conflict. It is a further advantage of the present invention that a plurality of conflict resolution methods will be applied to a particular conflict until resolution of the conflict is achieved or no more methods are available. It is a further advantage of the present invention that a plurality of conflict resolution methods will be applied in the user-specified priority order to a particular conflict.
  • a different update conflict resolution method or set of methods may be specified and employed for each column in a database table. It is a further advantage of the present invention that a different uniqueness conflict resolution method or set of methods may be specified and employed for each uniqueness constraint in a database table. It is a further advantage of the present invention that detection and resolution of update conflicts are based on column groups, each of which is a collection of columns logically treated as a single column, in a database table. It is a further advantage of the present invention that results of conflict resolutions can be collected optionally for analysis.
  • Figure 1 illustrates a typical distributed data processing system.
  • Figure 2 illustrates a typical architecture of a data processing node within a distributed processing system.
  • Figure 3 illustrates a typical update conflict scenario.
  • FIGS. 4 and 5 are flowcharts which illustrate the processing logic for detecting and resolving conflicts in the preferred embodiment.
  • Figure 6 illustrates a typical update ordering conflict scenario.
  • Figure 7 illustrates the conflict resolution methods provided in the preferred embodiment.
  • Figures 8 and 9 illustrate an example of priority groups.
  • Figure 10 illustrates the three types of conflict resolution provided by the present invention.
  • Figures 11-21 are flowcharts which illustrate the processing logic for detecting and resolving conflicts in the preferred embodiment.
  • FIG. 1 illustrates a typical distributed data processing system comprising node 1, node 2, and node 3 coupled together with network links 41 , 42, and 43. It will be apparent to those of ordinary skill in the art that an arbitrary number of nodes in the distributed system may be supported in an arbitrary configuration.
  • Each node such as node 1, comprises a data processing system 10 and a set of distributed programs and data 11 stored in a distributed data store (not shown).
  • the distributed data store can be main memory 104 or mass storage device 107 shown in Figure 2.
  • a typical architecture of data processing system 10 is described below in connection with Figure 2.
  • Distributed data 11 comprises a set of programs, data structures, and data objects which may be shared or replicated by other nodes in the distributed system.
  • Distributed data 11, for example comprises a set of data structures available to other nodes in the distributed processing system.
  • Data processing system 10 may directly access distributed data 11 ; because, distributed data 1 1 is local (i.e. located within the same node) to data processing system 10. In this situation, access between data processing system 10 and distributed data 11 does not require a network access.
  • nodes node 2 and node 3 of the distributed system illustrated in Figure 1 must access distributed data 11 via a network access.
  • node 2 and the data processing system 20 therein must access distributed data 11 via network link 41.
  • This network access can be the result of a distributed transaction or a replication operation, for example.
  • other forms of distributed transfer technologies can cause network accesses.
  • the data processing system 20 of node 2 accesses and manipulates distributed data 11 within the distributed data store of node 1.
  • the data processing system 20 of node 2 accesses distributed data 11 within the distributed data store of node 1 locally within node 2. This local copy of the distributed data 21 within the distributed data store of node 2 may subsequently be manipulated locally by data processing system 20.
  • nodes within the distributed processing system can be varied and diverse. There is no requirement in the present invention that each node have equivalent and compatible processing systems. It is only necessary that each node of the distributed processing system be able to communicate on a network or some communication path coupling the nodes together.
  • FIG. 2 illustrates a typical data processing system upon which one embodiment of the present invention is implemented. It will be apparent to those of ordinary skill in the art, however that other alternative systems of various system architectures may also be used.
  • the data processing system illustrated in Figure 2 includes a bus or other intemal communication means 101 for communicating information, and a processing means 102 coupled to the bus 101 for processing information.
  • the system further comprises a random access memory (RAM) or other volatile storage device 104 (referred to as main memory), coupled to bus 101 for storing information and instructions to be executed by processor 102.
  • Main memory 104 also may be used for storing temporary variables or other intermediate information during execution of instructions by processor 102.
  • the system also comprises a read only memory (ROM) and/or static storage device 106 coupled to bus 101 for storing static information and instructions for processor 102, and a data storage device 107 such as a magnetic disk or optical disk and its corresponding disk drive.
  • Data storage device 107 is coupled to bus 101 for storing information and instructions.
  • the system may further be coupled to a display device 121, such as a cathode ray tube (CRT) or a liquid crystal display (LCD) coupled to bus 101 for displaying information to a computer user.
  • An alphanumeric input device 122 including alphanumeric and other keys, may also be coupled to bus 101 for communicating information and command selections to processor 102.
  • cursor control 123 such as a mouse, a trackball, stylus, or cursor direction keys coupled to bus 101 for communicating direction information and command selections to processor 102, and for controlling cursor movement on display device 121.
  • cursor control 123 such as a mouse, a trackball, stylus, or cursor direction keys coupled to bus 101 for communicating direction information and command selections to processor 102, and for controlling cursor movement on display device 121.
  • Another device which may optionally be coupled to bus 101 is a hard copy device 124 which may be used for printing instructions, data, or other information on a medium such as paper, film, or similar types of media.
  • a communication device 125 is coupled to bus 101 for use in accessing other nodes of the distributed system via a network. This communication device 125 may include any of a number of commercially available networking peripheral devices such as those used for coupling to an Ethernet, token ring, Internet, or wide area network.
  • the data processing system illustrated in Figure 2 is an IBM® compatible personal computer.
  • Processor 102 may be one of the 80X86 compatible microprocessors such as the 80486 or PENTIUM® brand microprocessors manufactured by INTEL® Co ⁇ oration of Santa Clara, California.
  • the distributed data can be stored in main memory 104, mass storage device 107, or other storage medium locally accessible to processor 102. It will also be apparent to those of ordinary skill in the art that the methods and processes described herein can be implemented as software stored in main memory 104 or read only memory 106 and executed by processor 102. This software may also be resident on an article of manufacture comprising a computer usable mass storage medium 108 having computer readable program code embodied therein and being readable by the mass storage device 107 and for causing the processor 102 to perform configurable conflict resolution in accordance with the teachings herein.
  • FIG. 3 an example illustrates a typical update conflict scenario.
  • a master site A 310 and a master site B 320 is shown.
  • An update operation 312 is performed on a table 31 1 at site 310 to produce table 314.
  • a different update operation 322 is performed on table 321 at site 320 to produce table 324.
  • the present invention uses the modification information 316 and 326 and compares the old and new values of the row from the originating site with the old and current values for the same row at the receiving site 320.
  • the present invention detects a conflict 330 if there are any differences between these values for any column in the row. If no conflicts are detected, the row at the receiving site is modified to contain the new value for each column.
  • conflict resolution routines If a conflict is detected, the present invention applies one or more of the appropriate conflict resolution routines in priority order until the conflict is resolved, or until no more conflict resolution routines are available. Any unresolved conflicts are logged as errors. There are three types of conflicts detected by the preferred embodiment of the present invention. These conflicts are: 1 ) uniqueness conflicts, 2) update conflicts, and 3) delete conflicts. It will be apparent to those of ordinary skill in the art that other conflicts may equivalently be defined.
  • the conflict resolution functional block 1010 receives a table identifier and a modification operation related to the identified table. This information is used by the conflict resolution functional block 1010 to detect and resolve update conflicts in block
  • a column group links a collection of columns in a table to a single "logical column".
  • a column group can consist of any number of columns, from a single column, to all of the columns in a table. Each column, however, can belong to only one column group.
  • the update conflict detection mechanism or module of the preferred embodiment detects conflicts on a column group by column group basis, all columns must be a part of some column group. However, all of the columns in a table need not be assigned to a column group. Any column not assigned to a column group is automatically assigned by the present invention to a "shadow" column group for the purpose of conflict detection. This shadow column group is not user visible, and you cannot assign a conflict resolution method to the columns in the group. You can only designate a conflict resolution method for columns in a user-defined column group.
  • numeric data lends itself well to some sort of arithmetic method of resolving conflicts, while conflicts with character data might be better resolved using a timestamp to apply the most recent change.
  • portions of a row in a table may be updated using the data from the originating site, while other portions may maintain the values of the data at the destination site. That is, although the use of a conflict resolution mechanism may result in convergence
  • FIG. 4 the processing logic for detecting and resolving update conflicts in a column group is illustrated.
  • the change indicators for the row are forwarded to the receiving site.
  • the receiving site is then able to determine which column groups in the row have been updated at the originating site (Blocks 410 and 412 in Figure 4). Not all old and new values of all columns in a replicated row are copied and forwarded to the receiving site. Only the old and new values of a column group that has been updated are copied and forwarded to the receiving site.
  • the old primary key values for the row are always copied and forwarded to the receiving site.
  • the old primary key values of the replicated row are used to uniquely identify the row to apply the update. If there is no row or more than one row has been identified, then an unresolvable update conflict has occurred. When an unresolvable update conflict occurs, an error is raised, and the control is returned to the transaction that replicates the row.
  • the receiving site uses the change indicators to identify the column groups that have been changed at the originating site. For those column groups whose values have been modified at the originating site, it compares the old values of each of these column groups with the current values of the same column group at the receiving site.
  • a column group has an update conflict if both the originating site and the receiving site have updated the column group before propagating their updates to each other.
  • the update conflict for a column group is detected by comparing the old values from the originating site with the current values at the receiving site. If the old values do not match the current values, there is an update conflict. Conversely, if the old values match the current values, there is no update conflict (Block 414).
  • the current values at the receiving site are updated with the new values from the originating site (blocks 428 and 526).
  • the update conflict resolution routines that have been assigned to the column group are applied to resolve the update conflict.
  • the conflict resolution routines are applied one by one in priority order until the conflict is resolved (Block 420 in Figure 4 and Block 516 in Figure 5).
  • the conflict is resolved when one of the conflict resolution routines can determine the appropriate new values for the column group (Path 426).
  • the update conflict is resolved for the column group, the current values of the column grouped at the receiving site are updated with the newly determined values (blocks 428 and 526).
  • an error is raised, and the control is returned to the transaction that replicates the row.
  • the current values of all column groups at the receiving site can be updated with their new values using a single UPDATE command, instead of individual UPDATE commands from each column group.
  • the receiving site does not lock the row for update that it has uniquely identified to apply the update.
  • any local running transactions at the receiving site can potentially introduce new update conflicts while current update conflicts are being resolved.
  • the receiving site must check for any new update conflict after it has resolved the current update conflicts.
  • a database table can have one or more unique constraints.
  • a unique constraint can be on one or more columns.
  • a uniqueness constraint enforces that the values of those columns are unique in the database table.
  • a uniqueness conflict occurs when a uniqueness constraint is violated during an insert or update of the replicated row at the receiving site.
  • the uniqueness conflict is detected by comparing the new values from the originating site with the current values of all existing rows at the receiving site. The comparison is performed by each uniqueness constraint. There is a uniqueness conflict for a uniqueness constraint if there is already an existing row at the receiving site whose current values match the new values of those unique columns.
  • uniqueness conflicts can be detected by trapping exceptions from the violations of unique indexes or uniqueness constraints.
  • the conflict resolution routines that have been assigned to the uniqueness constraint are applied to resolve the uniqueness conflict.
  • the conflict resolution routines are applied one by one in priority order until the conflict is resolved.
  • the conflict is resolved when one of the conflict resolution routines can determine the appropriate new values for the unique columns from the originating site.
  • one of the conflict resolution routines can determine the appropriate new values for the unique columns from the originating site.
  • an error is raised, and control is returned to the transaction that replicates the row.
  • the replicated row with any newly determined new column values is then applied to the database table. Any local running transactions at the receiving site can potentially introduce new uniqueness conflicts while current uniqueness conflicts are being resolved. Thus, the receiving site must check for any new uniqueness conflicts after it has resolved all the current uniqueness conflicts.
  • the old primary key values of the deleted row from the originating site are used to uniquely identify the row to delete. If there is no row or more than one row has been identified, then an unresolvable delete conflict has occurred. When an unresolvable delete conflict occurs, an error is raised, and control is returned to the transaction that replicates the delete.
  • the receiving site Once the receiving site has uniquely identified the row to delete, it compares the old values of the deleted row with the current values of the selected row. If the old values match the current values, the selected row is deleted from the receiving site. On the other hand, if the old values do not match the current values, a delete conflict is detected.
  • the delete conflict resolution routines that have been assigned to the database table are applied to resolve the delete conflict.
  • the conflict resolution routines are applied one by one in priority order until the conflict is resolved. If there is no delete conflict resolution routine, or none of the assigned delete conflict resolution routines can resolve the delete conflict, an error is raised, and control is returned to the transaction that replicates the delete. Any local running transactions at the receiving site can potentially introduce a new delete conflict while the current delete conflict is being resolved. Thus, the receiving site must check for any new delete conflict after it has resolved the current delete conflict.
  • the present invention provides various declarative conflict resolution methods for each type of conflict.
  • the user can choose and declare one or more conflict resolution methods to be used for each possible conflict.
  • the user can also declare the priority order of applying the resolution methods when there is more than one resolution method for a possible conflict.
  • the "User Function” method provides user extensibility to standard conflict resolution routines that are provided by the present invention for update conflicts, uniqueness conflicts and delete conflicts. The user can write their own conflict resolution routines and use them together with the standard conflict resolution routines.
  • the user wants to use a user-defined conflict resolution routine for a possible conflict they declare a "User Function” method for the possible conflict and specify the user-defined conflict resolution routine as the User Function.
  • the "User Function” method can also be used by the user to do other useful tasks such as sending out notifications when a conflict cannot be resolved.
  • the user writes a notification routine and registers it as the User Function for the "User Function” method.
  • the standard conflict resolution routines do not support all cases in the preferred embodiment. For example, they do not support the following situations: 1) delete conflicts, 2) changes to a primary key, 3) NULLs or undefined values in the columns designated to be used to resolve the conflict, or 4) referential integrity constraint violations. For these situations, users must write their own conflict resolution routines, or define a method of resolving the conflicts after the errors have been raised and logged in some error tables.
  • the preferred embodiment provides the following methods of resolving update conflicts in replicated environments with any number of master sites. It will be apparent to those of ordinary skill in the art that other update conflict methods are suggested by this disclosure of the present invention. The methods are: 1) apply the data with the latest timestamp, 2) apply all data additively, 3) apply the minimum value, when the column value is always decreasing, and 4) apply the maximum value, when the column value is always increasing.
  • the preferred embodiment provides a variety of methods for resolving uniqueness conflicts: 1 ) append the global name of the originating site to the column value, 2) append a generated sequence number to the column value, or 3) discard the new transaction from the remote site.
  • the preferred embodiment provides one standard resolution method for resolving delete conflicts, which is the "User Function" method. Although the present invention has not provided any standard conflict resolution routine for a delete conflict, the delete conflict can still be detected, and resolved with a user- defined conflict resolution routine.
  • the "User Function” method allows the user to register a user-defined conflict resolution routine for a potential delete conflict.
  • Figure 6 illustrates the actions that may occur over the course of the day in a replicated environment that may lead to the inability of the multiple sites to make conflicts converge.
  • ordering conflicts can be avoided when using priority groups if it is required that the flow of ownership be ordered, as it is in the work flow model. That is, information must always flow from the ORDER site to the SHIP site to the BILL site, in a typical business example. If the billing site receives a change to a row from the ordering site after it has already received a change to that row from the shipping site, for example, the billing site will know to ignore the out-of-order change from the ordering site.
  • Figure 7 summarizes the standard conflict resolution methods provided in the preferred embodiment of the present invention. This figure also shows if these resolution methods can guarantee convergence (all sites ultimately agreeing on the same value) between multiple master sites and their associated snapshot sites. Each of these methods is explained in greater detail in the following sections.
  • the processing logic compares the new value from the originating site with the current value at the receiving site for a designated column in the column group. This column must be designated when the minimum value conflict resolution method is selected.
  • the conflict is not resolved, and the values of the columns in the column group remain unchanged. If the new value of the designated column is less than the current value, the column group values from the originating site are applied at the receiving site. If the new value of the designated column is greater than the current value, the conflict is resolved by leaving the current column group values unchanged.
  • the maximum value method works exactly the same as the minimum value method, except that the values from the originating site are only applied if the value of the designated column at the originating site is greater than the value of the designated column at the receiving site.
  • the earliest and latest timestamp methods are simply variations on the minimum and maximum value methods.
  • the designated column must be of the well known type DATE.
  • SYSDATE SYSTEM DATE
  • the timestamp value should be set to the timestamp value from the originating site. For example, suppose that a customer calls his local sales representative to update his address information. After hanging up the phone, he realizes that he gave the wrong zip code.
  • the present invention When he tries to call his sales representative with the correct zip code, he discovers that the phone lines in that area have gone down, so he calls the headquarters number instead, to update his address information again. When the network connecting the headquarters site with the sales site comes back up, the present invention will see two updates for the same address, and detect a conflict. By using the latest timestamp method, the present invention would select the later update, and apply the address with the correct zip code. If the replicated environment crosses time zones, all timestamps should be converted to a common time zone. Otherwise, although the data will converge, the most recent update may not be applied as expected.
  • the earliest timestamp method applies the changes from the site with the earliest timestamp
  • the latest timestamp method applies the changes from the site with the latest timestamp.
  • a backup method such as site priority, should be designated to be called in case two sites have the same timestamp.
  • the timestamping mechanism should be designed to be time zone independent (for example, by always converting the timestamp to a designated time zone, such as GMT).
  • the maximum value method can guarantee convergence if the value is always increasing, the latest timestamp method can guarantee convergence (assuming that you have properly designed your timestamping mechanism). The earliest timestamp method cannot guarantee convergence for more than two masters (since time is generally not always decreasing).
  • the earliest and latest timestamp methods require auditing of update timestamps. Users specify whether the present invention automatically audits update timestamps or whether an application explicitly audits update timestamps. If the application audits update timestamps, users must specify which column in the column group stores the last update timestamp. The datatype of the column must be of date.
  • the auditing column stores the last update timestamp of any column in the column group.
  • the auditing column always stores the timestamp when the column value(s) were first created. For example, if the earliest change is from a remote site, this method results in overwriting the current column value with the replicated column value from the remote site, and updating the auditing column with the timestamp when the replicated column value was first created, that is, from the remote site.
  • the additive and average conflict resolution routines work with column groups consisting of a single numeric column only.
  • the additive conflict resolution method provides convergence for any number of master sites.
  • current value (current value + new value )/2
  • the average method cannot guarantee convergence if the replicated environment has more than one master. This method is useful for an environment with a single master site and multiple updatable snapshots.
  • Priority groups allow a user to assign a priority level to each possible value of a particular column. As shown in the example illustrated in Figures 8 and 9, a Priority view 910 shows the priority level 914 assigned to each value 916 that the "priority" column can contain. A priority level for all possible values of the "priority" column must be specified.
  • the priority group method of conflict resolution is selected for a column group, the user designates which column in the column group is the "priority " column.
  • the priority group conflict resolution routine is invoked to resolve an update conflict in a column group, the processing logic compares the priority level of the new value from the originating site with that of the current value at the receiving site for the designated "priority" column in the column group.
  • the conflict is not resolved, and the current column group values at the receiving site remain unchanged. If the priority level of the new value in the designated "priority" column is greater than that of the current value at the receiving site, the current column group values at the receiving site are updated with the new column group values from the originating site. Conversely, if the priority level of the new value in the designated "priority” column is less than that of the current value at the receiving site, the conflict is resolved by leaving the current column group values at the receiving site unchanged. It will be apparent to one of ordinary skill in the art that the priority group conflict resolution routine may equivalently interpret that a lower priority level takes precedence over a higher priority level.
  • the Priority view 910 displays the values of all priority groups 912 defined at the current location.
  • the CREDIT_STATUS table 920 is shown in Figure 9 as using the "site-priority" priority group 912; because the site values 918 in the CREDIT_STATUS table 920 correspond to the "site-priority" values 916 shown in Figure 8.
  • Site priority is a special kind of priority group. With site priority, the "priority" column designated is automatically updated with the global database name of the site where the update originated.
  • the Priority view 910 in Figure 8 shows the priority level 914 assigned to each database site 916. Site priority can be useful if one site is considered to be more likely to have the most accurate information.
  • the Houston site (houston.world) may be the corporate headquarters, while the New York site (new_york. world) may be a sales office.
  • the headquarters office (houston.world) in this example, is considered more likely to have the most accurate information about the credit that can be extended to each customer.
  • the headquarters office (houston.world) priority is set higher (“ 1 ") than the sales office (new_york.world) priority ("2").
  • the values in the priority column correspond to an ordered sequence of events; for example: ordered, shipped, billed.
  • the site priority method requires auditing of the global name of the site that applied the last update. Users specify whether the present invention or an application maintains the audit information. If the application maintains the audit information, users must specify which column in the column group stores the site global name. The column specified always stores the global name of the site where the column value(s) were first created.
  • the overwrite and discard methods ignore the values from either the originating or receiving site and therefore can never guarantee convergence with more than one master site.
  • These methods are designed to be used by a single master site and multiple snapshot sites, or with some form of a user-defined notification facility. For example, if you have a single master site that you expect to be used primarily for queries, with all updates being performed at the snapshot sites, you might select the overwrite method. These methods are also useful if your primary concern is data convergence (and you have a single master site), and there is no particular business rule for selecting one update over the other. You may even choose to use one of these methods if you have multiple master sites, provided that you supply a notification facility. This allows you to notify the appropriated personnel, who will be responsible for ensuring that the data is correctly applied, instead of logging the conflict in the some error tables and leaving the resolution up of a local database administrator.
  • the overwrite routine overwrites the current value at the receiving site with the new value from the originating site. Conversely, the discard method 32 discards the new value from the originating site.
  • the append site name and append sequence conflict resolution routines work by appending a string to a column whose unique constraint has been violated, that is, duplicate column values have been detected. If the unique constraint is on more than one column, the user must specify one of the columns as the column to which the string is appended. Although this allows the column to be inserted or updated without violating a unique integrity constraint, it does not provide any form of convergence between multiple master sites. The resulting discrepancies must be manually resolved; therefore, these methods are meant to be used with some form of a notification facility. Both methods can be used on character columns only in the preferred embodiment.
  • the append site name routine appends the global database name of the site originating the transaction to the new column value from the originating site.
  • the append sequence routine appends a generated sequence number to the new column value from the originating site.
  • This conflict resolution routine resolves uniqueness conflicts by simply discarding the row from the originating site. This method never guarantees convergence with multiple masters and should be used with a notification facility.
  • this method is useful if you want to minimize the propagation of data until its accuracy can be verified.
  • Configurable conflict resolution allows users to freely mix their own functions with the standard conflict resolution routines that are provided by the present invention.
  • the present invention does not impose any restrictions on user functions. But, it requires that user functions adhere to the following interface specification in the preferred embodiment:
  • the user function should return FALSE if it has not successfully resolved the conflict. • The user function should accept column values as parameters.
  • the user function should accept old, new and current columns values.
  • the old, new and current values for a column are received consecutively. If it can resolve the conflict, it modifies only the new column values so that the replication procedure can update the current row with the new column values.
  • the function sets the flag to TRUE if it wants to discard the new column values; otherwise it sets the flag to FALSE.
  • Configurable conflict resolution allows users to freely mix their own functions with the standard conflict resolution routines that are provided by the present invention.
  • the present invention does not impose any restrictions on user functions. But, it requires that user functions adhere to the following interface specification:
  • the function should return FALSE if it has not successfully resolved the conflict. • The function should have all parameters with IN OUT mode.
  • the function should accept new column values. If it can resolve the conflict, it modifies the new column values so that the replication procedure can insert or update the current row with the new column values.
  • the last parameter of the function must be a boolean flag.
  • the function sets the flag to TRUE if it wants to discard the new column values; otherwise it sets the flag to FALSE. Users specify which columns in the table constitute the parameters for each user function.
  • Configurable conflict resolution allows users to freely mix their own functions.
  • the present invention does not impose any restrictions on user functions. But, it requires that user functions adhere to the following interface specification:
  • the function should accept old column values. If it can resolve the conflict, it modifies the old column values so that the replication procedure can delete the current row that matches all old column values.
  • the last parameter of the function must be a boolean flag.
  • the function sets the flag to TRUE if it wants to discard the old column values; otherwise it sets the flag to FALSE. Users specify which columns in the table constitute the parameters for each user function.
  • 15-21 are flowcharts illustrating the logic steps performed by the processing logic of the present invention for resolving conflicts with the automatic generation of audit information. These flowcharts describe a transaction between two sites (Master site I and Master site II) and illustrate the conflict resolution features of the present invention. These flowcharts are described in detail below.
  • a table (called “emp” in this example) is updated in the first site or Master site I (block 1 110).
  • a corresponding trigger (emp$rt) is fired when the table is updated.
  • the OLD and NEW values (values prior to the update and values subsequent to the update, respectively) of the table are written to a transfer buffer (DefTran) in the Master site I in block 1112.
  • a Master Site II replication update routine (emp$rp.rep_update) is called in block 1114 and processing for Master site I ends in Figure 11. The perspective of this description then shifts to the point of view of the Master site II as shown starting in Figure 12.
  • the "emp" table in Master site II is updated in block 1210.
  • the "emp" table in each of the Master sites I and II are replicated copies of each other that may be updated independently at each site where it resides.
  • a series of tests are executed to determine if the update caused any conflicts. First, a test is performed to determine if an update conflict is present. If so, the no_data__found exception is raised in block 1214 and processing continues at the bubble labeled N shown in Figure 13. If a uniqueness conflict occurred, the dup_val_on_index exception is raised in block 1218 and processing continues at the bubble labeled P shown in Figure 14. If some other exception occurred, the other exception is raised in block 1222 and processing continues at the bubble labeled R shown in Figure 13. If no exception is detected, processing continues at the bubble labeled S shown in Figure 13.
  • an update conflict handler (emp$rr.update_conflict_handler in this example) is called in the Master site II in block 1310.
  • the current table values in Master site II are queried in block 1312.
  • the column groups are tested to determine which column group has an update conflict in block 1314.
  • the corresponding update conflict resolution routine(s) are called in block 1316. If the update conflict is resolved, block 1320 in Master site II is executed to call the replication update routine emp$rp.rep_update in this example. Processing then continues at the bubble labeled E shown in Figure 12 where the emp table is again recursively checked for update conflicts. If in block 1318 the update conflict is not resolved, an error is posted in block 1322 and processing terminates at the End bubble shown in Figure 13.
  • a uniqueness conflict handler (emp$rr.unique_conflict_update_handler in this example) is called in the Master site II in block 1410.
  • the uniqueness constraint that was violated is determined in block 1414.
  • the corresponding uniqueness conflict resolution routine(s) are called in block 1416. Processing continues at the bubble labeled T shown in Figure 13 where a test is executed to determine if the uniqueness conflict was resolved.
  • Figures 15-21 are flowcharts illustrating the logic steps performed by the processing logic of the present invention for resolving conflicts with the automatic generation of audit information.
  • a table (called “emp” in this example) is updated in Master site I (block 1510).
  • a corresponding trigger (emp$rt) is fired when the table is updated.
  • the OLD and NEW values of the table are written to a transfer buffer (DefTran) in the Master site I in block 1512. Audit columns of the table (emp) are queried for the OLD audit values (block 1514).
  • the OLD audit values and NEW audit values are written to the transfer buffer DefTran in block 1516.
  • a Master Site II replication update routine (emp$rp.rep_update) is called in block 1518 and processing for Master site I ends in Figure 15. The perspective of this description then shifts to the point of view of the Master site II for the automatic generation of audit information as shown starting in Figure 16.
  • a shadow table (emp$ra) corresponding to the emp table is updated in the Master site II in block 1609. If an update conflict (block 1611) is found, processing continues at the bubble labeled J shown in Figure 18. If a uniqueness conflict (block 1613) in the shadow table is found, processing continues at the bubble labeled G shown in Figure 17 where the emp table is checked for conflicts. If another exception is found (block 1614), processing continues at the bubble RR shown in Figure 20. Otherwise, processing continues at bubble G shown in Figure 17.
  • the table (“emp” in this example) is checked for conflicts.
  • the "emp" table in Master site II is updated in block 1710.
  • a series of tests are executed to determine if the update caused any conflicts.
  • a test is performed to determine if an update conflict is present. If so, the no_data_found exception is raised in block 1714 and processing continues at the bubble labeled J shown in Figure 18. If a uniqueness conflict occurred, the dup_val_on_index exception is raised in block 1718 and processing continues at the bubble labeled K shown in Figure 19. If some other exception occurred, the other exception is raised in block 1722 and processing continues at the bubble labeled RR shown in Figure 20. If no exception is detected, processing continues at the bubble labeled SS shown in Figure 20.
  • an update conflict handler (emp$rr.update_conflict_handler, in this example) is called in the Master site II in block 2010.
  • the current table values in Master site II are queried in block 2012.
  • the column groups are tested to determine which column group has an update conflict in block 2014.
  • the corresponding update conflict resolution routine(s) are called in block 2016.
  • block 2020 in Master site II is executed to call the replication update routine emp$rp.rep_update, in this example. Processing then continues at the bubble labeled EE shown in Figure 16 where the emp table is again recursively checked for update conflicts. If in block 2018 the update conflict is not resolved, an error is posted in block 2022 and processing terminates at the End bubble shown in Figure 20.
  • conflict resolution methods require additional information typically stored in an audit table, such as last update timestamp or last update site.
  • the conflict resolution methods that require such audit information are listed below. A description of the audit information used by these methods is set forth above.
  • Every replicated table can have one and only one shadow table in the preferred embodiment.
  • the present invention creates a shadow table if none exists, or modifies the table definition if the generated shadow table already exists. Once a shadow table has been created, the present invention can only add new columns to the shadow table. It is possible that the present invention may fail to add new columns to a shadow table if there are obsolete columns in the shadow table. Some columns become obsolete when users change their use of conflict resolution methods that require automatic generation of audit information. It is recommended that the database administrator periodically removes obsolete columns from the shadow tables.
  • the present invention treats a shadow table as if it were an extension of the replicated table. Because of this, the "shadow table" for snapshot replicated object must be a snapshot at the snapshot site and must be in the same refresh group as the snapshot replicated object.
  • a shadow table has the following columns:
  • the name of the shadow table is derived by appending $ra to the table name.
  • the present invention truncates the table name to accommodate the additional characters when necessary. If the derived name conflicts with an existing one, the present invention uses an arbitrary name with $ra appended.
  • the present invention uses the exact primary key column names in the shadow table. As for other columns in the shadow table, the present invention derives their names according to the type of audit information.
  • the present invention uses "TIMESTAMP” as the name of a column used for storing timestamp, and "GLOBAL_NAME" as the name of a column used for storing site's global database name.
  • the present invention appends a unique integer to the column name if that column name already exists.

Abstract

An apparatus and method for providing adaptable and configurable conflict resolution within a replicated data environment is disclosed. In a distributed database system having a first node and a second node, the first node including a first data structure, the second node including a second data structure, a configurable conflict resolution apparatus is disclosed as comprising: 1) a conflict detection module for detecting a conflicting modification to corresponding portions of the first and the second data structures; 2) a plurality of conflict resolution methods, one or more of the plurality of conflict resolution methods being configurably associated with the corresponding portions of the first and the second data structures; and 3) a conflict resolution module for activating a first conflict resolution method of the one or more of the plurality of conflict resolution methods when the conflict detection module detects the conflicting modification to the corresponding portions of the first and the second data structures.

Description

Configurable Conflict Resolution in a Computer Implemented Distributed Database
FIELD OF THE INVENTION
The present invention relates to computer implemented database systems. Specifically, the present invention relates to conflict resolution within a database system.
DESCRIPTION OF RELATED ART
Modern data processing systems, once the domain of large centralized mainframe computers, have evolved into collections of dispersed independent processing systems interconnected by networked links. That is, modern systems are distributed over many interconnected processing nodes. The efficiency of these distributed systems depends not only upon the processing power of each independent node, but also upon the ability of the system to efficiently move information between processing nodes across the network. It is common for these distributed data processing systems to support the execution of complex application programs which typically access large databases of information. These application programs and distributed systems must be capable of operating across multiple processing platforms which can be geographically separated by great distances. For example, many commercial business operations are geographically dispersed. Some locations may perform manufacturing or product development while other locations perform sales, marketing, inventory control, billing, or various administrative functions. These business operations require the use of various types of business data including, for example, customer data, order data, shipping data, billing data, etc. It will be apparent to those of ordinary skill in the art that many other types of information may be required to allow a particular business operation to run efficiently. These various types of business information are stored in distributed systems for access by application programs executed by the data processing systems at the local and remote business locations. These data processing systems, called nodes herein, typically include computers, processors, or terminals physically resident at the same proximate location. The distributed systems being accessed by these complex application programs are typically comprised of many tables, arrays, files, and other complex interdependent data structures and related programs.
A conventional distributed database technology is replication. Using this methodology, multiple replicas or instantiations of data structures or programs can exist in more than one database in a distributed system. In a typical scenario, a remote database would be used to access a local node in which a desired distributed data structure is resident. The remote node would then transfer a copy of the data structure to itself for local processing without the need for distributed transactions. Although replication increases the speed of accessing a particular data structure, this methodology produces a problem of maintaining coherency between multiple replicas of the same data structure or program. For example, two different nodes may replicate a distributed data structure and concurrently modify the data structure differently. It then becomes very difficult to reconcile the two modified replicas or to merge the updates into a composite version of the data structure. Thus, multiple modifiable replicas of a data structure introduces a coherency problem. Within the replication methodology, two basic conventional techniques are used for maintaining coherency among multiple replicas of the same data structure. First, synchronous replication may be used. In synchronous (not asynchronous) replication, each update or modification of a data structure is immediately replicated to all other replicas of the data structure existing on any other processing node of the distributed system as part of their local transaction.
The data structure modification is not allowed to complete until all other replicas of the data structure have been similarly updated. In this manner, all replicas of the data structure across the distributed system are guaranteed to be the same.
Although the synchronous replication methodology provides a simple means for maintaining distributed system coherency, this method is sensitive to network latencies and intermittent network failures and does not work at all for dormant clients. Dormant clients are those that cannot perform an update to a data structure within a predetermined time period. Because each data structure modification is stalled until all nodes have been updated, network delays impact each such modification. Further, synchronous replication does not provide a means for differentiating particular data access transactions as higher or lower priority. Thus, low priority accesses can still produce significant system delays when this result may be unnecessary.
The second conventional method for maintaining data structure coherency in a replication methodology is asynchronous replication. Using this method, local replicas of a particular data structure can be slightly different for a time until an asynchronous update is performed. In asynchronous replication, a distributed node can modify its local copy of a data structure without forcing a network access as in the synchronous replication methodology. At a predetermined time interval or on demand, all previously un-replicated distributed data structure modifications are combined into a minimal set of updates which are transferred across the network at one time and applied to all other replicas of the data structure. The asynchronous replication method provides a means for balancing network traffic over time. The asynchronous method also provides an effective means for tuning the performance of a particular distributed system; however, a problem of data update conflicts is introduced. Data update conflicts occur when two distributed nodes modify the same data object in a distributed data structure before the asynchronous update process has executed. In this situation, two data objects may be validly updated locally at two different distributed processing nodes; however, an invalid combination of the two data objects may be produced when the updates to the data objects are propagated to other nodes. Although these conflicts can usually be reduced, it is important in designing any distributed system to understand and minimize these conflicts as much as possible.
Typically, remote accesses between nodes are performed using a conventional data manipulation language such as SQL or other conventional protocol. Alternatively, the remote database can use a remote procedure call (RPC) to activate a data access procedure on the local database in a synchronous RPC context. In a synchronous context, the remote database waits for the RPC to finish before completing the transaction. An RPC can also be used to queue a request on the local system in an asynchronous RPC context. In an asynchronous context, the remote database does not wait for the RPC to finish before completing the transaction. The use of a remote procedure call is well known to those of ordinary skill in the art.
Another form of replication is procedural replication. In procedural replication, a procedure invocation is replicated to other nodes in a distributed system. This is different from an RPC in that the RPC only produces a procedure execution in the remote node. In procedural replication, a procedure execution is produced in both the local node and the remote node. As in the use of RPCs, both synchronous and asynchronous forms of procedural replication can be implemented. The synchronous form stalls the transaction until the procedure execution is completed, while the asynchronous form allows the transaction to continue before the procedure execution has completed.
Multiple sites in a distributed computing environment often have need to share information. Various models have been employed in the prior art to support data sharing. One model employs a single centralized database which is shared among multiple users at each distributed site. Using this model, only one copy of the data is maintained in the centralized database. This model, however, suffers performance limitations because each of the distributed sites are competing for access to the same database. A second conventional shared database model used multiple copies of a database, each resident at the distributed computing sites. As long as each distributed site operates in a read-only mode (i.e., no data modifications to the shared database), multiple copies of the same data item can be located at multiple sites without raising any data integrity issues. However, as multiple users at multiple system locations begin to modify one or more of the copies of the same data items, data integrity becomes a critical issue. As a solution to the data integrity problems caused by data modifications or updates to a shared database, conventional systems employ a distributed database configuration wherein a master copy of the database resides at one site and slave copies reside at any of the other distributed sites. Any modifications to the database are performed only on the master site with the slave locations receiving a copy of the modified data after the modification is completed on the master site. Thus, a user at a slave location must access the master copy to modify a data item. This technique does not provide the ability to update any copy of the data and to propagate the changes performed on that copy to all other copies of the data.
A third technique for distributed data sharing is described in co-pending U.S. patent application serial number 08/126,586, titled "Method and Apparatus for Data Replication", filed September 24, 1993 and assigned to the assignee of the present application. In this co-pending application, data replication in a networked computing environment is described as allowing duplicate copies of the same data to be resident at more than one site on the network. Any data modifications or updates are replicated to all other sites in an asynchronous manner. In this way, modifications to the shared data can be duplicated at other sites thereby preserving data integrity at each of the distributed sites. Various details of the replication technique as employed with a database having multiple rows, each row having multiple columns or attributes, is described in the above- referenced co-pending patent application.
Another technique for asynchronous data replication is updatable snapshots. Updatable snapshots are defined to contain a full copy of a master table or a defined subset of the rows in the master table that satisfy a value-based selection criterion. Snapshots are refreshed from the master at time-based intervals or on demand. Any changes to the snapshot are propagated and applied to masters using asynchronous RPCs as the underlying mechanism. Any changes to the master table since the last refresh are then propagated and applied to the snapshot. A snapshot table can only have one master table, but a master table can be the master for many snapshots. Using replication, update conflicts may occur if two sites concurrently modify the same data item before the data modification can be propagated to other sites. If update conflicts are not handled in some convergent manner, the data integrity of the replicated copies will begin to diverge. It is therefore necessary to first detect update conflicts and secondly to resolve these conflicts in a manner that allows sustained data integrity.
Prior art data replication systems either do not handle update conflicts at all or handle conflicts in a rigid and fixed manner. Those conventional systems that do not handle update conflicts at all require that a particular shared database environment be configured in a way to prevent update conflicts such as by partitioning data appropriately. Other conventional data replication systems detect some update conflicts and respond in a predetermined and fixed manner. This rigid conflict resolution technique of the prior art, however, limits significantly the adaptability of a replication system to a particular application or user environment. Conflict resolution is usually highly application specific.
Depending upon the application of particular data items in a shared database, a desired conflict resolution method may vary significantly from one application to the next.
If all sites (or nodes) in a replicated environment, such as the one illustrated in Figure 1 , are propagating database changes synchronously, conflicting updates cannot occur, and there is no need to designate a conflict resolution method. However, if any sites in the replicated environment are propagating changes asynchronously, it is advisable to designate a conflict resolution method for each replicated table. Even if the data environment has been designed to avoid conflicts (for example, by partitioning data ownership) there should be provided at a minimum some form of notification mechanism to alert someone if an unexpected conflict does occur. Sometimes it may be necessary to provide multiple conflict resolution methods for a single column or group of columns. These methods would then be applied in order (using a specified priority ranking) until the conflict is resolved.
This is useful for a variety of reasons. For example, there might be a preferred method of resolving a conflict that might not always be successful. In this event, a backup method could be provided to have a greater chance of not requiring manual intervention to resolve the error. In another case, a user- defined method could be provided that performs logging or notification if the conflict cannot be resolved. It would be advantageous to be able to mix any number of user-defined and previously supplied conflict resolution routines.
Another reason for needing multiple conflict resolution methods is if a particular method is selected, such as latest timestamp, it would be advantageous to provide a backup method to guarantee success. The latest timestamp method uses a special timestamp column to determine and apply the most recent change. In the unlikely event that the row at the originating site and the row at another site were changed at precisely the same time, a backup method, such as site priority, could be provided to break this tie. Prior art systems do not provide a capability for handling database conflict problems using multiple conflict resolution methods.
Thus, a better means and method for handling update, uniqueness, and delete conflicts in a replicated database environment is needed.
SUMMARY OF THE INVENTION
The present invention is a means and method for providing adaptable and configurable conflict resolution within a replicated data environment. In a distributed database system having a first node and a second node, the first node including a first data structure, the second node including a second data structure, the present invention is a configurable conflict resolution apparatus comprising; 1) a conflict detection module for detecting a conflicting modification to corresponding portions of the first and the second data structures; 2) a plurality of conflict resolution methods, one or more of the plurality of conflict resolution methods being configurably associated with the corresponding portions of the first and the second data structures; and 3) a conflict resolution module for activating a first conflict resolution method of the one or more of the plurality of conflict resolution methods when the conflict detection module detects the conflicting modification to the corresponding portions of the first and the second data structures.
Therefore, it is an advantage of the present invention that configurable conflict resolution allows complete flexibility of conflict resolution specification. It is a further advantage of the present invention that configurable conflict resolution includes detection and resolution of update, uniqueness, and delete conflicts. It is a further advantage of the present invention that system-provided conflict resolution methods are declarative and user-extensible. It is a further advantage of the present invention that various system-provided or user-extended conflict resolution methods can be applied to resolve a particular conflict. It is a further advantage of the present invention that a plurality of conflict resolution methods will be applied to a particular conflict until resolution of the conflict is achieved or no more methods are available. It is a further advantage of the present invention that a plurality of conflict resolution methods will be applied in the user-specified priority order to a particular conflict. It is a further advantage of the present invention that a different update conflict resolution method or set of methods may be specified and employed for each column in a database table. It is a further advantage of the present invention that a different uniqueness conflict resolution method or set of methods may be specified and employed for each uniqueness constraint in a database table. It is a further advantage of the present invention that detection and resolution of update conflicts are based on column groups, each of which is a collection of columns logically treated as a single column, in a database table. It is a further advantage of the present invention that results of conflict resolutions can be collected optionally for analysis.
BRIEF DESCRIPTION OF THE DRAWINGS
The features and advantages of the present invention will be apparent from the accompanying drawings and from the detailed description of the preferred embodiment of the present invention as set forth below. Figure 1 illustrates a typical distributed data processing system.
Figure 2 illustrates a typical architecture of a data processing node within a distributed processing system.
Figure 3 illustrates a typical update conflict scenario.
Figures 4 and 5 are flowcharts which illustrate the processing logic for detecting and resolving conflicts in the preferred embodiment.
Figure 6 illustrates a typical update ordering conflict scenario.
Figure 7 illustrates the conflict resolution methods provided in the preferred embodiment.
Figures 8 and 9 illustrate an example of priority groups. Figure 10 illustrates the three types of conflict resolution provided by the present invention.
Figures 11-21 are flowcharts which illustrate the processing logic for detecting and resolving conflicts in the preferred embodiment.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
The present invention is a means and method for providing adaptable and configurable conflict resolution in a replicated database system. In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it will be apparent to one of ordinary skill in the art that these specific details need not be used to practice the present invention. In other instances, well-known structures, interfaces, and processes have not been shown in detail in order not to unnecessarily obscure the present invention. Figure 1 illustrates a typical distributed data processing system comprising node 1, node 2, and node 3 coupled together with network links 41 , 42, and 43. It will be apparent to those of ordinary skill in the art that an arbitrary number of nodes in the distributed system may be supported in an arbitrary configuration. Each node, such as node 1, comprises a data processing system 10 and a set of distributed programs and data 11 stored in a distributed data store (not shown). The distributed data store can be main memory 104 or mass storage device 107 shown in Figure 2. A typical architecture of data processing system 10 is described below in connection with Figure 2. Distributed data 11 comprises a set of programs, data structures, and data objects which may be shared or replicated by other nodes in the distributed system. Distributed data 11, for example, comprises a set of data structures available to other nodes in the distributed processing system. Data processing system 10 may directly access distributed data 11 ; because, distributed data 1 1 is local (i.e. located within the same node) to data processing system 10. In this situation, access between data processing system 10 and distributed data 11 does not require a network access. Typically, these local accesses can be performed more quickly than accesses requiring a network communication. Other nodes (node 2 and node 3) of the distributed system illustrated in Figure 1 must access distributed data 11 via a network access. For example, node 2 and the data processing system 20 therein must access distributed data 11 via network link 41.
This network access can be the result of a distributed transaction or a replication operation, for example. As described above, other forms of distributed transfer technologies can cause network accesses. In a distributed transaction system, the data processing system 20 of node 2 accesses and manipulates distributed data 11 within the distributed data store of node 1. In a replication system, after a local version of the distributed data 1 1 within the distributed data store of node 1 has been transferred to the distributed data store of node 2, the data processing system 20 of node 2 accesses distributed data 11 within the distributed data store of node 1 locally within node 2. This local copy of the distributed data 21 within the distributed data store of node 2 may subsequently be manipulated locally by data processing system 20. The hardware architecture of nodes within the distributed processing system, such as the one illustrated in Figure 1 , can be varied and diverse. There is no requirement in the present invention that each node have equivalent and compatible processing systems. It is only necessary that each node of the distributed processing system be able to communicate on a network or some communication path coupling the nodes together.
Figure 2 illustrates a typical data processing system upon which one embodiment of the present invention is implemented. It will be apparent to those of ordinary skill in the art, however that other alternative systems of various system architectures may also be used. The data processing system illustrated in Figure 2 includes a bus or other intemal communication means 101 for communicating information, and a processing means 102 coupled to the bus 101 for processing information. The system further comprises a random access memory (RAM) or other volatile storage device 104 (referred to as main memory), coupled to bus 101 for storing information and instructions to be executed by processor 102. Main memory 104 also may be used for storing temporary variables or other intermediate information during execution of instructions by processor 102. The system also comprises a read only memory (ROM) and/or static storage device 106 coupled to bus 101 for storing static information and instructions for processor 102, and a data storage device 107 such as a magnetic disk or optical disk and its corresponding disk drive. Data storage device 107 is coupled to bus 101 for storing information and instructions. The system may further be coupled to a display device 121, such as a cathode ray tube (CRT) or a liquid crystal display (LCD) coupled to bus 101 for displaying information to a computer user. An alphanumeric input device 122, including alphanumeric and other keys, may also be coupled to bus 101 for communicating information and command selections to processor 102. An additional user input device is cursor control 123, such as a mouse, a trackball, stylus, or cursor direction keys coupled to bus 101 for communicating direction information and command selections to processor 102, and for controlling cursor movement on display device 121. Another device which may optionally be coupled to bus 101 is a hard copy device 124 which may be used for printing instructions, data, or other information on a medium such as paper, film, or similar types of media. In the preferred embodiment, a communication device 125 is coupled to bus 101 for use in accessing other nodes of the distributed system via a network. This communication device 125 may include any of a number of commercially available networking peripheral devices such as those used for coupling to an Ethernet, token ring, Internet, or wide area network. Note that any or all of the components of the system illustrated in Figure 2 and associated hardware may be used in various embodiments of the present invention; however, it will be appreciated by those of ordinary skill in the art that any configuration of the system may be used for various purposes according to the particular implementation. In one embodiment of the present invention, the data processing system illustrated in Figure 2 is an IBM® compatible personal computer. Processor 102 may be one of the 80X86 compatible microprocessors such as the 80486 or PENTIUM® brand microprocessors manufactured by INTEL® Coφoration of Santa Clara, California.
The distributed data, such as distributed data 11 illustrated in Figure 1 , can be stored in main memory 104, mass storage device 107, or other storage medium locally accessible to processor 102. It will also be apparent to those of ordinary skill in the art that the methods and processes described herein can be implemented as software stored in main memory 104 or read only memory 106 and executed by processor 102. This software may also be resident on an article of manufacture comprising a computer usable mass storage medium 108 having computer readable program code embodied therein and being readable by the mass storage device 107 and for causing the processor 102 to perform configurable conflict resolution in accordance with the teachings herein.
The following sections describe how to resolve update, uniqueness, and delete conflicts in an asynchronously replicated environment using the preferred embodiment of the present invention. General topics within these sections include the following: 1) the types of conflicts detected and resolved in the preferred embodiment, 2) how column groups are used in conflict detection, 3) detecting and resolving update, uniqueness, and delete conflicts, 4) how to select a conflict resolution method, 5) how to avoid ordering conflicts, 6) the details of particular update, uniqueness and delete conflict resolution methods, 7) how to use column groups, priority groups, and site priority, and 8) how to provide automatic generation of audit information. It will be apparent to those of ordinary skill in the art that the preferred embodiment of the present invention is disclosed. Other alternative embodiments are within the scope of the present invention.
Types of Conflicts
Referring to Figure 3, an example illustrates a typical update conflict scenario. In the example of Figure 3, a master site A 310 and a master site B 320 is shown. An update operation 312 is performed on a table 31 1 at site 310 to produce table 314. A different update operation 322 is performed on table 321 at site 320 to produce table 324. The present invention uses the modification information 316 and 326 and compares the old and new values of the row from the originating site with the old and current values for the same row at the receiving site 320. The present invention detects a conflict 330 if there are any differences between these values for any column in the row. If no conflicts are detected, the row at the receiving site is modified to contain the new value for each column. If a conflict is detected, the present invention applies one or more of the appropriate conflict resolution routines in priority order until the conflict is resolved, or until no more conflict resolution routines are available. Any unresolved conflicts are logged as errors. There are three types of conflicts detected by the preferred embodiment of the present invention. These conflicts are: 1 ) uniqueness conflicts, 2) update conflicts, and 3) delete conflicts. It will be apparent to those of ordinary skill in the art that other conflicts may equivalently be defined.
Referring to Figure 10, the basic conflict resolution functional block 1010 of the preferred embodiment of the present invention is shown. The conflict resolution functional block 1010 receives a table identifier and a modification operation related to the identified table. This information is used by the conflict resolution functional block 1010 to detect and resolve update conflicts in block
1012, to detect and resolve uniqueness conflicts in block 1014, and to detect and resolve delete conflicts in block 1016. The details of the processing performed in block 1010 are shown and described in the following sections.
Detecting and Resolving Update Conflicts Column Groups
The present invention uses the concept of a column group to detect and resolve update conflicts. A column group links a collection of columns in a table to a single "logical column". A column group can consist of any number of columns, from a single column, to all of the columns in a table. Each column, however, can belong to only one column group.
Because the update conflict detection mechanism or module of the preferred embodiment detects conflicts on a column group by column group basis, all columns must be a part of some column group. However, all of the columns in a table need not be assigned to a column group. Any column not assigned to a column group is automatically assigned by the present invention to a "shadow" column group for the purpose of conflict detection. This shadow column group is not user visible, and you cannot assign a conflict resolution method to the columns in the group. You can only designate a conflict resolution method for columns in a user-defined column group.
By defining column groups, it is possible to associate different methods of resolving conflicts for different types of data. For example, numeric data lends itself well to some sort of arithmetic method of resolving conflicts, while conflicts with character data might be better resolved using a timestamp to apply the most recent change.
Because each column group is evaluated individually, portions of a row in a table may be updated using the data from the originating site, while other portions may maintain the values of the data at the destination site. That is, although the use of a conflict resolution mechanism may result in convergence
(ultimately all sites have the same values for a row), with multiple column groups, it might not result in data integrity. Thus, if two or more columns in a table must remain consistent (for example, if multiple columns are used to store address information), these columns should be placed in the same column group.
Referring now to Figures 4 and 5, the processing logic for detecting and resolving update conflicts in a column group is illustrated. There is a change indicator for every column group defined for a table. For every row updated, the change indicators are used to show which column groups have been updated. When an update to a row is replicated from the originating site to the receiving site, the change indicators for the row are forwarded to the receiving site. The receiving site is then able to determine which column groups in the row have been updated at the originating site (Blocks 410 and 412 in Figure 4). Not all old and new values of all columns in a replicated row are copied and forwarded to the receiving site. Only the old and new values of a column group that has been updated are copied and forwarded to the receiving site. However, the old primary key values for the row are always copied and forwarded to the receiving site. At the receiving site, the old primary key values of the replicated row are used to uniquely identify the row to apply the update. If there is no row or more than one row has been identified, then an unresolvable update conflict has occurred. When an unresolvable update conflict occurs, an error is raised, and the control is returned to the transaction that replicates the row. Once the receiving site has uniquely identified the row to apply the update, it uses the change indicators to identify the column groups that have been changed at the originating site. For those column groups whose values have been modified at the originating site, it compares the old values of each of these column groups with the current values of the same column group at the receiving site.
A column group has an update conflict if both the originating site and the receiving site have updated the column group before propagating their updates to each other. The update conflict for a column group is detected by comparing the old values from the originating site with the current values at the receiving site. If the old values do not match the current values, there is an update conflict. Conversely, if the old values match the current values, there is no update conflict (Block 414).
When there is no update conflict for a column group, the current values at the receiving site are updated with the new values from the originating site (blocks 428 and 526).
Once an update conflict has been detected for a column group (Path 416), the update conflict resolution routines that have been assigned to the column group are applied to resolve the update conflict. The conflict resolution routines are applied one by one in priority order until the conflict is resolved (Block 420 in Figure 4 and Block 516 in Figure 5). The conflict is resolved when one of the conflict resolution routines can determine the appropriate new values for the column group (Path 426). When the update conflict is resolved for the column group, the current values of the column grouped at the receiving site are updated with the newly determined values (blocks 428 and 526). On the other hand, if there is no conflict resolution routine assigned to the column group, or none of the assigned conflict resolution routines can resolve the update conflict (Path 512 in Figure 5), an error is raised, and the control is returned to the transaction that replicates the row.
For optimization, the current values of all column groups at the receiving site can be updated with their new values using a single UPDATE command, instead of individual UPDATE commands from each column group. For further optimization, the receiving site does not lock the row for update that it has uniquely identified to apply the update. As a result, any local running transactions at the receiving site can potentially introduce new update conflicts while current update conflicts are being resolved. Thus, if both optimization approaches are being used, the receiving site must check for any new update conflict after it has resolved the current update conflicts.
Detecting and Resolving Uniqueness Conflicts
A database table can have one or more unique constraints. A unique constraint can be on one or more columns. A uniqueness constraint enforces that the values of those columns are unique in the database table.
When an insert of a new row is replicated from the originating site to the receiving site, the new values of all columns are copied and forwarded to the receiving site.
A uniqueness conflict occurs when a uniqueness constraint is violated during an insert or update of the replicated row at the receiving site. The uniqueness conflict is detected by comparing the new values from the originating site with the current values of all existing rows at the receiving site. The comparison is performed by each uniqueness constraint. There is a uniqueness conflict for a uniqueness constraint if there is already an existing row at the receiving site whose current values match the new values of those unique columns. For optimization, uniqueness conflicts can be detected by trapping exceptions from the violations of unique indexes or uniqueness constraints.
Once a uniqueness conflict has been detected for a uniqueness constraint, the conflict resolution routines that have been assigned to the uniqueness constraint are applied to resolve the uniqueness conflict. The conflict resolution routines are applied one by one in priority order until the conflict is resolved.
The conflict is resolved when one of the conflict resolution routines can determine the appropriate new values for the unique columns from the originating site. On the other hand, if there are no conflict resolution routines assigned to the uniqueness constraint, or none of the assigned conflict resolution routines can resolve the uniqueness conflict, an error is raised, and control is returned to the transaction that replicates the row. When all uniqueness conflicts have been resolved, the replicated row with any newly determined new column values is then applied to the database table. Any local running transactions at the receiving site can potentially introduce new uniqueness conflicts while current uniqueness conflicts are being resolved. Thus, the receiving site must check for any new uniqueness conflicts after it has resolved all the current uniqueness conflicts.
Detecting and Resolving Delete Conflicts
When the deletion of a row is replicated from the originating site to the receiving site, the old values of all columns are copied and forwarded to the receiving site.
At the receiving site, the old primary key values of the deleted row from the originating site are used to uniquely identify the row to delete. If there is no row or more than one row has been identified, then an unresolvable delete conflict has occurred. When an unresolvable delete conflict occurs, an error is raised, and control is returned to the transaction that replicates the delete.
Once the receiving site has uniquely identified the row to delete, it compares the old values of the deleted row with the current values of the selected row. If the old values match the current values, the selected row is deleted from the receiving site. On the other hand, if the old values do not match the current values, a delete conflict is detected.
Once a delete conflict has been detected for a database table, the delete conflict resolution routines that have been assigned to the database table are applied to resolve the delete conflict. The conflict resolution routines are applied one by one in priority order until the conflict is resolved. If there is no delete conflict resolution routine, or none of the assigned delete conflict resolution routines can resolve the delete conflict, an error is raised, and control is returned to the transaction that replicates the delete. Any local running transactions at the receiving site can potentially introduce a new delete conflict while the current delete conflict is being resolved. Thus, the receiving site must check for any new delete conflict after it has resolved the current delete conflict.
Declarative Conflict Resolution Methods
The present invention provides various declarative conflict resolution methods for each type of conflict. The user can choose and declare one or more conflict resolution methods to be used for each possible conflict. The user can also declare the priority order of applying the resolution methods when there is more than one resolution method for a possible conflict.
The standard declarative conflict resolution methods that are provided by the present invention for each type of conflict are described in the following sections. It will be apparent to one of ordinary skill in the art that other conflict resolution methods for each type of conflict may be equivalently defined. One of the standard declarative conflict resolution methods is the "User
Function" method. The "User Function" method provides user extensibility to standard conflict resolution routines that are provided by the present invention for update conflicts, uniqueness conflicts and delete conflicts. The user can write their own conflict resolution routines and use them together with the standard conflict resolution routines. When the user wants to use a user-defined conflict resolution routine for a possible conflict, they declare a "User Function" method for the possible conflict and specify the user-defined conflict resolution routine as the User Function.
The "User Function" method can also be used by the user to do other useful tasks such as sending out notifications when a conflict cannot be resolved. The user writes a notification routine and registers it as the User Function for the "User Function" method.
The standard conflict resolution routines do not support all cases in the preferred embodiment. For example, they do not support the following situations: 1) delete conflicts, 2) changes to a primary key, 3) NULLs or undefined values in the columns designated to be used to resolve the conflict, or 4) referential integrity constraint violations. For these situations, users must write their own conflict resolution routines, or define a method of resolving the conflicts after the errors have been raised and logged in some error tables.
Standard Resolution Methods for Update Conflicts The preferred embodiment provides the following methods of resolving update conflicts in replicated environments with any number of master sites. It will be apparent to those of ordinary skill in the art that other update conflict methods are suggested by this disclosure of the present invention. The methods are: 1) apply the data with the latest timestamp, 2) apply all data additively, 3) apply the minimum value, when the column value is always decreasing, and 4) apply the maximum value, when the column value is always increasing.
There are several additional methods provided that can be used to resolve conflicts in replicated environments with no more than two master sites: 1) apply the data with the earliest timestamp, 2) apply the minimum value, 3) apply the maximum value, 4) apply the value from the site with the highest priority, and
5) apply the value assigned the highest priority. There are several methods that can be used to resolve conflicts in replicated environment with only one master site and multiple updatable snapshot sites: 1) average the values, 2) discard the values from the snapshot sites, or 3) overwrite the values at the master sites.
Finally, there is the "User Function" method that the user uses to resolve an update conflict with a user-defined conflict resolution routine. These methods are described in more detail below.
Standard Resolution Methods for Unique Constraint Conflicts
The preferred embodiment provides a variety of methods for resolving uniqueness conflicts: 1 ) append the global name of the originating site to the column value, 2) append a generated sequence number to the column value, or 3) discard the new transaction from the remote site.
If there is more than one master site, none of these routines result in convergence and they should only be used with some form of notification facility.
There is the "User Function" method that the user can use to resolve a uniqueness conflict with a user-defined conflict resolution routine.
Standard Resolution Method for Delete Conflicts The preferred embodiment provides one standard resolution method for resolving delete conflicts, which is the "User Function" method. Although the present invention has not provided any standard conflict resolution routine for a delete conflict, the delete conflict can still be detected, and resolved with a user- defined conflict resolution routine. The "User Function" method allows the user to register a user-defined conflict resolution routine for a potential delete conflict.
Avoiding Ordering Conflicts
As the above list indicates, many forms of conflict resolution cannot guarantee convergence if the replicated environment contains more than two masters. As shown in Figure 6, network failures and infrequent pushing of the deferred remote procedure call queue increase the chance of non-convergence there are more than two masters.
In the example shown in Figure 6, changes made at site A are given priority over conflicting changes at site B, and changes from site B have priority over changes from site C. Figure 6 illustrates the actions that may occur over the course of the day in a replicated environment that may lead to the inability of the multiple sites to make conflicts converge.
These types of ordering conflicts can be avoided when using priority groups if it is required that the flow of ownership be ordered, as it is in the work flow model. That is, information must always flow from the ORDER site to the SHIP site to the BILL site, in a typical business example. If the billing site receives a change to a row from the ordering site after it has already received a change to that row from the shipping site, for example, the billing site will know to ignore the out-of-order change from the ordering site.
Summary of Standard Conflict Resolution Methods
Figure 7 summarizes the standard conflict resolution methods provided in the preferred embodiment of the present invention. This figure also shows if these resolution methods can guarantee convergence (all sites ultimately agreeing on the same value) between multiple master sites and their associated snapshot sites. Each of these methods is explained in greater detail in the following sections.
Minimum and Maximum Update Conflict Methods
When the present invention detects a conflict with a column group and calls the minimum value conflict resolution routine, the processing logic compares the new value from the originating site with the current value at the receiving site for a designated column in the column group. This column must be designated when the minimum value conflict resolution method is selected.
If the two values for the designated column are the same or one of the two values is undefined, that is, NULL (for example, if the designated column was not the column causing the conflict), the conflict is not resolved, and the values of the columns in the column group remain unchanged. If the new value of the designated column is less than the current value, the column group values from the originating site are applied at the receiving site. If the new value of the designated column is greater than the current value, the conflict is resolved by leaving the current column group values unchanged. The maximum value method works exactly the same as the minimum value method, except that the values from the originating site are only applied if the value of the designated column at the originating site is greater than the value of the designated column at the receiving site.
There are no restrictions on the datatypes of the columns in the column group. For minimum value, convergence for more than two master sites is only guaranteed if the value of the designated column is always decreasing. For maximum value, the column value must always be increasing. Earliest and Latest Timestamp Update Conflict Resolution Methods
The earliest and latest timestamp methods are simply variations on the minimum and maximum value methods. For the timestamp method, the designated column must be of the well known type DATE. Whenever any column in a column group is updated, a conventional application updates the value of this timestamp column with the local SYSTEM DATE (SYSDATE). For a change that is applied from another site, the timestamp value should be set to the timestamp value from the originating site. For example, suppose that a customer calls his local sales representative to update his address information. After hanging up the phone, he realizes that he gave the wrong zip code. When he tries to call his sales representative with the correct zip code, he discovers that the phone lines in that area have gone down, so he calls the headquarters number instead, to update his address information again. When the network connecting the headquarters site with the sales site comes back up, the present invention will see two updates for the same address, and detect a conflict. By using the latest timestamp method, the present invention would select the later update, and apply the address with the correct zip code. If the replicated environment crosses time zones, all timestamps should be converted to a common time zone. Otherwise, although the data will converge, the most recent update may not be applied as expected.
The earliest timestamp method applies the changes from the site with the earliest timestamp, and the latest timestamp method applies the changes from the site with the latest timestamp. A backup method, such as site priority, should be designated to be called in case two sites have the same timestamp. The timestamping mechanism should be designed to be time zone independent (for example, by always converting the timestamp to a designated time zone, such as GMT).
Because the maximum value method can guarantee convergence if the value is always increasing, the latest timestamp method can guarantee convergence (assuming that you have properly designed your timestamping mechanism). The earliest timestamp method cannot guarantee convergence for more than two masters (since time is generally not always decreasing).
The earliest and latest timestamp methods require auditing of update timestamps. Users specify whether the present invention automatically audits update timestamps or whether an application explicitly audits update timestamps. If the application audits update timestamps, users must specify which column in the column group stores the last update timestamp. The datatype of the column must be of date.
The auditing column stores the last update timestamp of any column in the column group. The auditing column always stores the timestamp when the column value(s) were first created. For example, if the earliest change is from a remote site, this method results in overwriting the current column value with the replicated column value from the remote site, and updating the auditing column with the timestamp when the replicated column value was first created, that is, from the remote site.
Additive and Average Update Conflict Resolution Methods
The additive and average conflict resolution routines work with column groups consisting of a single numeric column only. The additive routine adds the difference between the old and new values at the originating site to the current value at the receiving site as follows: current value = current value + (new value - old value)
The additive conflict resolution method provides convergence for any number of master sites.
The average conflict resolution method averages the new column value from the originating site with the current value at the receiving site as follows: current value = (current value + new value )/2 The average method cannot guarantee convergence if the replicated environment has more than one master. This method is useful for an environment with a single master site and multiple updatable snapshots.
Priority Group and Site Priority Update Conflict Resolution Methods Priority groups allow a user to assign a priority level to each possible value of a particular column. As shown in the example illustrated in Figures 8 and 9, a Priority view 910 shows the priority level 914 assigned to each value 916 that the "priority" column can contain. A priority level for all possible values of the "priority" column must be specified. When the priority group method of conflict resolution is selected for a column group, the user designates which column in the column group is the "priority " column. When the priority group conflict resolution routine is invoked to resolve an update conflict in a column group, the processing logic compares the priority level of the new value from the originating site with that of the current value at the receiving site for the designated "priority" column in the column group.
If the two priority levels are the same, or one of them is undefined, the conflict is not resolved, and the current column group values at the receiving site remain unchanged. If the priority level of the new value in the designated "priority" column is greater than that of the current value at the receiving site, the current column group values at the receiving site are updated with the new column group values from the originating site. Conversely, if the priority level of the new value in the designated "priority" column is less than that of the current value at the receiving site, the conflict is resolved by leaving the current column group values at the receiving site unchanged. It will be apparent to one of ordinary skill in the art that the priority group conflict resolution routine may equivalently interpret that a lower priority level takes precedence over a higher priority level.
The Priority view 910 displays the values of all priority groups 912 defined at the current location. In the example shown in Figure 8, there are two different priority groups, "site-priority" and "order-status". The CREDIT_STATUS table 920 is shown in Figure 9 as using the "site-priority" priority group 912; because the site values 918 in the CREDIT_STATUS table 920 correspond to the "site-priority" values 916 shown in Figure 8.
Site priority is a special kind of priority group. With site priority, the "priority" column designated is automatically updated with the global database name of the site where the update originated. The Priority view 910 in Figure 8 shows the priority level 914 assigned to each database site 916. Site priority can be useful if one site is considered to be more likely to have the most accurate information. For example, in Figures 8 and 9, the Houston site (houston.world) may be the corporate headquarters, while the New York site (new_york. world) may be a sales office. The headquarters office (houston.world), in this example, is considered more likely to have the most accurate information about the credit that can be extended to each customer. Thus, the headquarters office (houston.world) priority is set higher (" 1 ") than the sales office (new_york.world) priority ("2").
When using site priority, convergence with more than two masters is not guaranteed. You can guarantee convergence with more than two masters when you are using priority groups, however, if the value of the "priority" column is always increasing. That is, the values in the priority column correspond to an ordered sequence of events; for example: ordered, shipped, billed.
The site priority method requires auditing of the global name of the site that applied the last update. Users specify whether the present invention or an application maintains the audit information. If the application maintains the audit information, users must specify which column in the column group stores the site global name. The column specified always stores the global name of the site where the column value(s) were first created.
Overwrite and Discard Update Conflict Resolution Methods.
The overwrite and discard methods ignore the values from either the originating or receiving site and therefore can never guarantee convergence with more than one master site. These methods are designed to be used by a single master site and multiple snapshot sites, or with some form of a user-defined notification facility. For example, if you have a single master site that you expect to be used primarily for queries, with all updates being performed at the snapshot sites, you might select the overwrite method. These methods are also useful if your primary concern is data convergence (and you have a single master site), and there is no particular business rule for selecting one update over the other. You may even choose to use one of these methods if you have multiple master sites, provided that you supply a notification facility. This allows you to notify the appropriated personnel, who will be responsible for ensuring that the data is correctly applied, instead of logging the conflict in the some error tables and leaving the resolution up of a local database administrator.
The overwrite routine overwrites the current value at the receiving site with the new value from the originating site. Conversely, the discard method 32 discards the new value from the originating site.
Append Site Name/Append Sequence Uniqueness Conflict Resolution Methods
The append site name and append sequence conflict resolution routines work by appending a string to a column whose unique constraint has been violated, that is, duplicate column values have been detected. If the unique constraint is on more than one column, the user must specify one of the columns as the column to which the string is appended. Although this allows the column to be inserted or updated without violating a unique integrity constraint, it does not provide any form of convergence between multiple master sites. The resulting discrepancies must be manually resolved; therefore, these methods are meant to be used with some form of a notification facility. Both methods can be used on character columns only in the preferred embodiment.
These methods can be useful when the availability of the data may be more important than the complete accuracy of the data. By selecting one of these methods, used with a notification scheme, instead of logging a conflict, you allow the data to become available as soon as it is replicated. By notifying the appropriate personnel, you ensure that the most knowledgeable person resolves the duplication. When a uniqueness conflict is encountered, the append site name routine appends the global database name of the site originating the transaction to the new column value from the originating site. Similarly, the append sequence routine appends a generated sequence number to the new column value from the originating site.
Discard Uniqueness Conflict Resolution Method
This conflict resolution routine resolves uniqueness conflicts by simply discarding the row from the originating site. This method never guarantees convergence with multiple masters and should be used with a notification facility.
Compared to the append methods, this method is useful if you want to minimize the propagation of data until its accuracy can be verified.
User Function Update Conflict Resolution Method
Users can also create their own specific conflict resolution functions. Configurable conflict resolution allows users to freely mix their own functions with the standard conflict resolution routines that are provided by the present invention.
The present invention does not impose any restrictions on user functions. But, it requires that user functions adhere to the following interface specification in the preferred embodiment:
• The user conflict resolution function should return TRUE if it has successfully resolved the conflict.
• The user function should return FALSE if it has not successfully resolved the conflict. • The user function should accept column values as parameters.
• The user function should accept old, new and current columns values. The old, new and current values for a column are received consecutively. If it can resolve the conflict, it modifies only the new column values so that the replication procedure can update the current row with the new column values.
• Parameters that accept new column values should use IN OUT parameter mode. Others should use IN parameter mode. • The last parameter of the function must be a boolean flag.
The function sets the flag to TRUE if it wants to discard the new column values; otherwise it sets the flag to FALSE.
Users specify which columns in the table constitute the parameters for each user function.
Users should avoid using the following commands in their conflict resolution functions. While configurable conflict resolution does not prohibit these commands, the outcome of replication becomes unpredictable if user functions use them. User functions should avoid the following: • Data Definition Language commands
• Transaction Control commands
• Session Control commands
• System Control commands
Note that any transaction or data change made by the user function will not automatically be replicated.
User Function Uniqueness Conflict Resolution Method
Users can create their own conflict resolution functions. Configurable conflict resolution allows users to freely mix their own functions with the standard conflict resolution routines that are provided by the present invention.
The present invention does not impose any restrictions on user functions. But, it requires that user functions adhere to the following interface specification:
• The function should return TRUE if it has successfully resolved the conflict.
• The function should return FALSE if it has not successfully resolved the conflict. • The function should have all parameters with IN OUT mode.
• The function should accept column values as parameters.
• The function should accept new column values. If it can resolve the conflict, it modifies the new column values so that the replication procedure can insert or update the current row with the new column values.
• The last parameter of the function must be a boolean flag. The function sets the flag to TRUE if it wants to discard the new column values; otherwise it sets the flag to FALSE. Users specify which columns in the table constitute the parameters for each user function.
Users should avoid using the following commands in their conflict resolution functions. While automatic conflict resolution does not prohibit these commands, the outcome of replication becomes unpredictable if user functions use them. User functions should avoid the following:
• Data Definition Language commands
• Transaction Control commands
• Session Control commands System Control commands Note that any transaction or data change made by the user function will not automatically be replicated.
User Function Delete Conflict Resolution Method
Users can create their own conflict resolution functions. Configurable conflict resolution allows users to freely mix their own functions.
The present invention does not impose any restrictions on user functions. But, it requires that user functions adhere to the following interface specification:
• The function should return TRUE if it has successfully resolved the conflict.
• The function should return FALSE if it has not successfully resolved the conflict.
• The function should have all parameters with IN OUT mode.
• The function should accept column values as parameters.
• The function should accept old column values. If it can resolve the conflict, it modifies the old column values so that the replication procedure can delete the current row that matches all old column values.
• The last parameter of the function must be a boolean flag. The function sets the flag to TRUE if it wants to discard the old column values; otherwise it sets the flag to FALSE. Users specify which columns in the table constitute the parameters for each user function.
Users should avoid using the following commands in their conflict resolution functions. While automatic conflict resolution does not prohibit these commands, the outcome of replication becomes unpredictable if user functions use them. User functions should avoid the following:
• Data Definition Language commands
• Transaction Control commands
• Session Control commands
• System Control commands Note that any transaction or data changes made by the user function will not automatically be replicated. Conflict Resolution Processing
The following is an example illustrating the preferred embodiment of the present invention. This example is a demonstration of the process of resolving both update and uniqueness conflicts that can occur for an update of a row. Referring to Figures 11-14, flowcharts illustrate the logic steps performed by the processing logic of the present invention for resolving conflicts. Figures
15-21 are flowcharts illustrating the logic steps performed by the processing logic of the present invention for resolving conflicts with the automatic generation of audit information. These flowcharts describe a transaction between two sites (Master site I and Master site II) and illustrate the conflict resolution features of the present invention. These flowcharts are described in detail below.
Referring to Figure 11, a table (called "emp" in this example) is updated in the first site or Master site I (block 1 110). A corresponding trigger (emp$rt) is fired when the table is updated. The OLD and NEW values (values prior to the update and values subsequent to the update, respectively) of the table are written to a transfer buffer (DefTran) in the Master site I in block 1112. A Master Site II replication update routine (emp$rp.rep_update) is called in block 1114 and processing for Master site I ends in Figure 11. The perspective of this description then shifts to the point of view of the Master site II as shown starting in Figure 12.
Referring now to Figure 12, the "emp" table in Master site II is updated in block 1210. The "emp" table in each of the Master sites I and II are replicated copies of each other that may be updated independently at each site where it resides. After the table is updated in block 1210, a series of tests are executed to determine if the update caused any conflicts. First, a test is performed to determine if an update conflict is present. If so, the no_data__found exception is raised in block 1214 and processing continues at the bubble labeled N shown in Figure 13. If a uniqueness conflict occurred, the dup_val_on_index exception is raised in block 1218 and processing continues at the bubble labeled P shown in Figure 14. If some other exception occurred, the other exception is raised in block 1222 and processing continues at the bubble labeled R shown in Figure 13. If no exception is detected, processing continues at the bubble labeled S shown in Figure 13.
Referring now to Figure 13, the processing performed when an update conflict is detected is shown. This processing implements the logic for detecting and resolving conflicts shown in Figures 4 and 5, which are described above. Referring again to Figure 13 , an update conflict handler (emp$rr.update_conflict_handler in this example) is called in the Master site II in block 1310. The current table values in Master site II are queried in block 1312. The column groups are tested to determine which column group has an update conflict in block 1314. The corresponding update conflict resolution routine(s) are called in block 1316. If the update conflict is resolved, block 1320 in Master site II is executed to call the replication update routine emp$rp.rep_update in this example. Processing then continues at the bubble labeled E shown in Figure 12 where the emp table is again recursively checked for update conflicts. If in block 1318 the update conflict is not resolved, an error is posted in block 1322 and processing terminates at the End bubble shown in Figure 13.
Referring now to Figure 14, the processing performed when a uniqueness conflict is detected is shown. First, a uniqueness conflict handler (emp$rr.unique_conflict_update_handler in this example) is called in the Master site II in block 1410. The uniqueness constraint that was violated is determined in block 1414. The corresponding uniqueness conflict resolution routine(s) are called in block 1416. Processing continues at the bubble labeled T shown in Figure 13 where a test is executed to determine if the uniqueness conflict was resolved.
Figures 15-21 are flowcharts illustrating the logic steps performed by the processing logic of the present invention for resolving conflicts with the automatic generation of audit information. Referring to Figure 15, a table (called "emp" in this example) is updated in Master site I (block 1510). A corresponding trigger (emp$rt) is fired when the table is updated. The OLD and NEW values of the table are written to a transfer buffer (DefTran) in the Master site I in block 1512. Audit columns of the table (emp) are queried for the OLD audit values (block 1514). The OLD audit values and NEW audit values are written to the transfer buffer DefTran in block 1516. A Master Site II replication update routine (emp$rp.rep_update) is called in block 1518 and processing for Master site I ends in Figure 15. The perspective of this description then shifts to the point of view of the Master site II for the automatic generation of audit information as shown starting in Figure 16.
Referring now to Figure 16, a shadow table (emp$ra) corresponding to the emp table is updated in the Master site II in block 1609. If an update conflict (block 1611) is found, processing continues at the bubble labeled J shown in Figure 18. If a uniqueness conflict (block 1613) in the shadow table is found, processing continues at the bubble labeled G shown in Figure 17 where the emp table is checked for conflicts. If another exception is found (block 1614), processing continues at the bubble RR shown in Figure 20. Otherwise, processing continues at bubble G shown in Figure 17.
Referring now to Figure 17, the table ("emp" in this example) is checked for conflicts. The "emp" table in Master site II is updated in block 1710. After the table is updated in block 1710, a series of tests are executed to determine if the update caused any conflicts. First, a test is performed to determine if an update conflict is present. If so, the no_data_found exception is raised in block 1714 and processing continues at the bubble labeled J shown in Figure 18. If a uniqueness conflict occurred, the dup_val_on_index exception is raised in block 1718 and processing continues at the bubble labeled K shown in Figure 19. If some other exception occurred, the other exception is raised in block 1722 and processing continues at the bubble labeled RR shown in Figure 20. If no exception is detected, processing continues at the bubble labeled SS shown in Figure 20.
Referring now to Figure 18, the processing performed when an update conflict is detected in the audit context is shown. First, the update to the audit table (emp$ra) is rolled back in block 1809. Next, processing then continues at the bubble labeled NN shown in Figure 20 where the emp table is processed for update conflicts.
Referring now to Figure 19, the processing performed when a uniqueness conflict is detected in the audit context is shown. First, the update to the audit table (emp$ra) is rolled back in block 1909. Next, processing then continues at the bubble labeled PP shown in Figure 21 where the emp table is processed for uniqueness conflicts. Once the conflict is resolved or the determination is made that the conflict cannot be resolved, processing is terminated as shown in Figure 20.
Referring now to Figure 20, the processing performed when an update conflict is detected is shown with the automatic generation of audit information selected. First, an update conflict handler (emp$rr.update_conflict_handler, in this example) is called in the Master site II in block 2010. The current table values in Master site II are queried in block 2012. The column groups are tested to determine which column group has an update conflict in block 2014. The corresponding update conflict resolution routine(s) are called in block 2016. If the update conflict is resolved, block 2020 in Master site II is executed to call the replication update routine emp$rp.rep_update, in this example. Processing then continues at the bubble labeled EE shown in Figure 16 where the emp table is again recursively checked for update conflicts. If in block 2018 the update conflict is not resolved, an error is posted in block 2022 and processing terminates at the End bubble shown in Figure 20.
Referring now to Figure 21 , the processing performed when a uniqueness conflict is detected is shown with the automatic generation of audit information se le cted . Fi rs t , a u n i q u e ne s s c onfl i c t han dl e r (emp$rr.unique_conflict_update_handler, in this example) is called in the Master site II in block 2110. The uniqueness constraint that was violated is determined in block 2114. The corresponding uniqueness conflict resolution routine(s) are called in block 2116. Processing continues at the bubble labeled TT shown in Figure 20 where a test is executed to determine if the uniqueness conflict was resolved.
This completes the conflict resolution process.
Automatic Generation of Audit Information
Some conflict resolution methods require additional information typically stored in an audit table, such as last update timestamp or last update site. The conflict resolution methods that require such audit information are listed below. A description of the audit information used by these methods is set forth above.
• Earliest timestamp
• Latest timestamp • Site Priority
When users use one of these conflict resolution methods, they have the option of generating and maintaining the audit information themselves in existing columns of the same table. If users do not specify which existing columns in the table store the audit information, the present invention will automatically generate and maintain the audit information in a shadow table. If users do specify which existing columns in the table store the audit information, the present invention will assume that the application will generate and maintain the audit information.
Every replicated table can have one and only one shadow table in the preferred embodiment. The present invention creates a shadow table if none exists, or modifies the table definition if the generated shadow table already exists. Once a shadow table has been created, the present invention can only add new columns to the shadow table. It is possible that the present invention may fail to add new columns to a shadow table if there are obsolete columns in the shadow table. Some columns become obsolete when users change their use of conflict resolution methods that require automatic generation of audit information. It is recommended that the database administrator periodically removes obsolete columns from the shadow tables.
The present invention treats a shadow table as if it were an extension of the replicated table. Because of this, the "shadow table" for snapshot replicated object must be a snapshot at the snapshot site and must be in the same refresh group as the snapshot replicated object.
A shadow table has the following columns:
• The primary key columns of the table.
• Optionally a date column for every column group that requires auditing of timestamp.
• Optionally a varchar2(128) column for every column for every column group that requires auditing of site's global database name. In the preferred embodiment of the present invention, the name of the shadow table is derived by appending $ra to the table name. The present invention truncates the table name to accommodate the additional characters when necessary. If the derived name conflicts with an existing one, the present invention uses an arbitrary name with $ra appended.
The present invention uses the exact primary key column names in the shadow table. As for other columns in the shadow table, the present invention derives their names according to the type of audit information. The present invention uses "TIMESTAMP" as the name of a column used for storing timestamp, and "GLOBAL_NAME" as the name of a column used for storing site's global database name. The present invention appends a unique integer to the column name if that column name already exists.
Thus, an apparatus and method for providing adaptable and configurable conflict resolution within a replicated data environment is disclosed. Numerous modifications in form and detail may be made by those of ordinary skill in the art without departing from the scope of the present invention. Although this invention has been shown in relation to a particular preferred embodiment, it should not be considered so limited. Rather, the present invention is limited only by the scope of the appended claims.

Claims

We claim:
1. In a distributed database system having a first node and a second node, said first node including a first data structure, said second node including a second data structure, a configurable conflict resolution apparatus comprising: a conflict detection module for detecting a conflicting modification to corresponding portions of said first and said second data structures;
a plurality of conflict resolution methods, one or more of said plurality of conflict resolution methods being configurably associated with said corresponding portions of said first and said second data structures; and
a conflict resolution module for activating a first conflict resolution method of said one or more of said plurality of conflict resolution methods when said conflict detection module detects said conflicting modification to said corresponding portions of said first and said second data structures.
2. The configurable conflict resolution apparatus as claimed in Claim 1 wherein said conflict resolution module further includes processing logic for determining if said first conflict resolution method successfully resolved said conflicting modification.
3. The configurable conflict resolution apparatus as claimed in Claim 2 wherein said conflict resolution module further includes processing logic for activating a second conflict resolution module if said first conflict resolution method was not successful in resolving said conflicting modification.
4. The configurable conflict resolution apparatus as claimed in Claim 2 wherein said conflict resolution module further includes processing logic for activating each of said one or more of said plurality of conflict resolution methods until one or more of said conflict resolution methods is successful in resolving said conflicting modification or until all conflict resolution methods have been tried at least once.
5. The configurable conflict resolution apparatus as claimed in Claim 4 wherein said conflict resolution module further includes processing logic for storing non- conflicting values in said corresponding portions of said first and said second data structures if one of said conflict resolution methods was successful in resolving said conflicting modification.
6. The configurable conflict resolution apparatus as claimed in Claim 5 further including a means for deferring storing of non-conflicting values until after a complete portion of said first and said second data structures is checked for other conflicting modifications.
7. The configurable conflict resolution apparatus as claimed in Claim 4 wherein said conflict resolution module further includes processing logic for logging an error if none of said conflict resolution methods was successful in resolving said conflicting modification.
8. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is registered by a user.
9. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is a minimum value method.
10. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is a maximum value method.
11. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is an earliest timestamp method.
12. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is a priority group method.
13. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is a site priority method.
14. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is an overwrite method.
15. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is a discard method.
16. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is an average value method.
17. The configurable conflict resolution apparatus as claimed in Claim 1 wherein at least one of said plurality of conflict resolution methods is an additive value method.
18. The configurable conflict resolution apparatus as claimed in Claim 1 wherein said conflicting modification is an update conflict.
19. The configurable conflict resolution apparatus as claimed in Claim 1 wherein said conflicting modification is a uniqueness conflict.
20. The configurable conflict resolution apparatus as claimed in Claim 1 wherein said conflicting modification is a delete conflict.
21. The configurable conflict resolution apparatus as claimed in Claim 1 further including audit logic for automatically generating and maintaining audit information in a shadow data structure.
22. The configurable conflict resolution apparatus as claimed in Claim 1 further including recursion logic for recursively activating said first conflict resolution method until no further modification to said corresponding portions of said first and said second data structures is detected.
23. The configurable conflict resolution apparatus as claimed in Claim 4 wherein a priority is assigned to each of said plurality of conflict resolution methods and said conflict resolution module activates each of said plurality of conflict resolution methods in priority order.
24. In a distributed database system having a first node and a second node, said first node including a first data structure, said second node including a second data structure, a method for configurable conflict resolution comprising the steps of: detecting a conflicting modification to corresponding portions of said first and said second data structures; providing a plurality of conflict resolution methods, one or more of said plurality of conflict resolution methods being configurably associated with said corresponding portions of said first and said second data structures; and
activating a first conflict resolution method of said one or more of said plurality of conflict resolution methods when said conflicting modification is detected in said detecting step.
25. For use in a distributed database system having a first node and a second node, said first node including a first data structure, said second node including a second data structure, an article of manufacture comprising a computer usable mass storage medium having computer readable program code embodied therein for causing a processing means to perform configurable conflict resolution, said computer readable program code in said article of manufacture comprising: a conflict detection module for causing said processing means to detect a conflicting modification to corresponding portions of said first and said second data structures;
a plurality of conflict resolution methods, one or more of said plurality of conflict resolution methods being configurably associated with said corresponding portions of said first and said second data structures; and
a conflict resolution module for causing said processing means to activate a first conflict resolution method of said one or more of said plurality of conflict resolution methods when said conflict detection module detects said conflicting modification to said corresponding portions of said first and said second data structures.
PCT/US1997/004399 1996-03-19 1997-03-19 Configurable conflict resolution in a computer implemented distributed database WO1997035270A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
AU23353/97A AU2335397A (en) 1996-03-19 1997-03-19 Configurable conflict resolution in a computer implemented distributed database
CA002249422A CA2249422C (en) 1996-03-19 1997-03-19 Configurable conflict resolution in a computer implemented distributed database
GB9819943A GB2326495B (en) 1996-03-19 1997-03-19 Configurable conflict resolution in a computer implemented distributed database

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/618,507 1996-03-19
US08/618,507 US5806074A (en) 1996-03-19 1996-03-19 Configurable conflict resolution in a computer implemented distributed database

Publications (2)

Publication Number Publication Date
WO1997035270A1 true WO1997035270A1 (en) 1997-09-25
WO1997035270A9 WO1997035270A9 (en) 1998-02-12

Family

ID=24478006

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1997/004399 WO1997035270A1 (en) 1996-03-19 1997-03-19 Configurable conflict resolution in a computer implemented distributed database

Country Status (5)

Country Link
US (2) US5806074A (en)
AU (1) AU2335397A (en)
CA (1) CA2249422C (en)
GB (1) GB2326495B (en)
WO (1) WO1997035270A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6138143A (en) * 1999-01-28 2000-10-24 Genrad, Inc. Method and apparatus for asynchronous transaction processing
US7149759B2 (en) 2002-03-25 2006-12-12 International Business Machines Corporation Method and system for detecting conflicts in replicated data in a database network
EP3155527A4 (en) * 2014-07-02 2018-04-18 Pure Storage, Inc. Redundant, fault-tolerant, distributed remote procedure call cache in a storage system
WO2019217481A1 (en) * 2018-05-07 2019-11-14 Microsoft Technology Licensing, Llc Conflict resolution for multi-master distributed databases
US10540340B2 (en) 2016-02-09 2020-01-21 International Business Machines Corporation Performing conflict analysis of replicated changes among nodes in a network
US11126610B1 (en) * 2017-11-22 2021-09-21 Amazon Technologies, Inc. Conflict resolution in a data proxy
WO2022272084A1 (en) * 2021-06-25 2022-12-29 Bristol-Myers Squibb Company Indicating differences in and reconciling data stored in disparate data storage devices

Families Citing this family (262)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6105025A (en) * 1996-03-08 2000-08-15 Oracle Corporation Method for using an index as a workspace for deferred enforcement of uniqueness constraints
US6412017B1 (en) * 1996-07-01 2002-06-25 Microsoft Corporation Urgent replication facility
US6125408A (en) * 1997-03-10 2000-09-26 Compaq Computer Corporation Resource type prioritization in generating a device configuration
US5964871A (en) * 1997-03-10 1999-10-12 Compaq Computer Corporation Resolution of resource conflicts by reduction of systems to solve
FI105870B (en) * 1997-03-21 2000-10-13 Nokia Networks Oy A method for preventing inconsistency between data in the master center and the backup center
US5963959A (en) * 1997-05-30 1999-10-05 Oracle Corporation Fast refresh of snapshots
JP3367385B2 (en) * 1997-06-27 2003-01-14 日本電気株式会社 Distributed transaction matching method and machine-readable recording medium recording program
US5924096A (en) * 1997-10-15 1999-07-13 Novell, Inc. Distributed database using indexed into tags to tracks events according to type, update cache, create virtual update log on demand
US6529904B1 (en) * 1998-05-28 2003-03-04 Oracle Corp. Deployment of snapshots with parameterized data description language strings
US7162689B2 (en) 1998-05-28 2007-01-09 Oracle International Corporation Schema evolution in replication
US6163777A (en) * 1998-06-23 2000-12-19 Microsoft Corporation System and method for reducing location conflicts in a database
US7016921B1 (en) * 1998-07-27 2006-03-21 Siemens Aktiengesellschaft Method, arrangement and set of a plurality of arrangements for remedying at least one inconsistency in a group of databases which comprises a database and at least one copy database of the database
US6377540B1 (en) * 1998-07-31 2002-04-23 Motorola, Inc. Method and apparatus for managing resource allocation conflicts in a communications systems
US6810405B1 (en) * 1998-08-18 2004-10-26 Starfish Software, Inc. System and methods for synchronizing data between multiple datasets
US6339769B1 (en) * 1998-09-14 2002-01-15 International Business Machines Corporation Query optimization by transparently altering properties of relational tables using materialized views
CA2255047A1 (en) * 1998-11-30 2000-05-30 Ibm Canada Limited-Ibm Canada Limitee Comparison of hierarchical structures and merging of differences
US6587856B1 (en) 1998-12-07 2003-07-01 Oracle International Corporation Method and system for representing and accessing object-oriented data in a relational database system
US7801913B2 (en) * 1998-12-07 2010-09-21 Oracle International Corporation System and method for querying data for implicit hierarchies
US6748374B1 (en) 1998-12-07 2004-06-08 Oracle International Corporation Method for generating a relational database query statement using one or more templates corresponding to search conditions in an expression tree
US6473776B2 (en) 1999-04-16 2002-10-29 International Business Machines Corporation Automatic prunning for log-based replication
US7526481B1 (en) * 1999-04-19 2009-04-28 Oracle International Corporation Web servers with queryable dynamic caches
GB2349718A (en) * 1999-05-07 2000-11-08 Quality Systems & Software Ltd Producing systems engineering models
US6430576B1 (en) * 1999-05-10 2002-08-06 Patrick Gates Distributing and synchronizing objects
WO2000070531A2 (en) * 1999-05-17 2000-11-23 The Foxboro Company Methods and apparatus for control configuration
US7089530B1 (en) 1999-05-17 2006-08-08 Invensys Systems, Inc. Process control configuration system with connection validation and configuration
US6484159B1 (en) * 1999-05-20 2002-11-19 At&T Corp. Method and system for incremental database maintenance
US6411967B1 (en) * 1999-06-18 2002-06-25 Reliable Network Solutions Distributed processing system with replicated management information base
US6374267B1 (en) * 1999-08-13 2002-04-16 Unisys Corporation Database backup system and method utilizing numerically identified files for incremental dumping
US6411969B1 (en) * 1999-08-13 2002-06-25 Unisys Corporation Enhanced system and method for management of system database utilities
EP1247221A4 (en) 1999-09-20 2005-01-19 Quintiles Transnat Corp System and method for analyzing de-identified health care data
US6578054B1 (en) * 1999-10-04 2003-06-10 Microsoft Corporation Method and system for supporting off-line mode of operation and synchronization using resource state information
US6944642B1 (en) * 1999-10-04 2005-09-13 Microsoft Corporation Systems and methods for detecting and resolving resource conflicts
US6973464B1 (en) 1999-11-15 2005-12-06 Novell, Inc. Intelligent replication method
US6615223B1 (en) * 2000-02-29 2003-09-02 Oracle International Corporation Method and system for data replication
US6820088B1 (en) 2000-04-10 2004-11-16 Research In Motion Limited System and method for synchronizing data records between multiple databases
US7162499B2 (en) * 2000-06-21 2007-01-09 Microsoft Corporation Linked value replication
US6529917B1 (en) 2000-08-14 2003-03-04 Divine Technology Ventures System and method of synchronizing replicated data
US6671686B2 (en) 2000-11-02 2003-12-30 Guy Pardon Decentralized, distributed internet data management
US6516317B1 (en) * 2000-12-21 2003-02-04 Oracle Corporation Method and apparatus for facilitating compartmentalized database user management
US7725423B1 (en) * 2001-02-08 2010-05-25 Teradata Us, Inc. Analyzing associations in the order of transactions
US6879564B2 (en) * 2001-02-28 2005-04-12 Microsoft Corp. Method for designating communication paths in a network
US7103586B2 (en) 2001-03-16 2006-09-05 Gravic, Inc. Collision avoidance in database replication systems
US7177866B2 (en) * 2001-03-16 2007-02-13 Gravic, Inc. Asynchronous coordinated commit replication and dual write with replication transmission and locking of target database on updates only
AU2002248638A1 (en) * 2001-03-16 2002-10-03 Orchid Systems, Inc. System for and method of synchronizing a standard-protocol database with a legacy data
CN1240244C (en) * 2001-05-11 2006-02-01 诺基亚公司 Data element information management in a network environment
US7945551B1 (en) * 2001-06-14 2011-05-17 Oracle International Corporation Redirection of misses in queryable caches
US20020198725A1 (en) * 2001-06-21 2002-12-26 International Business Machines Corporation Method and system for managing a relationship with a venture company
US6981250B1 (en) * 2001-07-05 2005-12-27 Microsoft Corporation System and methods for providing versioning of software components in a computer programming language
US6772178B2 (en) * 2001-07-27 2004-08-03 Sun Microsystems, Inc. Method and apparatus for managing remote data replication in a distributed computer system
US9659292B1 (en) * 2001-08-30 2017-05-23 EMC IP Holding Company LLC Storage-based replication of e-commerce transactions in real time
US7299230B2 (en) * 2001-08-30 2007-11-20 Siebel Systems, Inc. Method, apparatus and system for transforming, converting and processing messages between multiple systems
US7054890B2 (en) 2001-09-21 2006-05-30 Sun Microsystems, Inc. Method and apparatus for managing data imaging in a distributed computer system
US6996587B2 (en) * 2001-09-27 2006-02-07 Sun Microsystems, Inc. Method and apparatus for managing data volumes in a distributed computer system
US7146617B2 (en) 2001-09-29 2006-12-05 Siebel Systems, Inc. Method, apparatus, and system for implementing view caching in a framework to support web-based applications
US7203948B2 (en) * 2001-09-29 2007-04-10 Siebel Systems, Inc. Method, apparatus, and system for implementing caching of view custom options in a framework to support web-based applications
US6907451B1 (en) * 2001-09-29 2005-06-14 Siebel Systems, Inc. Method, apparatus, and system for immediate posting of changes in a client server environment
US7461119B2 (en) * 2001-09-29 2008-12-02 Siebel Systems, Inc. Method, apparatus, and system for managing status of requests in a client server environment
US7885996B2 (en) * 2001-09-29 2011-02-08 Siebel Systems, Inc. Method, apparatus, and system for implementing notifications in a framework to support web-based applications
US8359335B2 (en) 2001-09-29 2013-01-22 Siebel Systems, Inc. Computing system and method to implicitly commit unsaved data for a world wide web application
US7870492B2 (en) * 2001-10-02 2011-01-11 Siebel Systems, Inc. Method, apparatus, and system for managing commands in a client server environment
US6983465B2 (en) * 2001-10-11 2006-01-03 Sun Microsystems, Inc. Method and apparatus for managing data caching in a distributed computer system
US6938031B1 (en) * 2001-10-19 2005-08-30 Data Return Llc System and method for accessing information in a replicated database
US7444393B2 (en) * 2001-10-30 2008-10-28 Keicy K. Chung Read-only storage device having network interface, a system including the device, and a method of distributing files over a network
US7000235B2 (en) * 2001-10-30 2006-02-14 Sun Microsystems, Inc. Method and apparatus for managing data services in a distributed computer system
US6910075B2 (en) * 2001-11-14 2005-06-21 Emc Corporation Dynamic RDF groups
US7159015B2 (en) * 2001-11-16 2007-01-02 Sun Microsystems, Inc. Method and apparatus for managing configuration information in a distributed computer system
US7035922B2 (en) * 2001-11-27 2006-04-25 Microsoft Corporation Non-invasive latency monitoring in a store-and-forward replication system
US7185359B2 (en) * 2001-12-21 2007-02-27 Microsoft Corporation Authentication and authorization across autonomous network systems
EP3401794A1 (en) 2002-01-08 2018-11-14 Seven Networks, LLC Connection architecture for a mobile network
US7024429B2 (en) 2002-01-31 2006-04-04 Nextpage,Inc. Data replication based upon a non-destructive data model
US6909910B2 (en) 2002-02-01 2005-06-21 Microsoft Corporation Method and system for managing changes to a contact database
US6721871B2 (en) * 2002-04-02 2004-04-13 Nokia Corporation Method and apparatus for synchronizing data stores with respect to changes in folders
US9052944B2 (en) * 2002-07-16 2015-06-09 Oracle America, Inc. Obstruction-free data structures and mechanisms with separable and/or substitutable contention management mechanisms
US7139690B2 (en) * 2002-07-22 2006-11-21 Microsoft Corporation Object-level conflict detection in an object-relational database system
US10706417B2 (en) * 2002-08-02 2020-07-07 Ncr Corporation System and method for execution of customer-specific marketing, discounts, and promotions
US7487140B2 (en) 2002-12-17 2009-02-03 International Business Machines Corporation Method for executing a query having multiple distinct key columns
AU2003299837B2 (en) * 2002-12-23 2010-03-25 Antenna Dexterra, Inc. Mobile data and software update system and method
US7853563B2 (en) 2005-08-01 2010-12-14 Seven Networks, Inc. Universal data aggregation
US7917468B2 (en) 2005-08-01 2011-03-29 Seven Networks, Inc. Linking of personal information management data
US8468126B2 (en) 2005-08-01 2013-06-18 Seven Networks, Inc. Publishing data in an information community
US7366460B2 (en) * 2003-01-23 2008-04-29 Dexterra, Inc. System and method for mobile data update
US7865536B1 (en) 2003-02-14 2011-01-04 Google Inc. Garbage collecting systems and methods
US7822757B2 (en) * 2003-02-18 2010-10-26 Dun & Bradstreet, Inc. System and method for providing enhanced information
US20040249836A1 (en) * 2003-03-24 2004-12-09 John Reynders Synchronized data-centric and document-centric knowledge management system for drug discovery and development
WO2004092982A2 (en) * 2003-04-07 2004-10-28 Dexterra, Inc. System and method for context sensitive mobile data and software update
US7069351B2 (en) * 2003-06-02 2006-06-27 Chung Keicy K Computer storage device having network interface
US7660833B2 (en) 2003-07-10 2010-02-09 Microsoft Corporation Granular control over the authority of replicated information via fencing and unfencing
US7051623B2 (en) * 2003-07-28 2006-05-30 Easco Hand Tools, Inc. Pawl-less ratchet wrench
US7398285B2 (en) 2003-07-30 2008-07-08 International Business Machines Corporation Apparatus and system for asynchronous replication of a hierarchically-indexed data store
US8166101B2 (en) 2003-08-21 2012-04-24 Microsoft Corporation Systems and methods for the implementation of a synchronization schemas for units of information manageable by a hardware/software interface system
US20050055354A1 (en) * 2003-08-21 2005-03-10 Microsoft Corporation Systems and methods for representing units of information manageable by a hardware/software interface system but independent of physical representation
US7349913B2 (en) * 2003-08-21 2008-03-25 Microsoft Corporation Storage platform for organizing, searching, and sharing data
US7616333B2 (en) * 2003-08-21 2009-11-10 Microsoft Corporation Electronic ink processing and application programming interfaces
EP1656618A4 (en) * 2003-08-21 2007-10-17 Microsoft Corp Electronic ink processing
US7739316B2 (en) * 2003-08-21 2010-06-15 Microsoft Corporation Systems and methods for the implementation of base schema for organizing units of information manageable by a hardware/software interface system
US8131739B2 (en) 2003-08-21 2012-03-06 Microsoft Corporation Systems and methods for interfacing application programs with an item-based storage platform
ATE530992T1 (en) * 2003-08-21 2011-11-15 Microsoft Corp ELECTRONIC INK PROCESSING
US7483915B2 (en) * 2003-08-21 2009-01-27 Microsoft Corporation Systems and method for representing relationships between units of information manageable by a hardware/software interface system
US7529811B2 (en) * 2003-08-21 2009-05-05 Microsoft Corporation Systems and methods for the implementation of a core schema for providing a top-level structure for organizing units of information manageable by a hardware/software interface system
US7401104B2 (en) 2003-08-21 2008-07-15 Microsoft Corporation Systems and methods for synchronizing computer systems through an intermediary file system share or device
US8238696B2 (en) * 2003-08-21 2012-08-07 Microsoft Corporation Systems and methods for the implementation of a digital images schema for organizing units of information manageable by a hardware/software interface system
WO2005029391A1 (en) * 2003-08-21 2005-03-31 Microsoft Corporation Electronic ink processing
US7512638B2 (en) * 2003-08-21 2009-03-31 Microsoft Corporation Systems and methods for providing conflict handling for peer-to-peer synchronization of units of information manageable by a hardware/software interface system
US7483923B2 (en) * 2003-08-21 2009-01-27 Microsoft Corporation Systems and methods for providing relational and hierarchical synchronization services for units of information manageable by a hardware/software interface system
US7555497B2 (en) * 2003-08-21 2009-06-30 Microsoft Corporation Systems and methods for separating units of information manageable by a hardware/software interface system from their physical organization
US7590643B2 (en) 2003-08-21 2009-09-15 Microsoft Corporation Systems and methods for extensions and inheritance for units of information manageable by a hardware/software interface system
US7502812B2 (en) * 2003-08-21 2009-03-10 Microsoft Corporation Electronic ink processing
US7644376B2 (en) * 2003-10-23 2010-01-05 Microsoft Corporation Flexible architecture for notifying applications of state changes
US20050097144A1 (en) * 2003-11-04 2005-05-05 Taiwan Semiconductor Manufacturing Co. Performance tuning at CM loader program while replicating EQP list for IBM SiView
DE112004002233B4 (en) * 2003-11-19 2018-07-12 Nimcat Networks Inc. Time and data synchronization between network devices
US20050131965A1 (en) * 2003-12-11 2005-06-16 Lam Wai T. System and method for replicating data
US7478211B2 (en) * 2004-01-09 2009-01-13 International Business Machines Corporation Maintaining consistency for remote copy using virtualization
US20050154786A1 (en) * 2004-01-09 2005-07-14 International Business Machines Corporation Ordering updates in remote copying of data
US7624120B2 (en) * 2004-02-11 2009-11-24 Microsoft Corporation System and method for switching a data partition
US7801866B1 (en) 2004-02-20 2010-09-21 Microsoft Corporation Systems and methods for reading only durably committed data in a system that otherwise permits lazy commit of transactions
US7490083B2 (en) * 2004-02-27 2009-02-10 International Business Machines Corporation Parallel apply processing in data replication with preservation of transaction integrity and source ordering of dependent updates
US8688634B2 (en) * 2004-02-27 2014-04-01 International Business Machines Corporation Asynchronous peer-to-peer data replication
US20050218739A1 (en) * 2004-04-01 2005-10-06 Microsoft Corporation System and method for sharing objects between computers over a network
US7533134B2 (en) * 2004-04-01 2009-05-12 Microsoft Corporation Systems and methods for the propagation of conflict resolution to enforce item convergence (i.e., data convergence)
US7778962B2 (en) * 2004-04-30 2010-08-17 Microsoft Corporation Client store synchronization through intermediary store change packets
US7143120B2 (en) * 2004-05-03 2006-11-28 Microsoft Corporation Systems and methods for automated maintenance and repair of database and file systems
US7644239B2 (en) * 2004-05-03 2010-01-05 Microsoft Corporation Non-volatile memory cache performance improvement
EP1607883B1 (en) * 2004-06-15 2009-08-12 Sap Ag A data processing system and method for monitoring database replication
US7487471B2 (en) * 2004-07-23 2009-02-03 Sap Ag User interface for conflict resolution management
US7353227B2 (en) * 2004-07-23 2008-04-01 Sap Aktiengesellschaft Conflict resolution engine
US20070143768A1 (en) * 2004-07-28 2007-06-21 Satoshi Inami Conflict resolution apparatus
US20060031681A1 (en) * 2004-08-05 2006-02-09 Motorola, Inc. Method and system for controlling access to a wireless client device
FR2876816A1 (en) * 2004-10-15 2006-04-21 France Telecom METHOD AND SYSTEM FOR ASYMMETRIC DATA SYNCHRONIZATION
US8010082B2 (en) 2004-10-20 2011-08-30 Seven Networks, Inc. Flexible billing architecture
WO2006045102A2 (en) 2004-10-20 2006-04-27 Seven Networks, Inc. Method and apparatus for intercepting events in a communication system
US7490197B2 (en) 2004-10-21 2009-02-10 Microsoft Corporation Using external memory devices to improve system performance
US7706781B2 (en) 2004-11-22 2010-04-27 Seven Networks International Oy Data security in a mobile e-mail service
FI117152B (en) 2004-12-03 2006-06-30 Seven Networks Internat Oy E-mail service provisioning method for mobile terminal, involves using domain part and further parameters to generate new parameter set in list of setting parameter sets, if provisioning of e-mail service is successful
EP1828932A4 (en) * 2004-12-10 2008-03-05 Seven Networks Internat Oy Database synchronization
US9020887B2 (en) 2004-12-21 2015-04-28 Proofpoint, Inc. Managing the status of documents in a distributed storage system
FI120165B (en) * 2004-12-29 2009-07-15 Seven Networks Internat Oy Synchronization of a database through a mobile network
US7509354B2 (en) * 2005-01-07 2009-03-24 International Business Machines Corporation System, method, and computer program product for multi-master replication conflict resolution
US8682844B2 (en) * 2005-01-28 2014-03-25 Sap Ag Method and apparatus for collision resolution in an asynchronous database system
US7712078B1 (en) * 2005-02-02 2010-05-04 Teradata Us, Inc. Techniques for data store population
US9286346B2 (en) * 2005-02-18 2016-03-15 International Business Machines Corporation Replication-only triggers
US8037056B2 (en) * 2005-02-18 2011-10-11 International Business Machines Corporation Online repair of a replicated table
US7376675B2 (en) * 2005-02-18 2008-05-20 International Business Machines Corporation Simulating multi-user activity while maintaining original linear request order for asynchronous transactional events
US8214353B2 (en) * 2005-02-18 2012-07-03 International Business Machines Corporation Support for schema evolution in a multi-node peer-to-peer replication environment
US7805422B2 (en) 2005-02-28 2010-09-28 Microsoft Corporation Change notification query multiplexing
US7877703B1 (en) 2005-03-14 2011-01-25 Seven Networks, Inc. Intelligent rendering of information in a limited display environment
US8214754B2 (en) 2005-04-15 2012-07-03 Microsoft Corporation Registration of applications and complimentary features for interactive user interfaces
US7796742B1 (en) 2005-04-21 2010-09-14 Seven Networks, Inc. Systems and methods for simplified provisioning
US8438633B1 (en) 2005-04-21 2013-05-07 Seven Networks, Inc. Flexible real-time inbox access
WO2006136660A1 (en) 2005-06-21 2006-12-28 Seven Networks International Oy Maintaining an ip connection in a mobile network
US8069166B2 (en) 2005-08-01 2011-11-29 Seven Networks, Inc. Managing user-to-user contact with inferred presence information
US8533169B1 (en) * 2005-09-21 2013-09-10 Infoblox Inc. Transactional replication
US8290910B2 (en) * 2005-09-21 2012-10-16 Infoblox Inc. Semantic replication
US8250030B2 (en) 2005-09-21 2012-08-21 Infoblox Inc. Provisional authority in a distributed database
US20070143368A1 (en) * 2005-12-15 2007-06-21 Lundsgaard Soren K Conflict resolution in highly available network element
US8914557B2 (en) 2005-12-16 2014-12-16 Microsoft Corporation Optimizing write and wear performance for a memory
US7769395B2 (en) 2006-06-20 2010-08-03 Seven Networks, Inc. Location-based operations and messaging
US8359297B2 (en) 2006-06-29 2013-01-22 International Business Machines Corporation Multiple source data management using a conflict rule
US8688749B1 (en) 2011-03-31 2014-04-01 Palantir Technologies, Inc. Cross-ontology multi-master replication
US8515912B2 (en) 2010-07-15 2013-08-20 Palantir Technologies, Inc. Sharing and deconflicting data changes in a multimaster database system
US9355273B2 (en) 2006-12-18 2016-05-31 Bank Of America, N.A., As Collateral Agent System and method for the protection and de-identification of health care data
US8693494B2 (en) 2007-06-01 2014-04-08 Seven Networks, Inc. Polling
US8805425B2 (en) 2007-06-01 2014-08-12 Seven Networks, Inc. Integrated messaging
US8566296B2 (en) 2007-09-14 2013-10-22 Oracle International Corporation Support for compensation aware data types in relational database systems
US8769185B2 (en) 2007-10-23 2014-07-01 Keicy Chung Computer storage device having separate read-only space and read-write space, removable media component, system management interface, and network interface
US8364181B2 (en) 2007-12-10 2013-01-29 Seven Networks, Inc. Electronic-mail filtering for mobile devices
US8631203B2 (en) 2007-12-10 2014-01-14 Microsoft Corporation Management of external memory functioning as virtual cache
US8793305B2 (en) 2007-12-13 2014-07-29 Seven Networks, Inc. Content delivery to a mobile device from a content service
US9002828B2 (en) 2007-12-13 2015-04-07 Seven Networks, Inc. Predictive content delivery
US8107921B2 (en) 2008-01-11 2012-01-31 Seven Networks, Inc. Mobile virtual network operator
US8862657B2 (en) 2008-01-25 2014-10-14 Seven Networks, Inc. Policy based content service
US20090193338A1 (en) 2008-01-28 2009-07-30 Trevor Fiatal Reducing network and battery consumption during content delivery and playback
US8787947B2 (en) 2008-06-18 2014-07-22 Seven Networks, Inc. Application discovery on mobile devices
US8984301B2 (en) * 2008-06-19 2015-03-17 International Business Machines Corporation Efficient identification of entire row uniqueness in relational databases
US8078158B2 (en) 2008-06-26 2011-12-13 Seven Networks, Inc. Provisioning applications for a mobile device
US8032707B2 (en) 2008-09-15 2011-10-04 Microsoft Corporation Managing cache data and metadata
US9032151B2 (en) 2008-09-15 2015-05-12 Microsoft Technology Licensing, Llc Method and system for ensuring reliability of cache data and metadata subsequent to a reboot
US7953774B2 (en) 2008-09-19 2011-05-31 Microsoft Corporation Aggregation of write traffic to a data store
US8909759B2 (en) 2008-10-10 2014-12-09 Seven Networks, Inc. Bandwidth measurement
US8364636B2 (en) * 2009-09-14 2013-01-29 International Business Machines Corporation Real time data replication
US8683498B2 (en) * 2009-12-16 2014-03-25 Ebay Inc. Systems and methods for facilitating call request aggregation over a network
WO2011126889A2 (en) 2010-03-30 2011-10-13 Seven Networks, Inc. 3d mobile user interface with configurable workspace management
US8838783B2 (en) 2010-07-26 2014-09-16 Seven Networks, Inc. Distributed caching for resource and mobile network traffic management
GB2500333B (en) 2010-07-26 2014-10-08 Seven Networks Inc Mobile application traffic optimization
US9077630B2 (en) 2010-07-26 2015-07-07 Seven Networks, Inc. Distributed implementation of dynamic wireless traffic policy
US9043433B2 (en) 2010-07-26 2015-05-26 Seven Networks, Inc. Mobile network traffic coordination across multiple applications
WO2012060995A2 (en) 2010-11-01 2012-05-10 Michael Luna Distributed caching in a wireless network of content delivered for a mobile application over a long-held request
US8326985B2 (en) 2010-11-01 2012-12-04 Seven Networks, Inc. Distributed management of keep-alive message signaling for mobile network resource conservation and optimization
US8166164B1 (en) 2010-11-01 2012-04-24 Seven Networks, Inc. Application and network-based long poll request detection and cacheability assessment therefor
US8484314B2 (en) 2010-11-01 2013-07-09 Seven Networks, Inc. Distributed caching in a wireless network of content delivered for a mobile application over a long-held request
US8843153B2 (en) 2010-11-01 2014-09-23 Seven Networks, Inc. Mobile traffic categorization and policy for network use optimization while preserving user experience
WO2012061437A1 (en) 2010-11-01 2012-05-10 Michael Luna Cache defeat detection and caching of content addressed by identifiers intended to defeat cache
US9330196B2 (en) 2010-11-01 2016-05-03 Seven Networks, Llc Wireless traffic management system cache optimization using http headers
GB2499534B (en) 2010-11-01 2018-09-19 Seven Networks Llc Caching adapted for mobile application behavior and network conditions
US9060032B2 (en) 2010-11-01 2015-06-16 Seven Networks, Inc. Selective data compression by a distributed traffic management system to reduce mobile data traffic and signaling traffic
GB2500327B (en) 2010-11-22 2019-11-06 Seven Networks Llc Optimization of resource polling intervals to satisfy mobile device requests
GB2495463B (en) 2010-11-22 2013-10-09 Seven Networks Inc Aligning data transfer to optimize connections established for transmission over a wireless network
US8341134B2 (en) 2010-12-10 2012-12-25 International Business Machines Corporation Asynchronous deletion of a range of messages processed by a parallel database replication apply process
GB2501416B (en) 2011-01-07 2018-03-21 Seven Networks Llc System and method for reduction of mobile network traffic used for domain name system (DNS) queries
WO2012145533A2 (en) 2011-04-19 2012-10-26 Seven Networks, Inc. Shared resource and virtual resource management in a networked environment
WO2012149434A2 (en) 2011-04-27 2012-11-01 Seven Networks, Inc. Detecting and preserving state for satisfying application requests in a distributed proxy and cache system
GB2504037B (en) 2011-04-27 2014-12-24 Seven Networks Inc Mobile device which offloads requests made by a mobile application to a remote entity for conservation of mobile device and network resources
WO2013015994A1 (en) 2011-07-27 2013-01-31 Seven Networks, Inc. Monitoring mobile application activities for malicious traffic on a mobile device
WO2013019894A1 (en) * 2011-08-01 2013-02-07 Tagged, Inc. Systems and methods for asynchronous distributed database management
US9547705B2 (en) 2011-09-23 2017-01-17 Hybrid Logic Ltd System for live-migration and automated recovery of applications in a distributed system
US10311027B2 (en) 2011-09-23 2019-06-04 Open Invention Network, Llc System for live-migration and automated recovery of applications in a distributed system
GB2495079A (en) 2011-09-23 2013-04-03 Hybrid Logic Ltd Live migration of applications and file systems in a distributed system
US10331801B2 (en) 2011-09-23 2019-06-25 Open Invention Network, Llc System for live-migration and automated recovery of applications in a distributed system
US9483542B2 (en) 2011-09-23 2016-11-01 Hybrid Logic Ltd System for live-migration and automated recovery of applications in a distributed system
US9501543B2 (en) * 2011-09-23 2016-11-22 Hybrid Logic Ltd System for live-migration and automated recovery of applications in a distributed system
US9477739B2 (en) 2011-09-23 2016-10-25 Hybrid Logic Ltd System for live-migration and automated recovery of applications in a distributed system
US8868753B2 (en) 2011-12-06 2014-10-21 Seven Networks, Inc. System of redundantly clustered machines to provide failover mechanisms for mobile traffic management and network resource conservation
US8934414B2 (en) 2011-12-06 2015-01-13 Seven Networks, Inc. Cellular or WiFi mobile traffic optimization based on public or private network destination
US9009250B2 (en) 2011-12-07 2015-04-14 Seven Networks, Inc. Flexible and dynamic integration schemas of a traffic management system with various network operators for network traffic alleviation
WO2013086447A1 (en) 2011-12-07 2013-06-13 Seven Networks, Inc. Radio-awareness of mobile device for sending server-side control signals using a wireless network optimized transport protocol
US9832095B2 (en) 2011-12-14 2017-11-28 Seven Networks, Llc Operation modes for mobile traffic optimization and concurrent management of optimized and non-optimized traffic
EP2792188B1 (en) 2011-12-14 2019-03-20 Seven Networks, LLC Mobile network reporting and usage analytics system and method using aggregation of data in a distributed traffic optimization system
US8861354B2 (en) 2011-12-14 2014-10-14 Seven Networks, Inc. Hierarchies and categories for management and deployment of policies for distributed wireless traffic optimization
US8768927B2 (en) * 2011-12-22 2014-07-01 Sap Ag Hybrid database table stored as both row and column store
US8818949B2 (en) * 2011-12-30 2014-08-26 Bmc Software, Inc. Systems and methods for migrating database data
GB2499306B (en) 2012-01-05 2014-10-22 Seven Networks Inc Managing user interaction with an application on a mobile device
US8782004B2 (en) * 2012-01-23 2014-07-15 Palantir Technologies, Inc. Cross-ACL multi-master replication
WO2013116856A1 (en) 2012-02-02 2013-08-08 Seven Networks, Inc. Dynamic categorization of applications for network access in a mobile network
US9326189B2 (en) 2012-02-03 2016-04-26 Seven Networks, Llc User as an end point for profiling and optimizing the delivery of content and data in a wireless network
US8812695B2 (en) 2012-04-09 2014-08-19 Seven Networks, Inc. Method and system for management of a virtual network connection without heartbeat messages
US10263899B2 (en) 2012-04-10 2019-04-16 Seven Networks, Llc Enhanced customer service for mobile carriers using real-time and historical mobile application and traffic or optimization data associated with mobile devices in a mobile network
US8543540B1 (en) * 2012-05-09 2013-09-24 Bertec Corporation System and method for the merging of databases
US8700569B1 (en) 2012-05-09 2014-04-15 Bertec Corporation System and method for the merging of databases
US9043278B1 (en) 2012-05-09 2015-05-26 Bertec Corporation System and method for the merging of databases
WO2014011216A1 (en) 2012-07-13 2014-01-16 Seven Networks, Inc. Dynamic bandwidth adjustment for browsing or streaming activity in a wireless network based on prediction of user behavior when interacting with mobile applications
US9081975B2 (en) 2012-10-22 2015-07-14 Palantir Technologies, Inc. Sharing information between nexuses that use different classification schemes for information access control
US9161258B2 (en) 2012-10-24 2015-10-13 Seven Networks, Llc Optimized and selective management of policy deployment to mobile clients in a congested network to prevent further aggravation of network congestion
US9501761B2 (en) 2012-11-05 2016-11-22 Palantir Technologies, Inc. System and method for sharing investigation results
US9307493B2 (en) 2012-12-20 2016-04-05 Seven Networks, Llc Systems and methods for application management of mobile device radio state promotion and demotion
US9824132B2 (en) * 2013-01-08 2017-11-21 Facebook, Inc. Data recovery in multi-leader distributed systems
US9241314B2 (en) 2013-01-23 2016-01-19 Seven Networks, Llc Mobile device with application or context aware fast dormancy
US8874761B2 (en) 2013-01-25 2014-10-28 Seven Networks, Inc. Signaling optimization in a wireless network for traffic utilizing proprietary and non-proprietary protocols
US8750123B1 (en) 2013-03-11 2014-06-10 Seven Networks, Inc. Mobile device equipped with mobile network congestion recognition to make intelligent decisions regarding connecting to an operator network
US9679037B2 (en) * 2013-06-07 2017-06-13 Wipro Limited System and method for implementing database replication configurations using replication modeling and transformation
US9065765B2 (en) 2013-07-22 2015-06-23 Seven Networks, Inc. Proxy server associated with a mobile carrier for enhancing mobile traffic management in a mobile network
US9183526B2 (en) 2013-09-11 2015-11-10 Oracle International Corporation Metadata-driven audit reporting system that applies data security to audit data
US9690838B2 (en) * 2013-10-31 2017-06-27 Microsoft Technology Licensing, Llc Master data management
US9569070B1 (en) * 2013-11-11 2017-02-14 Palantir Technologies, Inc. Assisting in deconflicting concurrency conflicts
US9569481B1 (en) 2013-12-10 2017-02-14 Google Inc. Efficient locking of large data collections
WO2015097991A1 (en) * 2013-12-24 2015-07-02 日本電気株式会社 Transaction distribution processing device, method, system and memory medium
US9727625B2 (en) 2014-01-16 2017-08-08 International Business Machines Corporation Parallel transaction messages for database replication
US9965359B2 (en) 2014-11-25 2018-05-08 Sap Se Log forwarding to avoid deadlocks during parallel log replay in asynchronous table replication
US10929431B2 (en) * 2015-08-28 2021-02-23 Hewlett Packard Enterprise Development Lp Collision handling during an asynchronous replication
US10620626B2 (en) 2015-11-24 2020-04-14 Nova Dynamics, Llc Conflict resolution via contingency plan execution or human interaction
US11624631B2 (en) 2015-11-24 2023-04-11 Daxbot Inc. Autonomous robots and methods for determining, mapping, and traversing routes for autonomous robots
US10650300B2 (en) 2015-11-24 2020-05-12 Nova Dynamics, Llc Component-based decision-making with centralized officiating and the modification of those decisions through success measurements
US10578443B2 (en) 2015-11-24 2020-03-03 Nova Dynamics, Llc Method for re-mapping safe and traversable routes
US10578447B2 (en) 2015-11-24 2020-03-03 Nova Dynamics, Llc Method for identifying safe and traversable paths
CN107395384A (en) 2016-05-17 2017-11-24 阿里巴巴集团控股有限公司 Versions of data comparison method and device between across time zone website
US10585854B2 (en) * 2016-06-24 2020-03-10 Box, Inc. Establishing and enforcing selective object deletion operations on cloud-based shared content
US20180173778A1 (en) * 2016-12-16 2018-06-21 Linkedin Corporation Database uniqueness constraints
US11210212B2 (en) 2017-08-21 2021-12-28 Western Digital Technologies, Inc. Conflict resolution and garbage collection in distributed databases
US11055266B2 (en) 2017-08-21 2021-07-06 Western Digital Technologies, Inc. Efficient key data store entry traversal and result generation
US10824612B2 (en) 2017-08-21 2020-11-03 Western Digital Technologies, Inc. Key ticketing system with lock-free concurrency and versioning
US11210211B2 (en) 2017-08-21 2021-12-28 Western Digital Technologies, Inc. Key data store garbage collection and multipart object management
US10942823B2 (en) 2018-01-29 2021-03-09 Guy Pardon Transaction processing system, recovery subsystem and method for operating a recovery subsystem
US20190239037A1 (en) * 2018-02-01 2019-08-01 Blackberry Limited System and method for managing items in a list shared by a group of mobile devices
US11204940B2 (en) * 2018-11-16 2021-12-21 International Business Machines Corporation Data replication conflict processing after structural changes to a database

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4432057A (en) * 1981-11-27 1984-02-14 International Business Machines Corporation Method for the dynamic replication of data under distributed system control to control utilization of resources in a multiprocessing, distributed data base system
US5212788A (en) * 1990-05-22 1993-05-18 Digital Equipment Corporation System and method for consistent timestamping in distributed computer databases
US5241675A (en) * 1992-04-09 1993-08-31 Bell Communications Research, Inc. Method for enforcing the serialization of global multidatabase transactions through committing only on consistent subtransaction serialization by the local database managers
US5386559A (en) * 1992-07-16 1995-01-31 International Business Machines Corporation Variant domains and variant maps in a versioned database management system
US5434994A (en) * 1994-05-23 1995-07-18 International Business Machines Corporation System and method for maintaining replicated data coherency in a data processing system

Family Cites Families (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS56168263A (en) * 1980-05-30 1981-12-24 Hitachi Ltd Program making device
US4558413A (en) * 1983-11-21 1985-12-10 Xerox Corporation Software version management system
JPH0640302B2 (en) * 1984-01-30 1994-05-25 株式会社日立製作所 Schematic / source program automatic generation method
US4888690A (en) * 1985-01-11 1989-12-19 Wang Laboratories, Inc. Interactive error handling means in database management
US4734854A (en) * 1985-10-08 1988-03-29 American Telephone And Telegraph Company System for generating software source code components
US5202996A (en) * 1985-10-11 1993-04-13 Hitachi, Ltd. Software structuring system and method by data table translation
JPS62128332A (en) * 1985-11-30 1987-06-10 Toshiba Corp Data processor
US4833604A (en) * 1986-01-13 1989-05-23 International Business Machines Corporation Method for the relocation of linked control blocks
US4866611A (en) * 1987-01-29 1989-09-12 International Business Machines Corporation Method for automatically reconciling entries on two copies of independently maintained electronic calendars
US4939689A (en) * 1987-04-09 1990-07-03 Crowninshield Software, Inc. Outline-driven database editing and retrieval system
US4809170A (en) * 1987-04-22 1989-02-28 Apollo Computer, Inc. Computer device for aiding in the development of software system
US4930071A (en) * 1987-06-19 1990-05-29 Intellicorp, Inc. Method for integrating a knowledge-based system with an arbitrary database system
JPH01162979A (en) * 1987-12-19 1989-06-27 Ricoh Co Ltd Extension mechanism for data base control system
US4866638A (en) * 1988-03-04 1989-09-12 Eastman Kodak Company Process for producing human-computer interface prototypes
US5084813A (en) * 1988-04-20 1992-01-28 Kabushiki Kaisha Toshiba Rule based system for synthesizing a program suited for a target system in response to an input target system specification
US5016204A (en) * 1989-02-28 1991-05-14 Digital Equipment Corporation Expert system for performing diagnostic and redesign operations incorporating multiple levels of simulation detail
ES2106771T3 (en) * 1989-11-30 1997-11-16 Seer Technologies Inc COMPUTER ASSISTED SOFTWARE ENGINEERING UTILITY.
US5327555A (en) * 1991-02-14 1994-07-05 Hewlett-Packard Company Method for reconciling entries in a plurality of schedules
US5261094A (en) * 1991-04-08 1993-11-09 International Business Machines Corporation Asynchronous replication of data changes by distributed update requests
US5301316A (en) * 1991-10-30 1994-04-05 Sun Microsystems, Inc. System for determination of the equivalence of two objects without compromising then done by a third object nominated by one and accetped by the other
AU7684094A (en) * 1993-09-24 1995-04-10 Oracle Corporation Method and apparatus for data replication
EP0678812A1 (en) * 1994-04-20 1995-10-25 Microsoft Corporation Replication verification
US5625818A (en) * 1994-09-30 1997-04-29 Apple Computer, Inc. System for managing local database updates published to different online information services in different formats from a central platform
US5790848A (en) * 1995-02-03 1998-08-04 Dex Information Systems, Inc. Method and apparatus for data access and update in a shared file environment
AU6500496A (en) * 1995-07-20 1997-02-18 Novell, Inc. Transaction clash management in a disconnectable computer and network
US5870759A (en) * 1996-10-09 1999-02-09 Oracle Corporation System for synchronizing data between computers using a before-image of data
US5870765A (en) * 1996-10-09 1999-02-09 Oracle Corporation Database synchronizer
US5884325A (en) * 1996-10-09 1999-03-16 Oracle Corporation System for synchronizing shared data between computers
US5926816A (en) * 1996-10-09 1999-07-20 Oracle Corporation Database Synchronizer

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4432057A (en) * 1981-11-27 1984-02-14 International Business Machines Corporation Method for the dynamic replication of data under distributed system control to control utilization of resources in a multiprocessing, distributed data base system
US5212788A (en) * 1990-05-22 1993-05-18 Digital Equipment Corporation System and method for consistent timestamping in distributed computer databases
US5241675A (en) * 1992-04-09 1993-08-31 Bell Communications Research, Inc. Method for enforcing the serialization of global multidatabase transactions through committing only on consistent subtransaction serialization by the local database managers
US5386559A (en) * 1992-07-16 1995-01-31 International Business Machines Corporation Variant domains and variant maps in a versioned database management system
US5434994A (en) * 1994-05-23 1995-07-18 International Business Machines Corporation System and method for maintaining replicated data coherency in a data processing system

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
PROC. OF THE TWENTY-SECOND INTL. CONF. ON SYSTEM SCIENCES, Vol. III, 3-6 January 1989, CAREY et al., "ICRSS: Interactive Resolution Support System for Inter-Group Situations", p. 512-16. *
REAL-TIME SYSTEMS SYMPOSIUM, 5-7 December 1989, IEEE COMPUTER SOCIETY PRESS, HUANG et al., "Experimental Evaluation of Real-Time Transaction Processing", p. 144-53. *
THE FIRST INTL. CONF. ON AI ON WALL STREET, 9-11 October 1991, FISHMAN et al., "A New Perspective on Conflict Resolution in Market Forecasting", p. 97-102. *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6138143A (en) * 1999-01-28 2000-10-24 Genrad, Inc. Method and apparatus for asynchronous transaction processing
US7149759B2 (en) 2002-03-25 2006-12-12 International Business Machines Corporation Method and system for detecting conflicts in replicated data in a database network
EP3155527A4 (en) * 2014-07-02 2018-04-18 Pure Storage, Inc. Redundant, fault-tolerant, distributed remote procedure call cache in a storage system
US11176118B2 (en) 2016-02-09 2021-11-16 International Business Machines Corporation Performing conflict analysis of replicated changes among nodes in a network
US10540340B2 (en) 2016-02-09 2020-01-21 International Business Machines Corporation Performing conflict analysis of replicated changes among nodes in a network
US10585878B2 (en) 2016-02-09 2020-03-10 International Business Machines Corporation Performing conflict analysis of replicated changes among nodes in a network
US11126610B1 (en) * 2017-11-22 2021-09-21 Amazon Technologies, Inc. Conflict resolution in a data proxy
US11030185B2 (en) 2018-05-07 2021-06-08 Microsoft Technology Licensing, Llc Schema-agnostic indexing of distributed databases
US10970270B2 (en) 2018-05-07 2021-04-06 Microsoft Technology Licensing, Llc Unified data organization for multi-model distributed databases
US10970269B2 (en) 2018-05-07 2021-04-06 Microsoft Technology Licensing, Llc Intermediate consistency levels for database configuration
US10885018B2 (en) 2018-05-07 2021-01-05 Microsoft Technology Licensing, Llc Containerization for elastic and scalable databases
US10817506B2 (en) 2018-05-07 2020-10-27 Microsoft Technology Licensing, Llc Data service provisioning, metering, and load-balancing via service units
WO2019217481A1 (en) * 2018-05-07 2019-11-14 Microsoft Technology Licensing, Llc Conflict resolution for multi-master distributed databases
US11321303B2 (en) 2018-05-07 2022-05-03 Microsoft Technology Licensing, Llc Conflict resolution for multi-master distributed databases
US11379461B2 (en) 2018-05-07 2022-07-05 Microsoft Technology Licensing, Llc Multi-master architectures for distributed databases
US11397721B2 (en) 2018-05-07 2022-07-26 Microsoft Technology Licensing, Llc Merging conflict resolution for multi-master distributed databases
WO2022272084A1 (en) * 2021-06-25 2022-12-29 Bristol-Myers Squibb Company Indicating differences in and reconciling data stored in disparate data storage devices

Also Published As

Publication number Publication date
GB9819943D0 (en) 1998-11-04
CA2249422A1 (en) 1997-09-25
CA2249422C (en) 2005-05-10
US5806074A (en) 1998-09-08
GB2326495A (en) 1998-12-23
US6058401A (en) 2000-05-02
AU2335397A (en) 1997-10-10
GB2326495B (en) 2000-12-13

Similar Documents

Publication Publication Date Title
US5806074A (en) Configurable conflict resolution in a computer implemented distributed database
WO1997035270A9 (en) Configurable conflict resolution in a computer implemented distributed database
US5806075A (en) Method and apparatus for peer-to-peer data replication
US7076778B2 (en) Method and apparatus for upgrading a software application in the presence of user modifications
US6189011B1 (en) Method of maintaining a network of partially replicated database system
US6604117B2 (en) Method of maintaining a network of partially replicated database system
US5434994A (en) System and method for maintaining replicated data coherency in a data processing system
Corbett et al. Spanner: Google’s globally distributed database
US5799306A (en) Method and apparatus for facilitating data replication using object groups
Davcev et al. Consistency and Recovery Control for Replicated Files.
US7672966B2 (en) Adding extrinsic data columns to an existing database schema using a temporary column pool
US9165047B2 (en) Data synchronization system and method
US6266669B1 (en) Partially replicated distributed database with multiple levels of remote clients
US6216135B1 (en) Method of determining visibility to a remote database client of a plurality of database transactions having variable visibility strengths
US6446089B1 (en) Method of using a cache to determine the visibility to a remote database client of a plurality of database transactions
US6178425B1 (en) Method of determining the visibility to a remote database client of a plurality of database transactions using simplified visibility rules
US6047289A (en) Method and apparatus for directed data propagation
US6343299B1 (en) Method and apparatus for random update synchronization among multiple computing devices
US8386646B2 (en) Simplified application object data synchronization for optimized data storage
US6889229B1 (en) Techniques for peer-to-peer replication of objects in a relational database
Pacitti et al. Replica consistency in lazy master replicated databases
GB2366421A (en) Ordering relational database operations according to referential integrity constraints
US20090300037A1 (en) Enhanced database structure configuration
Ekenstam et al. The Bengal database replication system
US20030088615A1 (en) Update resolution procedure for a directory server

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AL AM AT AT AU AZ BA BB BG BR BY CA CH CN CU CZ CZ DE DE DK DK EE EE ES FI FI GB GE GH HU IL IS JP KE KG KP KR KZ LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SK TJ TM TR TT UA UG US UZ VN YU

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH KE LS MW SD SZ UG AM AZ BY KG KZ MD RU TJ TM AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
COP Corrected version of pamphlet

Free format text: PAGES 1-43, DESCRIPTION, REPLACED BY NEW PAGES 1-28; PAGES 44-48, CLAIMS, REPLACED BY NEW PAGES 29-32; PAGES 1/20-20/20, DRAWINGS, REPLACED BY NEW PAGES BEARING THE SAME NUMBER

ENP Entry into the national phase

Ref country code: GB

Ref document number: 9819943

Kind code of ref document: A

Format of ref document f/p: F

ENP Entry into the national phase

Ref document number: 2249422

Country of ref document: CA

Ref country code: CA

Ref document number: 2249422

Kind code of ref document: A

Format of ref document f/p: F

NENP Non-entry into the national phase

Ref country code: JP

Ref document number: 97533661

Format of ref document f/p: F

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase