Abstract:
Techniques for restoring point-in-time and transaction consistency across consistency groups between a first and a second independent database management system (DBMS) for a disaster recovery. Several consistency groups (CGs) are defined for replication. For each CG in the first DBMS data changes are transmitted to a second DBMS. A timestamp representing a most recently received commit log record or a heartbeat during periods of inactivity for a CG is stored in a database table at regular intervals. At regular intervals, the timestamp is compared with timestamps for other CGs to identify a common time at which data to be applied to the CGs in the second DBMS have been received into a recoverable data store. The received data is applied to the CGs in the second DBMS up to the common time.
Abstract:
A method for configuring an asynchronous Disaster Recovery (DR) process over a storage system comprising a plurality of storage sites, the method comprising: receiving data indicative of the storage sites; calculating information dispersal parameters that enable dispersing slices of DR enabling data based on original data written to a primary storage site out of the storage sites between at least two DR storage sites out of the storage sites, wherein at least one of the DR storage sites is a remote storage site, wherein each of the slices of DR enabling data is smaller than the original data and wherein the DR enabling data is larger than the original data; and configuring the asynchronous DR process in accordance with the calculated information dispersal parameters.
Abstract:
A system, method, and computer readable medium for reliable messaging between two or more servers. The computer readable medium includes computer-executable instructions for execution by a processing system. Primary applications runs on primary hosts and one or more replicated instances of each primary application run on one or more backup hosts. The reliable messaging ensures consistent ordered delivery of messages in the event that messages are lost; arrive out of order, or in duplicate. The messaging layer operates over TCP or UDP with our without multi-cast and broad-cast and requires no modification to applications, operating system or libraries.
Abstract:
Several different embodiments of a massively scalable object storage system are described. The object storage system is particularly useful for storage in a cloud computing installation whereby shared servers provide resources, software, and data to computers and other devices on demand. In several embodiments, the object storage system includes a ring implementation used to associate object storage commands with particular physical servers such that certain guarantees of consistency, availability, and performance can be met. In other embodiments, the object storage system includes a synchronization protocol used to order operations across a distributed system. In a third set of embodiments, the object storage system includes a metadata management system. In a fourth set of embodiments, the object storage system uses a structured information synchronization system. Features from each set of embodiments can be used to improve the performance and scalability of a cloud computing object storage system.
Abstract:
Replication is performed with consideration of data integrity between each subsystem when replication of a computer system that uses data by processing data and transferring the data to a next subsystem is performed. A management device that manages a computer system including a second subsystem which performs a predetermined process for data processed by a first subsystem and generates data which is a target of data processing by a third subsystem obtains process history information in which information indicating an input source and an output destination subsystem of data that is processed by each subsystem is included and trigger information in which information indicating a trigger for data input and output of the input source and the output destination subsystems is included, thereafter detects a dependence relationship of data input and output between each subsystem from the process history information, calculates a replication trigger for subsystems subsequent to a next subsystem for each of the subsystems subsequent to the next subsystem that is next to a subsystem of which an input source is not present with reference to the trigger information, and generates, in response to the replication trigger, a replication of each subsystem in another different computer system.
Abstract:
A method for error tracking a log subsystem of a file system is provided. The method includes: when a data block of the log subsystem is recovered to an original position in the file system, calculating a verification code of the data block to obtain a second verification code; determining whether a verification result between the second verification code and a first verification code of the data block stored in a spare space in a submit block of the log subsystem in a disk is consistent; and when the verification result is inconsistent, processing the data block corresponding to the inconsistent verification result. With the above method, given that system performance is least affected, an error and a position of the error of the log subsystem of the file system can be more accurately detected to enhance the reliability of the log subsystem.
Abstract:
Embodiments are directed to simultaneously backing up data from multiple locations to a single location and to restoring data from a single location to multiple locations simultaneously. In an embodiment, a computer system receives a backup command from a user which indicates that various specified portions of data are to be backed up. The data portions are stored in multiple different locations in a parallel database. The computer system generates multiple backup commands specific for each location in which the data is stored. The backup commands initiate a parallel backup from the various locations to a single location in a data store. The computer system also executes the generated backup commands in parallel, so that the data from each of the different locations is backed up in the data store.
Abstract:
A method for confirming replicated data at a data site, including utilizing a hash function, computing a first hash value based on first data at a first data site and utilizing the same hash function, computing a second hash value based on second data at a second data site, wherein the first data had previously been replicated from the first data site to the second data site as the second data. The method also includes comparing the first and second hash values to determine whether the second data is a valid replication of the first data. In additional embodiments, the first data may be modified based on seed data prior to computing the first hash value and the second data may be modified based on the same seed data prior to computing the second hash value. The process can be repeated to increase reliability of the results.
Abstract:
Embodiments of the present invention provide a data processing device and a data processing method. In the data processing device and the data processing method provided by the embodiments of the present invention, first data in a memory is written into a first non-volatile storage unit in a log file form, and a log file of the first data written into the first non-volatile storage unit is written into a second non-volatile storage unit. Because a data write speed of the first non-volatile storage unit is higher than a data write speed of the second non-volatile storage unit, fast backup of the data in the memory can be achieved, and when the data in the memory is lost in an abnormal situation, security of the data in the memory can be ensured.
Abstract:
Performing data management operations on replicated data in a computer network. Log entries are generated for data management operations of an application executing on a source system. Consistency point entries are used to indicate a time of a known good, or recoverable, state of the application. A destination system is configured to process a copy of the log and consistency point entries to replicate data in a replication volume, the replicated data being a copy of the application data on the source system. When the replicated data represents a known good state of the application, as determined by the consistency point entries, the destination system(s) may perform a storage operation (e.g., snapshot, backup) to copy the replicated data and to logically associate the copied data with a time information (e.g., time stamp) indicative of the source system time when the application was in the known good state.