G06F11/16

ENABLING DATA INTEGRITY CHECKING AND FASTER APPLICATION RECOVERY IN SYNCHRONOUS REPLICATED DATASETS
20220171544 · 2022-06-02 ·

One or more techniques and/or computing devices are provided for utilizing snapshots for data integrity validation and/or faster application recovery. For example, a first storage controller, hosting first storage, has a synchronous replication relationship with a second storage controller hosting second storage. A snapshot replication policy rule is defined to specify that a replication label is to be used for snapshot create requests, targeting the first storage, that are to be replicated to the second storage. A snapshot creation policy is created to issue snapshot create requests comprising the replication label. Thus a snapshot of the first storage and a replication snapshot of the second storage are created based upon a snapshot create request comprising the replication label. The snapshot and the replication snapshot may be compared for data integrity validation (e.g., determine whether the snapshots comprise the same data) and/or quickly recovering an application after a disaster.

Workload Repetition Redundancy

A graphics processing system includes a plurality of processing units for processing tasks, each processing unit being configured to process a task independently from any other processing unit of the plurality of processing units; a check unit operable to form a signature which is characteristic of an output of a processing unit on processing a task; and a fault detection unit operable to compare signatures formed at the check unit; wherein the graphics processing system is configured to process each task of a first type first and second times at the plurality of processing units so as to, respectively, generate first and second processed outputs, wherein the check unit is configured to form first and second signatures which are characteristic of, respectively, the first and second processed outputs, and wherein the fault detection unit is configured to compare the first and second signatures and raise a fault signal if the first and second signatures do not match.

DECENTRALIZED AND DISTRIBUTED CONTINUOUS REPLICATION SYSTEM FOR MOVING DEVICES
20220174112 · 2022-06-02 ·

A replication system for data of mobile devices is disclosed. The data of a mobile device is uploaded to stations in an area. Metadata associated with the objects is stored in a centralized or decentralized system. The metadata can be accessed to identify the stations storing the device's objects and the data of the mobile device can then be retrieved from the stations and reconstructed.

DATA TRANSMISSION

A device, comprising: a main module; a plurality of secondary modules; and a data bus configured to enable data transmission between the main module and the plurality of secondary modules over a data line of the data bus; wherein each of the plurality of secondary modules is configured with a unique secondary address used by the main module to communicate with the respective secondary module over the data line, wherein the main module is operable to configure a first two or more of the plurality of secondary modules with a first common secondary address for simultaneous data transmission from the main module to the first two or more of the plurality of secondary modules over the data line.

SERVICING DATA STORAGE DEVICES IN A DATA STORAGE ARRAY
20220171677 · 2022-06-02 ·

Systems and methods for replacing and testing a data storage device are disclosed. In disclosed embodiments, a system including a data storage array (DSA) including a plurality of data storage devices (DSDs) in an enclosure. The system further includes an I/O server coupling the DSA to a client node and configured to provide data access between the client node and the DSA. The system further includes a management server coupled to the DSA, configured to detect a failed DSD in the DSA, detect a replacement DSD in the enclosure that replaces the failed DSD, and add the replacement DSD to a logical path of the DSA. The management server is further configured to display an indication of a state of the DSA based on the comparing.

IDENTIFYING FAULT DOMAINS FOR DELTA COMPONENTS OF A DISTRIBUTED DATA OBJECT

The disclosure herein describes placing delta components of a base component in target fault domains. One or more delta components are generated. When a first fault domain that lacks a sibling component of the base component is identified, the first fault domain is selected as a single delta target fault domain and a single delta component is placed on the single delta target fault domain. When a second fault domain that includes a first sibling component of the base component is identified and a third fault domain that includes a second sibling component of the base component is identified, the second fault domain and the third fault domain are selected as a first double delta target fault domain and a second double delta target fault domain, and a first double delta component and a second double delta component are placed on the first and second double delta target fault domains.

Service takeover method, storage device, and service takeover apparatus
11347603 · 2022-05-31 · ·

The present disclosure describes example service takeover methods, storage devices, and service takeover apparatuses. In one example method, when a communication fault occurs between two storage devices in a storage system, the two storage devices respectively obtain running statuses of the two storage devices. A running status can reflect current usage of one or more system resources of a particular storage device. Then, a delay duration is determined according to the running statuses, where the delay duration is a duration for which the storage device waits before sending an arbitration request to a quorum server. The two storage devices respectively send, after the delay duration, arbitration requests to the quorum server to request to take over a service. The quorum server then can select a storage device in a relatively better running status to take over a host service.

Evaluation for rebuilding performance of redundant arrays of independent disks

Embodiments of the present disclosure provide a solution of evaluating a rebuilding performance of a redundant array of independent disks. In some embodiments, there is provided a computer-implemented method, comprising: simulating, based on a first group of redundant arrays of independent disks, a rebuilding process for a second group of redundant arrays of independent disks; obtaining a first performance metric of the simulated rebuilding process; and identifying a factor associated with the rebuilding performance of the second group of redundant arrays of independent disks based on the first performance metric.

Methods for managing communications involving a lockstep processing system

A method for managing communications involving a lockstep processing comprising at least a first processor and a second processor can include receiving, at a data synchronizer, a first signal from a first device. The method can also include receiving, at the data synchronizer, a second signal from a second device. In addition, the method can include determining, by the data synchronizer, whether the first signal is equal to the second signal. When the first signal is equal to the second signal, the method can include transmitting, by the data synchronizer, the first signal to the first processor and the second signal to the second processor. Specifically, in example embodiments, transmitting the first signal to the first processor can occur synchronously with transmitting the second signal to the second processor.

Database transaction log migration

A system, method and apparatus for storing metadata in a metadata store in a robust and efficient manner including receiving a request from a client to perform a data transaction, updating a key-value pair in a metadata store based on the request, entering the data transaction in a transaction log, replicating the last transaction log entry in at least one other storage node in the metadata store and, in response to a threshold event, migrating a first portion of the transaction log from the a data storage medium to a second data storage medium of each of the plurality of data storage nodes.