Described embodiments provide systems and methods for operating a storage system. One or more production volumes of the storage system are selected for continuous replication. A number, N, is selected that is associated with a number of damaged volumes the storage system can sustain and maintain data consistency. Write transactions from a host to an associated one of the selected one or more production volumes are intercepted. The intercepted write transactions are sent to the associated production volume and to a plurality of copy volumes. When acknowledgments of the write transaction have been received from N copy volumes, the write transaction is acknowledged to the host.
|
1. A method comprising:
selecting one or more production volumes of a storage system for continuous replication;
selecting a number, N, associated with a number of damaged volumes the storage system can sustain and maintain data consistency;
intercepting one or more write transactions from a host to an associated one of the selected one or more production volumes;
sending the intercepted one or more write transactions to the associated production volume and to a plurality of copy volumes;
when acknowledgments of the one or more write transactions have been received from N copy volumes, acknowledging the one or more write transactions to the host;
assigning each of the intercepted one or more write transactions a unique transaction number, the transaction number monotonically increasing based upon an order in which the intercepted one or more write transactions are intercepted; and
verifying one or more of the plurality of copy volumes as synchronized with an associated production volume based upon the transaction numbers of the intercepted one or more write transactions written to each of the copy volumes.
8. A system comprising:
a processor; and
memory storing computer program code that when executed on the processor causes the processor to operate a storage system, the storage system operable to perform the operations of:
selecting one or more production volumes of a storage system for continuous replication;
selecting a number, N, associated with a number of damaged volumes the storage system can sustain and maintain data consistency, wherein N is an integer greater than zero and less than or equal to a sum of a number of production volumes and a number of copy volumes of the storage system;
intercepting one or more write transactions from a host to an associated one of the selected one or more production volumes;
sending the intercepted one or more write transactions to the associated production volume and to the copy volumes; and
when acknowledgments of the one or more write transactions have been received from N copy volumes, acknowledging the one or more write transactions to the host;
assigning each of the intercepted one or more write transactions a unique transaction number, the transaction number monotonically increasing based upon an order in which the one or more write transactions are intercepted; and
verifying one or more of the plurality of copy volumes as synchronized with an associated production volume based upon the transaction numbers of write transactions written to each of the copy volumes.
14. A computer program product including a non-transitory computer readable storage medium having computer program code encoded thereon that when executed on a processor of a computer causes the computer to operate a storage system, the computer program product comprising:
computer program code for selecting one or more production volumes of a storage system for continuous replication;
computer program code for selecting a number, N, associated with a number of damaged volumes the storage system can sustain and maintain data consistency, wherein N is an integer greater than zero and less than or equal to a sum of a number of production volumes and a number of copy volumes of the storage system;
computer program code for intercepting one or more write transactions from a host to an associated one of the selected one or more production volumes;
computer program code for sending the intercepted one or more write transactions to the associated production volume and to the copy volumes;
when acknowledgments of the one or more write transactions have been received from N copy volumes, computer program code for acknowledging the one or more write transactions to the host;
computer program code for assigning each of the intercepted one or more write transactions a unique transaction number, the transaction number monotonically increasing based upon an order in which the one or more write transactions are intercepted; and
computer program code for verifying one or more of the plurality of copy volumes as synchronized with an associated production volume based upon the transaction numbers of the one or more write transactions written to each of the copy volumes.
2. The method of
determining a highest transaction number written to each copy volume;
determining one or more copy volumes having a highest transaction number; and
indicating the determined one or more copy volumes as being synchronized with the associated production volume.
3. The method of
4. The method of
determining one or more write transactions received by all the copy volumes;
tracking metadata associated with write transactions received by fewer than all the copy volumes; and
generating at least one synchronous copy volume based upon the tracked metadata and the one or more write transactions received by all the copy volumes.
5. The method of
6. The method of
7. The method of
9. The system of
determining a highest transaction number written to each copy volume;
determining one or more copy volumes having a highest transaction number; and
indicating the determined one or more copy volumes as being synchronized with the associated production volume.
10. The system of
11. The system of
determining one or more write transactions received by all the copy volumes;
tracking metadata associated with write transactions received by fewer than all the copy volumes; and
generating at least one synchronous copy volume based upon the tracked metadata and the one or more write transactions received by all the copy volumes.
12. The system of
13. The system of
15. The computer program product of
computer program code for determining a highest transaction number written to each copy volume;
computer program code for determining one or more copy volumes having a highest transaction number; and
computer program code for indicating the determined one or more copy volumes as being synchronized with the associated production volume.
16. The computer program product of
17. The computer program product of
computer program code for determining one or more write transactions received by all the copy volumes;
computer program code for tracking metadata associated with write transactions received by fewer than all the copy volumes; and
computer program code for generating at least one synchronous copy volume based upon the tracked metadata and the one or more write transactions received by all the copy volumes.
|
A distributed storage system may include a plurality of storage devices (e.g., storage arrays) to provide data storage to a plurality of nodes. The plurality of storage devices and the plurality of nodes may be situated in the same physical location, or in one or more physically remote locations. A distributed storage system may include data protection systems that back up production site data by replicating production site data on a secondary backup storage system. The production site data may be replicated on a periodic basis and/or may be replicated as changes are made to the production site data. The backup storage system may be situated in the same physical location as the production storage system, or in a physically remote location.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
One aspect provides a method of operating a storage system. One or more production volumes of the storage system are selected for continuous replication. A number, N, is selected that is associated with a number of damaged volumes the storage system can sustain and maintain data consistency. Write transactions from a host to an associated one of the selected one or more production volumes are intercepted. The intercepted write transactions are sent to the associated production volume and to a plurality of copy volumes. When acknowledgments of the write transaction have been received from N copy volumes, the write transaction is acknowledged to the host.
Another aspect provides a system that includes a processor and memory storing computer program code that when executed on the processor causes the processor to operate a storage system. The storage system is operable to perform the operations of selecting one or more production volumes of the storage system for continuous replication. A number, N, is selected that is associated with a number of damaged volumes the storage system can sustain and maintain data consistency. Write transactions from a host to an associated one of the selected one or more production volumes are intercepted. The intercepted write transactions are sent to the associated production volume and to a plurality of copy volumes. When acknowledgments of the write transaction have been received from N copy volumes, the write transaction is acknowledged to the host.
Another aspect provides a computer program product including a non-transitory computer readable storage medium having computer program code encoded thereon that when executed on a processor of a computer causes the computer to operate a storage system. The computer program product includes computer program code for selecting one or more production volumes of the storage system for continuous replication. A number, N, is selected that is associated with a number of damaged volumes the storage system can sustain and maintain data consistency. Write transactions from a host to an associated one of the selected one or more production volumes are intercepted. The intercepted write transactions are sent to the associated production volume and to a plurality of copy volumes. When acknowledgments of the write transaction have been received from N copy volumes, the write transaction is acknowledged to the host.
Objects, aspects, features, and advantages of embodiments disclosed herein will become more fully apparent from the following detailed description, the appended claims, and the accompanying drawings in which like reference numerals identify similar or identical elements. Reference numerals that are introduced in the specification in association with a drawing figure may be repeated in one or more subsequent figures without additional description in the specification in order to provide context for other features. For clarity, not every element may be labeled in every figure. The drawings are not necessarily to scale, emphasis instead being placed upon illustrating embodiments, principles, and concepts. The drawings are not meant to limit the scope of the claims included herewith.
Before describing embodiments of the concepts, structures, and techniques sought to be protected herein, some terms are explained. In some embodiments, the term “I/O request” or simply “I/O” may be used to refer to an input or output request. In some embodiments, an I/O request may refer to a data read or data write request. In some embodiments, the term “storage system” may encompass physical computing systems, cloud or virtual computing systems, or a combination thereof. In some embodiments, the term “storage device” may refer to any non-volatile memory (NVM) device, including hard disk drives (HDDs), solid state drivers (SSDs), flash devices (e.g., NAND flash devices), and similar devices that may be accessed locally and/or remotely (e.g., via a storage attached network (SAN)). In some embodiments, the term “storage device” may also refer to a storage array including multiple storage devices.
Referring to the example embodiment shown in
In certain embodiments, Site I 100a and Site II 100b may be remote from one another. In other embodiments, Site I 100a and Site II 100b may be local to one another and may be connected via a local area network (LAN). In some embodiments, local data protection may have the advantage of minimizing data lag between target and source, and remote data protection may have the advantage of being robust in the event that a disaster occurs at the source site.
In particular embodiments, data protection system 100 may include a failover mode of operation, wherein the direction of replicated data flow is reversed. In particular, in some embodiments, Site I 100a may behave as a target site and Site II 100b may behave as a source site. In some embodiments, failover may be triggered manually (e.g., by a user) or automatically. In many embodiments, failover may be performed in the event of a disaster at Site I 100a. In some embodiments, both Site I 100a and Site II 100b may behave as source site for some stored data and may behave simultaneously as a target site for other stored data. In certain embodiments, a portion of stored data may be replicated from one site to the other, and another portion may not be replicated.
In some embodiments, Site I 100a corresponds to a production site (e.g., a facility where one or more hosts run data processing applications that write data to a storage system and read data from the storage system) and Site II 100b corresponds to a backup or replica site (e.g., a facility where replicated production site data is stored). Thus, in some embodiments, Site II 100b may be responsible for replicating production site data and may enable rollback of data of Site I 100a to an earlier point in time. In some embodiments, rollback may be used in the event of data corruption of a disaster, or alternatively in order to view or to access data from an earlier point in time.
Described embodiments of Site I 100a may include a source host 104, a source storage system (or “storage array”) 108, and a source data protection appliance (DPA) 112 coupled via a first storage area network (SAN). Similarly, in some embodiments, Site II 100b may include a target host 116, a target storage system 120, and a target DPA 124 coupled via a second SAN. In some embodiments, each SAN may include one or more devices (or “nodes”) that may be designated an “initiator,” a “target”, or both. For example, in some embodiments, the first SAN may include a first fiber channel switch 148 and the second SAN may include a second fiber channel switch 168. In some embodiments, communication links between each host 104 and 116 and its corresponding storage system 108 and 120 may be any appropriate medium suitable for data transfer, such as fiber communication channel links. In many embodiments, a host communicates with its corresponding storage system over a communication link, such as an InfiniBand (IB) link or Fibre Channel (FC) link, and/or a network, such as an Ethernet or Internet (e.g., TCP/IP) network that may employ, for example, the iSCSI protocol.
In some embodiments, each storage system 108 and 120 may include storage devices for storing data, such as disks or arrays of disks. Typically, in such embodiments, storage systems 108 and 120 may be target nodes. In some embodiments, in order to enable initiators to send requests to storage system 108, storage system 108 may provide (e.g., expose) one or more logical units (LU) to which commands are issued. Thus, in some embodiments, storage systems 108 and 120 may be SAN entities that provide multiple logical units for access by multiple SAN initiators. In some embodiments, an LU is a logical entity provided by a storage system for accessing data stored therein. In some embodiments, a logical unit may be a physical logical unit or a virtual logical unit. In some embodiments, a logical unit may be identified by a unique logical unit number (LUN).
In the embodiment shown in
As shown in
In some embodiments, source host 104 may act as a SAN initiator that issues I/O requests through host device 140 to LU A 136 using, for example, SCSI commands. In some embodiments, such requests may be transmitted to LU A 136 with an address that includes a specific device identifier, an offset within the device, and a data size.
In some embodiments, source DPA 112 and target DPA 124 may perform various data protection services, such as data replication of a storage system, and journaling of I/O requests issued by hosts 104 and/or 116. In some embodiments, when acting as a target DPA, a DPA may also enable rollback of data to an earlier point-in-time (PIT), and enable processing of rolled back data at the target site. In some embodiments, each DPA 112 and 124 may be a physical device, a virtual device, or may be a combination of a virtual and physical device.
In some embodiments, a DPA may be a cluster of such computers. In some embodiments, use of a cluster may ensure that if a DPA computer is down, then the DPA functionality switches over to another computer. In some embodiments, the DPA computers within a DPA cluster may communicate with one another using at least one communication link suitable for data transfer, for example, an InfiniBand (TB) link, a Fibre Channel (FC) link, and/or a network link, such as an Ethernet or Internet (e.g., TCP/IP) link to transfer data via fiber channel or IP based protocols, or other such transfer protocols. In some embodiments, one computer from the DPA cluster may serve as the DPA leader. In some embodiments, the DPA cluster leader may coordinate between the computers in the cluster, and may also perform other tasks that require coordination between the computers, such as load balancing.
In certain embodiments, a DPA may be a standalone device integrated within a SAN. Alternatively, in some embodiments, a DPA may be integrated into storage system. In some embodiments, the DPAs communicate with their respective hosts through communication links suitable for data transfer, for example, an InfiniBand (TB) link, a Fibre Channel (FC) link, and/or a network link, such as an Ethernet or Internet (e.g., TCP/IP) link to transfer data via, for example, SCSI commands or any other protocol.
In various embodiments, the DPAs may act as initiators in the SAN. For example, in some embodiments, the DPAs may issue I/O requests using, for example, SCSI commands, to access LUs on their respective storage systems. In some embodiments, each DPA may also be configured with the necessary functionality to act as targets, e.g., to reply to I/O requests, such as SCSI commands, issued by other initiators in the SAN, including their respective hosts. In some embodiments, being target nodes, the DPAs may dynamically expose or remove one or more LUs. As described herein, in some embodiments, Site I 100a and Site II 100b may each behave simultaneously as a production site and a backup site for different logical units. As such, in some embodiments, DPA 112 and DPA 124 may each behave as a source DPA for some LUs and as a target DPA for other LUs, at the same time.
In the example embodiment shown in
In various embodiments, a protection agent may change its behavior for handling SCSI commands, for example as a result of an instruction received from the DPA. For example, in some embodiments, the behavior of a protection agent for a certain host device may depend on the behavior of its associated DPA with respect to the LU of the host device. In some embodiments, when a DPA behaves as a source site DPA for a certain LU, then during normal course of operation, the associated protection agent may split I/O requests issued by a host to the host device corresponding to that LU. Similarly, in some embodiments, when a DPA behaves as a target device for a certain LU, then during normal course of operation, the associated protection agent fails I/O requests issued by host to the host device corresponding to that LU.
In some embodiments, communication between protection agents 144 and 164 and a respective DPA 112 and 124 may use any protocol suitable for data transfer within a SAN, such as fiber channel, SCSI over fiber channel, or other protocols. In some embodiments, the communication may be direct, or via a logical unit exposed by the DPA.
In certain embodiments, protection agents may be drivers located in their respective hosts. Alternatively, in some embodiments, a protection agent may also be located in a fiber channel switch, or in any other device situated in a data path between a host and a storage system or on the storage system itself. In some embodiments, in a virtualized environment, the protection agent may run at the hypervisor layer or in a virtual machine providing a virtualization layer.
As shown in the example embodiment shown in
Some embodiments of data protection system 100 may be provided as physical systems for the replication of physical LUs, or as virtual systems for the replication of virtual LUs. For example, in one embodiment, a hypervisor may consume LUs and may generate a distributed file system on the logical units such as Virtual Machine File System (VMFS) that may generate files in the file system and expose the files as LUs to the virtual machines (each virtual machine disk is seen as a SCSI device by virtual hosts). In another embodiment, a hypervisor may consume a network based file system and exposes files in the Network File System (NFS) as SCSI devices to virtual hosts.
In some embodiments, in normal operation (sometimes referred to as “production mode”), DPA 112 may act as a source DPA for LU A 136. Thus, in some embodiments, protection agent 144 may act as a source protection agent, specifically by splitting I/O requests to host device 140 (“Device A”). In some embodiments, protection agent 144 may send an I/O request to source DPA 112 and, after receiving an acknowledgement from source DPA 112, may send the I/O request to LU A 136. In some embodiments, after receiving an acknowledgement from storage system 108, host 104 may acknowledge that the I/O request has successfully completed.
In some embodiments, when source DPA 112 receives a replicated I/O request from protection agent 144, source DPA 112 may transmit certain I/O information characterizing the write request, packaged as a “write transaction”, over WAN 128 to target DPA 124 for journaling and for incorporation within target storage system 120. In some embodiments, when applying write operations to storage system 120, target DPA 124 may act as an initiator, and may send SCSI commands to LU B 156.
In some embodiments, source DPA 112 may send its write transactions to target DPA 124 using a variety of modes of transmission, including (i) a synchronous mode, (ii) an asynchronous mode, and (iii) a snapshot mode.
In some embodiments, in synchronous (or continuous) mode, source DPA 112 may send each write transaction to target DPA 124, may receive back an acknowledgement from the target DPA 124, and in turn may send an acknowledgement back to protection agent 144. In some embodiments, in synchronous mode, protection agent 144 may wait until receipt of such acknowledgement before sending the I/O request to LU 136.
Thus, as described herein, in some embodiments, in continuous mode, every write I/O to a production volume may be intercepted and sent to both the production volume and a replica volume. Thus, in some embodiments, continuous mode may provide a very low Recovery Point Objective (RPO), meaning that data on a replica volume lags data on the production volume by only a short time period (e.g., a few seconds). In some embodiments, RPO is an amount of data that the user is willing to lose in case of production disaster (e.g., an amount of time between replications). In some embodiments (e.g., for particularly important data), continuous mode may provide an RPO of zero (e.g., data on the replica volume is the same as data on the production volume) and high granularity of points in time (PITs) for restoring the production volume (e.g., since continuous replication may generate a replica each time there is a write operation to the production volume).
In some embodiments, in asynchronous mode, source DPA 112 may send an acknowledgement to protection agent 144 upon receipt of each I/O request, before receiving an acknowledgement back from target DPA 124.
In some embodiments, in snapshot mode, source DPA 112 may receive several I/O requests and combine them into an aggregate “snapshot” or “batch” of write activity performed in the multiple I/O requests, and may send the snapshot to target DPA 124 for journaling and incorporation in target storage system 120. In some embodiments, in snapshot mode, source DPA 112 may send an acknowledgement to protection agent 144 upon receipt of each I/O request, before receiving an acknowledgement back from target DPA 124.
Thus, in some embodiments, in snapshot mode, snapshot replicas of a production volume may be periodically generated, and changes in data may be tracked between consecutive snapshot replicas. For example, one or more write operations may modify data on the production volume between generation of snapshot replicas. In some embodiments, regions of the production volume that are modified, and the changed data written to the regions, may be tracked. In some embodiments, when a new snapshot replica is generated, modified regions may be read from the production volume and sent to the replica volume.
In some embodiments, a snapshot replica may be a differential representation of a volume. For example, the snapshot may include pointers to the original volume, and may point to log volumes for locations of the original volume that store data changed by one or more I/O requests. In some embodiments, snapshots may be combined into a snapshot array, which may represent different images over a time period (e.g., for multiple PITs).
In some embodiments, in normal operation, LU B 156 may be used as a backup of LU A 136. In such embodiments, while data written to LU A 136 by host 104 is replicated from LU A 136 to LU B 156, target host 116 should not send I/O requests to LU B 156. To prevent such I/O requests from being sent, in some embodiments, protection agent 164 may act as a target site protection agent for host device B 160 and may fail I/O requests sent from host 116 to LU B 156 through host device B 160. In some embodiments, in a recovery mode, target DPA 124 may undo the write transactions in journal LU 176 so as to restore the target storage system 120 to an earlier state.
Referring to
Referring to both
Since the journal contains the “undo” information necessary to rollback storage system 120, in some embodiments, data that was stored in specific memory locations at a specified point in time may be obtained by undoing write transactions that occurred subsequent to such point in time (PIT).
In some embodiments, each of the four streams may hold a plurality of write transaction data. In some embodiments, as write transactions are received dynamically by the target DPA, the write transactions may be recorded at the end of the DO stream and the end of the DO METADATA stream, prior to committing the transaction.
In some embodiments, a metadata stream (e.g., UNDO METADATA stream or the DO METADATA stream) and the corresponding data stream (e.g., UNDO stream or DO stream) may be kept in a single stream by interleaving metadata and data.
Some described embodiments may validate that point-in-time (PIT) data replicas (e.g., data replicated to LU B 156) are valid and usable, for example to verify that the data replicas are not corrupt due to a system error or inconsistent due to violation of write order fidelity. In some embodiments, validating data replicas can be important, for example, in data replication systems employing incremental backup where an undetected error in an earlier data replica may lead to corruption of future data replicas.
In conventional systems, validating data replicas can increase the journal lag, which may increase a recovery time objective (RTO) of a data protection system (e.g., an elapsed time between replicas or PITs). In such conventional systems, if the journal lag time is significant, the journal may become full and unable to account for data changes due to subsequent transactions. Further, in such conventional systems, validating data replicas may consume system resources (e.g., processor time, memory, communication link bandwidth, etc.), resulting in reduced performance for system tasks.
Referring to
In an embodiment, datastore 306 may include one or more differential virtual machine disks, shown as differential VMDKs 346. Some embodiments of datastore 306 may also include journal virtual machine disk 348. In some embodiments, differential VMDKs 346 and journal VMDK 348 may be stored in datastore 306, and one or more production virtual machine disks 342 may be stored in datastore 307. In some embodiments, datastore 306 and datastore 307 are separate physical devices so that access to differential VMDKs does not affect performance of production VMDKs. In some embodiments, differential VMDKs 346 may be used to store differential snapshot data representative of changes that happened to data stored on production VMDK 342. In one example, a first differential VMDK 346 may include changes due to writes that occurred to production VMDK 342 from time t1 to time t2, a second differential VMDK 346 may include the changes due to writes that occurred to production VMDK 342 from time t2 to time t3, and so forth.
In some embodiments, differential VMDKs 346 may be thin provisioned. Thin provisioning allocates storage space to volumes of a SAN in a flexible manner among multiple volumes based on a minimum space requirement for each volume at any given time.
In some embodiments, journal 352 may be stored in journal VMDK 348. Journal 352 includes one or more delta marker streams (DMS) 362. In some embodiments, each DMS 362 may include metadata associated with data that may be different between one differential VMDK and another differential VMDK.
In one example, DMS 362 may include the metadata differences between a current copy of the production VMDK 342 and a copy currently stored in backup storage 304. In some embodiments, journal 352 does not include the actual data changes, but rather metadata associated with the changes. In some embodiments, the data of the changes may be stored in the differential VMDKs.
Although not shown in
In some embodiments, data protection system may include one or more consistency groups. In some embodiments, a consistency group may treat source volumes (e.g., production volumes) and target volumes (e.g., backup volumes) as a single logical entity for data replication and migration.
Embodiments employing synchronous (or continuous) replication may provide zero data loss (e.g., a zero RPO) in case of a failed or damaged production volume. In some embodiments, since, in synchronous replication, data is sent to the replica “inline” (e.g., as part of the write transaction), data is not read from the production volume to generate a replica. However, since, in some embodiments, every write transaction sent to the production volume is also sent to the replica volume, network bandwidth requirements of continuous replication can be high (e.g., as high as the bandwidth of peak writes to the production volume). Further, since, in some embodiments, write transactions are sent to both the production and replica volumes, system performance can be lessened because each write transaction needs to be received by the replica volume(s) before the write transaction can be acknowledged.
Illustrative multi-copy data replication systems may commonly (but not always) be active-passive systems having one active production volume and one or more passive replica volumes. In illustrative active-passive systems, links between the production volume and each replica volume are configured independently (e.g., per link) as either synchronous or asynchronous. In such systems, if the link is synchronous then the replica volume is always consistent with the latest state of the production volume (for example, synchronous or continuous replication). In such systems, if the link is asynchronous, the replica volume eventually becomes consistent with the production volume (for example asynchronous or snapshot replication).
Some embodiments of data protection system 100 may be “active-active” systems having multiple nodes (e.g., multiple production sites and multiple target sites). Illustrative active-active systems may allow concurrent access to both a production site and a target site, for example as in a distributed database. In such active-active systems, a read-write quorum may be employed when all nodes need to be consistent (e.g., have synchronized data) with zero data loss (e.g., zero RPO). For example, in a conventional active-active system having a large number of nodes and/or employing database sharding (e.g., partitioning) across multiple nodes, each write transaction may need to reach at least a subset of the nodes before the write transaction is acknowledged. Thus, in conventional active-active systems, at least some of the copy nodes are defined as synchronous copies.
Some described embodiments provide multi-copy data replication having a zero RPO without defining any of the copies as a synchronous copy. As will be described, in some embodiments, each write transaction for a production volume may be intercepted (e.g., by protection agent 144 of
In some embodiments, acknowledgements of the write transaction from remaining copy volumes may be received while the data protection system is performing subsequent transactions. Thus, in some embodiments, performance of data protection system 100 can be improved when requiring synchronous (or continuous) replication by decreasing an elapsed time to perform write transactions (e.g., by not waiting for all copy volumes to acknowledge a write transaction before performing subsequent transactions). Thus, in some described embodiments, synchronous replication could be performed to multiple cloud (e.g., virtual) copy volumes without waiting for all the cloud copy volumes to acknowledge each write transaction. This may increase system performance of such embodiments since response time of cloud copy volumes could vary between each cloud copy volume (e.g., depending on location, performance of networks specific to each cloud volume, etc.). Further, some described embodiments do not rely upon a single cloud copy volume to store data copies, which may improve system robustness and improve average response time by having multiple copies.
Referring to
In some embodiments, at block 408, write transactions to the production volumes selected at block 404 are intercepted, for example by data protection agent 144. In some embodiments, at block 410, the data protection agent sends the write transaction to the production volumes and the copy volumes. In some embodiments, at block 412, the data protection agent waits to receive acknowledgments (acks) from the copy volumes. In some embodiments, at block 414, process 400 returns to block 412 until acks are received from N copy volumes. In some embodiments, at block 414, once acks are received from N copy volumes, then at block 416, the write transaction is acknowledged as completed, and synchronous replication process 400 completes at block 418.
Referring to
As shown in
Referring to
In some embodiments, process 600 may be performed, for example, after block 416 of process 400 of
In some described embodiments, hosts 104 and 116 of
Processes 400, 500 and 600 (
The processes described herein are not limited to the specific embodiments described. For example, processes 400, 500 and 600 are not limited to the specific processing order shown in
Processor 702 may be implemented by one or more programmable processors executing one or more computer programs to perform the functions of the system. As used herein, the term “processor” describes an electronic circuit that performs a function, an operation, or a sequence of operations. The function, operation, or sequence of operations may be hard coded into the electronic circuit or soft coded by way of instructions held in a memory device. A “processor” may perform the function, operation, or sequence of operations using digital values or using analog signals. In some embodiments, the “processor” can be embodied in one or more application specific integrated circuits (ASICs). In some embodiments, the “processor” may be embodied in one or more microprocessors with associated program memory. In some embodiments, the “processor” may be embodied in one or more discrete electronic circuits. The “processor” may be analog, digital or mixed-signal. In some embodiments, the “processor” may be one or more physical processors and/or one or more “virtual” (e.g., remotely located or “cloud”) processors.
Various functions of circuit elements may also be implemented as processing blocks in a software program. Such software may be employed in, for example, one or more digital signal processors, microcontrollers, or general purpose computers. Described embodiments may be implemented in hardware, a combination of hardware and software, software, or software in execution by one or more physical and/or one or more virtual processors.
Some embodiments may be implemented in the form of methods and apparatuses for practicing those methods. Described embodiments may also be implemented in the form of program code, for example, stored in a storage medium, loaded into and/or executed by a machine, or transmitted over some transmission medium or carrier, such as over electrical wiring or cabling, through fiber optics, or via electromagnetic radiation. A non-transitory machine-readable medium may include but is not limited to tangible media, such as magnetic recording media including hard drives, floppy diskettes, and magnetic tape media, optical recording media including compact discs (CDs) and digital versatile discs (DVDs), solid state memory such as flash memory, hybrid magnetic and solid state memory, non-volatile memory, volatile memory, and so forth, but does not include a transitory signal per se. When embodied in a non-transitory machine-readable medium and the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the method.
When implemented on a processing device, the program code segments combine with the processor to provide a unique device that operates analogously to specific logic circuits. Such processing devices may include, for example, a general purpose microprocessor, a digital signal processor (DSP), a reduced instruction set computer (RISC), a complex instruction set computer (CISC), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a programmable logic array (PLA), a microcontroller, an embedded controller, a multi-core processor, and/or others, including combinations of one or more of the above. Described embodiments may also be implemented in the form of a bitstream or other sequence of signal values electrically or optically transmitted through a medium, stored magnetic-field variations in a magnetic recording medium, etc., generated using a method and/or an apparatus as recited in the claims.
Various elements, which are described in the context of a single embodiment, may also be provided separately or in any suitable subcombination. It will be further understood that various changes in the details, materials, and arrangements of the parts that have been described and illustrated herein may be made by those skilled in the art without departing from the scope of the following claims.
Natanzon, Assaf, Bigman, Ron, Lieberman, Amit, Shemer, Jehuda, Baruch, Leehod
Patent | Priority | Assignee | Title |
10140039, | Dec 15 2016 | EMC IP HOLDING COMPANY LLC | I/O alignment for continuous replication in a storage system |
10223023, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Bandwidth reduction for multi-level data replication |
10235060, | Apr 14 2016 | EMC IP Holding Company, LLC | Multilevel snapshot replication for hot and cold regions of a storage system |
10235061, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Granular virtual machine snapshots |
10235087, | Mar 30 2016 | EMC IP HOLDING COMPANY LLC | Distributing journal data over multiple journals |
10235088, | Mar 30 2016 | EMC IP HOLDING COMPANY LLC | Global replication policy for multi-copy replication |
10235145, | Sep 13 2012 | EMC International Company | Distributed scale-out replication |
10235247, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Compressing memory snapshots |
10324798, | Sep 25 2014 | EMC IP HOLDING COMPANY LLC | Restoring active areas of a logical unit |
10366011, | May 03 2018 | EMC IP HOLDING COMPANY LLC | Content-based deduplicated storage having multilevel data cache |
10409629, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Automated host data protection configuration |
10409787, | Dec 22 2015 | EMC IP HOLDING COMPANY LLC | Database migration |
10409986, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Ransomware detection in a continuous data protection environment |
10423634, | Dec 27 2016 | EMC IP HOLDING COMPANY LLC | Temporal queries on secondary storage |
10437783, | Sep 25 2014 | EMC IP HOLDING COMPANY LLC | Recover storage array using remote deduplication device |
10467102, | Dec 15 2016 | EMC IP HOLDING COMPANY LLC | I/O score-based hybrid replication in a storage system |
10489321, | Jul 31 2018 | EMC IP HOLDING COMPANY LLC | Performance improvement for an active-active distributed non-ALUA system with address ownerships |
10496487, | Dec 03 2014 | EMC IP HOLDING COMPANY LLC | Storing snapshot changes with snapshots |
10496601, | Oct 09 2017 | EMC IP HOLDING COMPANY LLC | Efficient file system parsing using snap based replication |
10579282, | Mar 30 2016 | EMC IP HOLDING COMPANY LLC | Distributed copy in multi-copy replication where offset and size of I/O requests to replication site is half offset and size of I/O request to production volume |
10592166, | Aug 01 2018 | EMC IP HOLDING COMPANY LLC | Fast input/output in a content-addressable storage architecture with paged metadata |
10628268, | Dec 15 2016 | EMC IP HOLDING COMPANY LLC | Proof of data replication consistency using blockchain |
10713221, | Jul 30 2018 | EMC IP HOLDING COMPANY LLC | Dual layer deduplication for a file system running over a deduplicated block storage |
10747606, | Dec 21 2016 | EMC IP HOLDING COMPANY LLC | Risk based analysis of adverse event impact on system availability |
10747667, | Nov 02 2018 | EMC IP HOLDING COMPANY LLC | Memory management of multi-level metadata cache for content-based deduplicated storage |
10776211, | Dec 27 2016 | EMC IP HOLDING COMPANY LLC | Methods, systems, and apparatuses to update point in time journal using map reduce to create a highly parallel update |
10853181, | Jun 29 2015 | EMC IP HOLDING COMPANY LLC | Backing up volumes using fragment files |
10853286, | Jul 31 2018 | EMC IP HOLDING COMPANY LLC | Performance improvement for an active-active distributed non-ALUA system with address ownerships |
10944819, | Oct 26 2018 | Hewlett Packard Enterprise Development LP | Replication of an encrypted volume |
11016677, | Dec 13 2016 | EMC IP HOLDING COMPANY LLC | Dual-splitter for high performance replication |
11068364, | Sep 05 2019 | WESTERN DIGITAL TECHNOLOGIES INC ; INTELLIFLASH BY DDN, INC | Predictable synchronous data replication |
11093158, | Jan 29 2019 | EMC IP HOLDING COMPANY LLC | Sub-lun non-deduplicated tier in a CAS storage to reduce mapping information and improve memory efficiency |
11144247, | Aug 01 2018 | EMC IP HOLDING COMPANY LLC | Fast input/output in a content-addressable storage architecture with paged metadata |
11233850, | Apr 17 2018 | Hewlett Packard Enterprise Development LP | Replicating data over a public network |
11379246, | Jul 24 2019 | EMC IP HOLDING COMPANY LLC | Automatic configuration of multiple virtual storage processors |
11494108, | Sep 23 2019 | Amazon Technologies, Inc. | Cross-zone replicated block storage devices |
11537725, | Sep 23 2019 | Amazon Technologies, Inc. | Encrypted cross-zone replication for cross-zone replicated block storage devices |
11609931, | Jun 27 2019 | DATADOG, INC | Ring replication system |
11960504, | Sep 02 2021 | Bank of America Corporation | Data replication over low-latency network |
Patent | Priority | Assignee | Title |
5170480, | Sep 25 1989 | International Business Machines Corporation | Concurrently applying redo records to backup database in a log sequence using single queue server per queue at a time |
5249053, | Feb 05 1991 | DYCAM INC | Filmless digital camera with selective image compression |
5388254, | Mar 27 1992 | International Business Machines Corporation | Method and means for limiting duration of input/output (I/O) requests |
5499367, | Nov 15 1991 | Oracle International Corporation | System for database integrity with multiple logs assigned to client subsets |
5526397, | Apr 20 1992 | Hughes Electronics Corporation | Switching transcoder |
5864837, | Jun 12 1996 | Unisys Corporation | Methods and apparatus for efficient caching in a distributed environment |
5879459, | Aug 29 1997 | EUGENUS, INC | Vertically-stacked process reactor and cluster tool system for atomic layer deposition |
5990899, | Oct 27 1995 | Microsoft Technology Licensing, LLC | Method for compressing journal streams |
6042852, | Mar 23 1993 | General Mills, Inc | Low pressure refrigerated dough product |
6065018, | Mar 04 1998 | GLOBALFOUNDRIES Inc | Synchronizing recovery log having time stamp to a remote site for disaster recovery of a primary database having related hierarchial and relational databases |
6143659, | Nov 18 1997 | Samsung Electronics, Co., Ltd. | Method for manufacturing aluminum metal interconnection layer by atomic layer deposition method |
6148340, | Apr 30 1998 | IBM Corporation | Method and system for differencing container files |
6174377, | Mar 03 1997 | AIXTRON, INC | Processing chamber for atomic layer deposition processes |
6174809, | Dec 31 1997 | Samsung Electronics, Co., Ltd. | Method for forming metal layer using atomic layer deposition |
6203613, | Oct 19 1999 | International Business Machines Corporation | Atomic layer deposition with nitrate containing precursors |
6260125, | Dec 09 1998 | TERADATA US, INC | Asynchronous write queues, reconstruction and check-pointing in disk-mirroring applications |
6270572, | Aug 07 1998 | SAMSUNG ELECTRONICS CO , LTD | Method for manufacturing thin film using atomic layer deposition |
6272534, | Mar 04 1998 | Storage Technology Corporation | Method and system for efficiently storing web pages for quick downloading at a remote device |
6287965, | Jul 28 1997 | SAMSUNG ELECTRONICS, CO , LTD | Method of forming metal layer using atomic layer deposition and semiconductor device having the metal layer as barrier metal layer or upper or lower electrode of capacitor |
6467023, | Mar 23 1999 | AVAGO TECHNOLOGIES INTERNATIONAL SALES PTE LIMITED | Method for logical unit creation with immediate availability in a raid storage environment |
6574657, | May 03 1999 | CLOUDING CORP | Methods and apparatuses for file synchronization and updating using a signature list |
6621493, | Oct 27 1995 | Microsoft Technology Licensing, LLC | Metafile compression |
6804676, | Aug 31 1999 | International Business Machines Corporation | System and method in a data processing system for generating compressed affinity records from data records |
6947981, | Mar 26 2002 | Hewlett Packard Enterprise Development LP | Flexible data replication mechanism |
7043610, | Aug 19 2002 | ADAPTEC INC | System and method for maintaining cache coherency without external controller intervention |
7051126, | Aug 19 2003 | F5 Networks, Inc. | Hardware accelerated compression |
7076620, | Feb 27 2003 | Hitachi, LTD | Data processing system including storage systems |
7111197, | Sep 21 2001 | VALTRUS INNOVATIONS LIMITED | System and method for journal recovery for multinode environments |
7117327, | Apr 28 2004 | Hitachi, Ltd. | Data processing system |
7120768, | Mar 22 2002 | Hitachi, Ltd. | Snapshot acquisition method, storage system and disk apparatus |
7130975, | Sep 09 2003 | Hitachi, Ltd. | Data processing system |
7139927, | Mar 21 2002 | Wisconsin Alumni Research Foundation | Journaling and recovery method of shared disk file system |
7159088, | Aug 04 2004 | Hitachi, Ltd. | Storage system and data processing system |
7167963, | Apr 28 2004 | Hitachi, Ltd. | Storage system with multiple remote site copying capability |
7203741, | Oct 12 2000 | ZEPHYRTEL, INC | Method and system for accelerating receipt of data in a client-to-client network |
7222136, | May 23 2002 | Oracle International Corporation | Communicating data dictionary information of database objects through a redo stream |
7296008, | Aug 24 2004 | Veritas Technologies LLC | Generation and use of a time map for accessing a prior image of a storage device |
7328373, | Jan 30 2004 | GOOGLE LLC | Data processing system |
7353335, | Feb 03 2006 | Hitachi, Ltd. | Storage control method for database recovery in logless mode |
7360113, | Aug 30 2004 | Veritas Technologies LLC | Protocol for communicating data block copies in an error recovery environment |
7426618, | Sep 06 2005 | Dot Hill Systems Corp. | Snapshot restore method and apparatus |
7519625, | Sep 27 2005 | Hitachi, Ltd. | Snapshot management apparatus and method, and storage system |
7519628, | Jun 01 2004 | NetApp, Inc | Technique for accelerating log replay with partial cache flush |
7546485, | Aug 15 2006 | Hewlett Packard Enterprise Development LP | Method and system for efficient journal-based resynchronization |
7590887, | Sep 16 2003 | Hitachi, Ltd. | Mapping apparatus for backup and restoration of multi-generation recovered snapshots |
7606940, | Jun 23 2003 | Hitachi, Ltd. | Remote copy system |
7719443, | Jun 27 2008 | EMC BENELUX B V , S A R L | Compressing data in a continuous data protection environment |
7757057, | Nov 27 2006 | AVAGO TECHNOLOGIES GENERAL IP SINGAPORE PTE LTD | Optimized rollback of copy-on-write snapshot volumes |
7840536, | Dec 26 2007 | EMC BENELUX B V , S A R L | Methods and apparatus for dynamic journal expansion |
7840662, | Mar 28 2008 | EMC BENELUX B V , S A R L | Dynamically managing a network cluster |
7844856, | Dec 26 2007 | EMC BENELUX B V , S A R L | Methods and apparatus for bottleneck processing in a continuous data protection system having journaling |
7860836, | Dec 26 2007 | EMC BENELUX B V , S A R L | Method and apparatus to recover data in a continuous data protection environment using a journal |
7882286, | Sep 26 2008 | EMC BENELUX B V , S A R L | Synchronizing volumes for replication |
7934262, | Dec 26 2007 | EMC BENELUX B V , S A R L | Methods and apparatus for virus detection using journal data |
7958372, | Dec 26 2007 | EMC BENELUX B V , S A R L | Method and apparatus to convert a logical unit from a first encryption state to a second encryption state using a journal in a continuous data protection environment |
8037162, | Oct 12 2000 | ZEPHYRTEL, INC | Method and system for accelerating receipt of data in a client to client network |
8041940, | Dec 26 2007 | EMC BENELUX B V , S A R L | Offloading encryption processing in a storage area network |
8060713, | Dec 21 2005 | EMC BENELUX B V , S A R L | Consolidating snapshots in a continuous data protection system using journaling |
8060714, | Sep 26 2008 | EMC BENELUX B V , S A R L | Initializing volumes in a replication system |
8103937, | Mar 31 2010 | EMC IP HOLDING COMPANY LLC | Cas command network replication |
8108634, | Jun 27 2008 | EMC B.V., S.A.R.L. | Replicating a thin logical unit |
8205009, | Apr 25 2002 | EMC IP HOLDING COMPANY LLC; EMC Corporation | Apparatus for continuous compression of large volumes of data |
8214612, | Sep 28 2009 | EMC IP HOLDING COMPANY LLC | Ensuring consistency of replicated volumes |
8250149, | Oct 12 2000 | ZEPHYRTEL, INC | Method and system for accelerating receipt of data in a client to client network |
8271441, | Dec 26 2009 | EMC IP HOLDING COMPANY LLC | Virtualized CG |
8271447, | Jun 18 2010 | EMC International Company | Mirroring metadata in a continuous data protection environment |
8332687, | Jun 23 2010 | EMC IP HOLDING COMPANY LLC | Splitter used in a continuous data protection environment |
8335761, | Dec 02 2010 | EMC International Company | Replicating in a multi-copy environment |
8335771, | Sep 29 2010 | EMC IP HOLDING COMPANY LLC | Storage array snapshots for logged access replication in a continuous data protection system |
8341115, | Dec 26 2009 | EMC International Company | Dynamically switching between synchronous and asynchronous replication |
8370648, | Mar 15 2010 | EMC International Company | Writing and reading encrypted data using time-based encryption keys |
8380885, | Jun 30 2011 | EMC IP HOLDING COMPANY LLC | Handling abort commands in replication |
8392680, | Mar 30 2010 | EMC International Company | Accessing a volume in a distributed environment |
8429362, | Mar 31 2011 | EMC IP HOLDING COMPANY LLC | Journal based replication with a virtual service layer |
8433869, | Sep 27 2010 | EMC International Company | Virtualized consistency group using an enhanced splitter |
8438135, | Jun 18 2010 | EMC International Company | Mirroring metadata in a continuous data protection environment |
8464101, | Mar 31 2010 | EMC IP HOLDING COMPANY LLC | CAS command network replication |
8478955, | Sep 27 2010 | EMC International Company | Virtualized consistency group using more than one data protection appliance |
8495304, | Dec 23 2010 | EMC IP HOLDING COMPANY LLC | Multi source wire deduplication |
8510279, | Mar 15 2012 | EMC International Company | Using read signature command in file system to backup data |
8521691, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Seamless migration between replication technologies |
8521694, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Leveraging array snapshots for immediate continuous data protection |
8543609, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Snapshots in deduplication |
8583885, | Dec 01 2009 | EMC IP HOLDING COMPANY LLC | Energy efficient sync and async replication |
8600945, | Mar 29 2012 | EMC IP HOLDING COMPANY LLC | Continuous data replication |
8601085, | Mar 28 2011 | EMC IP HOLDING COMPANY LLC | Techniques for preferred path determination |
8627012, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for improving cache performance |
8683592, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | Associating network and storage activities for forensic analysis |
8694700, | Sep 29 2010 | EMC IP HOLDING COMPANY LLC | Using I/O track information for continuous push with splitter for storage device |
8706700, | Dec 23 2010 | EMC IP HOLDING COMPANY LLC | Creating consistent snapshots across several storage arrays or file systems |
8712962, | Dec 01 2011 | EMC IP HOLDING COMPANY LLC | Snapshots in de-duplication |
8719497, | Sep 21 2011 | EMC IP HOLDING COMPANY LLC | Using device spoofing to improve recovery time in a continuous data protection environment |
8725691, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Dynamic LUN resizing in a replication environment |
8725692, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Replication of xcopy command |
8726066, | Mar 31 2011 | EMC IP HOLDING COMPANY LLC | Journal based replication with enhance failover |
8738813, | Dec 27 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for round trip synchronous replication using SCSI reads |
8745004, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Reverting an old snapshot on a production volume without a full sweep |
8751828, | Dec 23 2010 | EMC IP HOLDING COMPANY LLC | Sharing encryption-related metadata between multiple layers in a storage I/O stack |
8769336, | Dec 27 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for preventing journal loss on failover in symmetric continuous data protection replication |
8805786, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Replicating selected snapshots from one storage array to another, with minimal data transmission |
8806161, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Mirroring splitter meta data |
8825848, | Mar 20 2012 | EMC IP HOLDING COMPANY LLC | Ordering of event records in an electronic system for forensic analysis |
8832399, | Sep 27 2010 | EMC International Company | Virtualized consistency group using an enhanced splitter |
8850143, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Point in time access in a replication environment with LUN resizing |
8850144, | Mar 29 2012 | EMC IP HOLDING COMPANY LLC | Active replication switch |
8862546, | Jun 30 2011 | EMC IP HOLDING COMPANY LLC | Virtual access roll |
8892835, | Jun 07 2012 | EMC IP HOLDING COMPANY LLC | Insertion of a virtualization layer into a replication environment |
8898112, | Sep 07 2011 | EMC IP HOLDING COMPANY LLC | Write signature command |
8898409, | Jun 27 2012 | EMC International Company | Journal-based replication without journal loss |
8898515, | Jun 28 2012 | EMC International Company | Synchronous replication using multiple data protection appliances across multiple storage arrays |
8898519, | Mar 30 2012 | EMC IP HOLDING COMPANY LLC | Method and apparatus for an asynchronous splitter |
8914595, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Snapshots in deduplication |
8924668, | Dec 23 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for an application- and object-level I/O splitter |
8930500, | Oct 12 2000 | ZEPHYRTEL, INC | Method and system for accelerating receipt of data in a client to client network |
8930947, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for live migration of a virtual machine with dedicated cache |
8935498, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Splitter based hot migration |
8949180, | Jun 28 2012 | EMC International Company | Replicating key-value pairs in a continuous data protection system |
8954673, | Mar 20 2012 | EMC International Company | Using a conditional read request and a hash to determine synchronization of data in a cache at a host with data in storage array |
8954796, | Jun 26 2012 | EMC International Company | Recovery of a logical unit in a consistency group while replicating other logical units in the consistency group |
8959054, | Mar 25 2010 | EMC IP HOLDING COMPANY LLC | Methods and apparatus for optimal journaling for continuous data replication |
8977593, | Dec 26 2009 | EMC IP HOLDING COMPANY LLC | Virtualized CG |
8977826, | Dec 28 2011 | EMC IP HOLDING COMPANY LLC | Extent commands in replication |
8996460, | Mar 14 2013 | EMC IP HOLDING COMPANY LLC | Accessing an image in a continuous data protection using deduplication-based storage |
8996461, | Mar 28 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for replicating the punch command |
8996827, | Dec 27 2011 | EMC IP HOLDING COMPANY LLC | Creating and maintaining clones in continuous data protection |
9003138, | Jun 30 2011 | EMC IP HOLDING COMPANY LLC | Read signature command |
9026696, | Sep 29 2010 | EMC IP HOLDING COMPANY LLC | Using I/O track information for continuous push with splitter for storage device |
9031913, | Dec 28 2011 | EMC IP HOLDING COMPANY LLC | File replication |
9032160, | Dec 29 2011 | EMC IP HOLDING COMPANY LLC | Continuous data replication |
9037818, | Mar 29 2012 | EMC IP HOLDING COMPANY LLC | Active replication switch |
9063994, | Mar 31 2011 | EMC IP HOLDING COMPANY LLC | Networked based replication of distributed volumes |
9069479, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Snapshots in deduplication |
9069709, | Jun 24 2013 | EMC International Company | Dynamic granularity in data replication |
9081754, | Mar 30 2012 | EMC IP HOLDING COMPANY LLC | Method and apparatus for cascaded replication using a multi splitter |
9081842, | Mar 15 2013 | EMC IP HOLDING COMPANY LLC | Synchronous and asymmetric asynchronous active-active-active data access |
9087008, | Jun 24 2013 | EMC International Company | Replicating a volume using snapshots |
9087112, | Jun 24 2013 | EMC International Company | Consistency across snapshot shipping and continuous replication |
9104529, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for copying a cache system |
9110914, | Mar 14 2013 | EMC IP HOLDING COMPANY LLC | Continuous data protection using deduplication-based storage |
9116811, | Jun 30 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9128628, | Mar 13 2013 | EMC IP HOLDING COMPANY LLC | Dynamic replication mode switching |
9128855, | Jun 26 2013 | EMC IP HOLDING COMPANY LLC | Flash cache partitioning |
9134914, | Dec 27 2012 | EMC Corporation | Deduplication |
9135119, | Sep 28 2012 | EMC IP HOLDING COMPANY LLC | System and method for data management |
9135120, | Jun 27 2012 | EMC IP HOLDING COMPANY LLC | Consistency group moving |
9146878, | Jun 25 2013 | EMC IP HOLDING COMPANY LLC | Storage recovery from total cache loss using journal-based replication |
9152339, | Mar 15 2013 | EMC IP HOLDING COMPANY LLC | Synchronization of asymmetric active-active, asynchronously-protected storage |
9152578, | Mar 12 2013 | EMC IP HOLDING COMPANY LLC | Securing data replication, backup and mobility in cloud storage |
9152814, | Mar 15 2010 | EMC International Company | Writing and reading encrypted data using time-based encryption keys |
9158578, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for migrating virtual machines |
9158630, | Dec 19 2013 | EMC IP HOLDING COMPANY LLC | Testing integrity of replicated storage |
9160526, | Dec 23 2010 | EMC IP HOLDING COMPANY LLC | Sharing encryption-related metadata between a host and an external intermediate device |
9177670, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Method and apparatus for flash cache management |
9189339, | Mar 28 2014 | EMC IP HOLDING COMPANY LLC | Replication of a virtual distributed volume with virtual machine granualarity |
9189341, | Mar 30 2012 | EMC IP HOLDING COMPANY LLC | Method and apparatus for multi-copy replication using a multi-splitter |
9201736, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Methods and apparatus for recovery of complex assets in distributed information processing systems |
9223659, | Jun 28 2012 | EMC International Company | Generating and accessing a virtual volume snapshot in a continuous data protection system |
9225529, | Dec 10 2013 | EMC IP HOLDING COMPANY LLC | Encrypted virtual machines in a cloud |
9235481, | Dec 29 2011 | EMC IP HOLDING COMPANY LLC | Continuous data replication |
9235524, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for improving cache performance |
9235632, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Synchronization of replication |
9244997, | Mar 15 2013 | EMC IP HOLDING COMPANY LLC | Asymmetric active-active access of asynchronously-protected data storage |
9256605, | Aug 03 2011 | EMC International Company | Reading and writing to an unexposed device |
9274718, | Jun 20 2014 | EMC IP HOLDING COMPANY LLC | Migration in replication system |
9275063, | Sep 30 2011 | EMC IP HOLDING COMPANY LLC | Replication optimizated IO |
9286052, | Sep 15 2011 | EMC International Company | Upgrading software on a pair of nodes in a clustered environment |
9305009, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Synchronous replication of virtualized storage processors |
9323750, | Sep 29 2010 | EMC IP HOLDING COMPANY LLC | Storage array snapshots for logged access replication in a continuous data protection system |
9330155, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Unified management of sync and async replication for block and file objects |
9336094, | Sep 13 2012 | EMC International Company | Scaleout replication of an application |
9336230, | Dec 28 2011 | EMC Corporation | File replication |
9367260, | Dec 13 2013 | EMC IP HOLDING COMPANY LLC | Dynamic replication system |
9378096, | Jun 30 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9378219, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Metro-cluster based on synchronous replication of virtualized storage processors |
9378261, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Unified synchronous replication for block and file objects |
9383937, | Mar 14 2013 | EMC IP HOLDING COMPANY LLC | Journal tiering in a continuous data protection system using deduplication-based storage |
9389800, | Mar 27 2014 | EMC IP HOLDING COMPANY LLC | Synthesizing virtual machine disk backups |
9405481, | Dec 17 2014 | EMC IP HOLDING COMPANY LLC | Replicating using volume multiplexing with consistency group file |
9405684, | Sep 28 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9405765, | Dec 17 2013 | EMC IP HOLDING COMPANY LLC | Replication of virtual machines |
9411535, | Mar 27 2015 | EMC IP HOLDING COMPANY LLC | Accessing multiple virtual devices |
9459804, | Mar 29 2012 | EMC IP HOLDING COMPANY LLC | Active replication switch |
9460028, | Dec 27 2012 | EMC IP HOLDING COMPANY LLC | Non-disruptive and minimally disruptive data migration in active-active clusters |
9471579, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Replicating selected snapshots from one storage array to another, with minimal data transmission |
9477407, | Jun 28 2013 | EMC IP HOLDING COMPANY LLC | Intelligent migration of a virtual storage unit to another data storage system |
9501542, | Mar 11 2008 | EMC Corporation | Methods and apparatus for volume synchronization |
9507732, | Sep 28 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9507845, | Mar 27 2014 | EMC IP HOLDING COMPANY LLC | Virtual splitter |
9514138, | Mar 15 2012 | EMC International Company | Using read signature command in file system to backup data |
9524218, | Sep 09 2013 | EMC IP HOLDING COMPANY LLC | Leverage fast VP extent-level statistics within CDP environments |
9529885, | Sep 29 2014 | EMC IP HOLDING COMPANY LLC | Maintaining consistent point-in-time in asynchronous replication during virtual machine relocation |
9535800, | Sep 30 2014 | EMC IP HOLDING COMPANY LLC | Concurrent data recovery and input/output processing |
9535801, | Jun 30 2011 | EMC IP HOLDING COMPANY LLC | Xcopy in journal based replication |
9547459, | Jun 27 2013 | EMC IP HOLDING COMPANY LLC | Techniques for data relocation based on access patterns |
9547591, | Sep 28 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9552405, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Methods and apparatus for recovery of complex assets in distributed information processing systems |
9557921, | Mar 26 2015 | EMC IP HOLDING COMPANY LLC | Virtual volume converter |
9557925, | Dec 29 2014 | EMC IP HOLDING COMPANY LLC | Thin replication |
9563517, | Dec 30 2013 | EMC IP HOLDING COMPANY LLC | Cloud snapshots |
9563684, | Dec 31 2013 | EMC IP HOLDING COMPANY LLC | Replication cookie |
9575851, | Jun 27 2012 | EMC IP HOLDING COMPANY LLC | Volume hot migration |
9575857, | Jun 27 2012 | EMC IP HOLDING COMPANY LLC | Active/active replication |
9575894, | Mar 27 2015 | EMC IP HOLDING COMPANY LLC | Application aware cache coherency |
9582382, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Snapshot hardening |
9588703, | Mar 28 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for replicating the punch command |
9588847, | Mar 25 2014 | EMC IP HOLDING COMPANY LLC | Recovering corrupt virtual machine disks |
9594822, | Mar 13 2013 | EMC IP HOLDING COMPANY LLC | Method and apparatus for bandwidth management in a metro cluster environment |
9600377, | Dec 03 2014 | EMC IP HOLDING COMPANY LLC | Providing data protection using point-in-time images from multiple types of storage devices |
9619543, | Jun 23 2014 | EMC IP HOLDING COMPANY LLC; EMC Corporation | Replicating in virtual desktop infrastructure |
9632881, | Mar 24 2015 | EMC IP HOLDING COMPANY LLC | Replication of a virtual distributed volume |
9665305, | Jun 26 2015 | EMC IP HOLDING COMPANY LLC | Tiering data between two deduplication devices |
9710177, | Dec 27 2011 | EMC IP HOLDING COMPANY LLC | Creating and maintaining clones in continuous data protection |
9720618, | Dec 16 2013 | EMC IP HOLDING COMPANY LLC | Maintaining backup snapshots using continuous replication from multiple sources |
9722788, | Jun 29 2015 | EMC IP HOLDING COMPANY LLC | Rekeying encrypted virtual machines in a cloud |
9727429, | Mar 31 2015 | EMC IP HOLDING COMPANY LLC | Method and system for immediate recovery of replicated virtual machines |
9733969, | Jun 30 2015 | EMC IP HOLDING COMPANY LLC | Method and system for malware detection in virtual machines |
9737111, | Mar 15 2013 | Removable shoe insert for corrective sizing | |
9740572, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Replication of xcopy command |
9740573, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Dynamic LUN resizing in a replication environment |
9740880, | Dec 10 2013 | EMC Corporation | Encrypted virtual machines in a cloud |
9749300, | Mar 30 2015 | EMC IP HOLDING COMPANY LLC | Method and system for immediate recovery of virtual machines encrypted in the cloud |
9772789, | Mar 25 2014 | EMC IP HOLDING COMPANY LLC | Alignment fixing on a data protection system during continuous data replication to deduplicated storage |
9798472, | Sep 29 2015 | EMC Corporation | Extent level cache destaging |
9798490, | Sep 30 2011 | EMC IP HOLDING COMPANY LLC | Replication optimizated IO |
9804934, | Dec 30 2014 | EMC IP HOLDING COMPANY LLC | Production recovery using a point in time snapshot |
9811431, | Mar 31 2011 | EMC IP HOLDING COMPANY LLC | Networked based replication of distributed volumes |
9823865, | Jun 30 2015 | EMC IP HOLDING COMPANY LLC | Replication based security |
9823973, | Dec 22 2014 | EMC IP HOLDING COMPANY LLC | Creating consistent snapshots in a virtualized environment |
9832261, | Sep 30 2014 | EMC IP HOLDING COMPANY LLC | Cloud consistency technology |
9846698, | Dec 16 2013 | EMC IP HOLDING COMPANY LLC | Maintaining point-in-time granularity for backup snapshots |
20020129168, | |||
20030048842, | |||
20030061537, | |||
20030110278, | |||
20030145317, | |||
20030196147, | |||
20040205092, | |||
20040250032, | |||
20040254964, | |||
20050015663, | |||
20050028022, | |||
20050049924, | |||
20050172092, | |||
20050273655, | |||
20060031647, | |||
20060047996, | |||
20060064416, | |||
20060107007, | |||
20060117211, | |||
20060161810, | |||
20060179343, | |||
20060195670, | |||
20070055833, | |||
20070180304, | |||
20070198602, | |||
20070198791, | |||
20150310044, | |||
20160320986, | |||
20170249222, | |||
EP1154356, | |||
WO45581, |
Date | Maintenance Fee Events |
Dec 16 2021 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Jul 10 2021 | 4 years fee payment window open |
Jan 10 2022 | 6 months grace period start (w surcharge) |
Jul 10 2022 | patent expiry (for year 4) |
Jul 10 2024 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jul 10 2025 | 8 years fee payment window open |
Jan 10 2026 | 6 months grace period start (w surcharge) |
Jul 10 2026 | patent expiry (for year 8) |
Jul 10 2028 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jul 10 2029 | 12 years fee payment window open |
Jan 10 2030 | 6 months grace period start (w surcharge) |
Jul 10 2030 | patent expiry (for year 12) |
Jul 10 2032 | 2 years to revive unintentionally abandoned end. (for year 12) |