In one aspect, a method includes moving metadata from a memory of a first deduplication device to a disk of the first deduplication device, moving the data related to the metadata or portions of the data to a second deduplication device, deleting the data from the first deduplication device and adding a pointer to the metadata at the first deduplication device indicating where the data is in the second deduplication device.
|
1. A method comprising:
moving metadata associated with data from a memory of a first deduplication device to a disk of the first deduplication device if the data has not been accessed in a predetermined amount of time, wherein the memory comprises the data, wherein data used more than once in the first deduplication is stored in one location on the first deduplication device;
not moving the metadata associated with the data from the memory of the first deduplication device to the disk of the first deduplication device if the data has been accessed in the predetermined amount of time;
moving the data related to the metadata or portions of the data from the memory to a second deduplication device after the metadata has moved to the disk, wherein data used more than once in the second deduplication is stored in one location on the second deduplication device;
deleting the data from the first deduplication device after the data has moved to the second deduplication device; and
adding a pointer to the metadata at the disk, after the metadata has moved to the disk, indicating where the data is located in the second deduplication device.
15. An article comprising:
a non-transitory computer-readable medium that stores computer-executable instructions, the instructions causing a machine to:
move metadata associated with data from a memory of a first deduplication device to a disk of the first deduplication device if the data has not been accessed in a predetermined amount of time, wherein the memory comprises the data;
not move the metadata associated with the data from the memory of the first deduplication device to the disk of the first deduplication device if the data has been accessed in the predetermined amount of time, wherein the memory comprises the data;
move the data related to the metadata or portions of the data from the memory to a second deduplication device after the metadata has moved to the disk, wherein data used more than once in the second deduplication is stored in one location on the second deduplication device;
delete the data from the first deduplication device after the data has moved to the second deduplication device; and
add a pointer to the metadata, after the metadata has moved to the disk, indicating where the data is located in the second deduplication device.
8. An apparatus, comprising:
electronic hardware circuitry configured to:
move metadata associated with data from a memory of a first deduplication device to a disk of the first deduplication device if the data has not been accessed in a predetermined amount of time, wherein the memory comprises the data, wherein data used more than once in the first deduplication is stored in one location on the first deduplication device;
not move the metadata associated with the data from the memory of the first deduplication device to the disk of the first deduplication device if the data has been accessed in the predetermined amount of time;
move the data related to the metadata or portions of the data from the memory to a second deduplication device after the metadata has moved to the disk, wherein data used more than once in the second deduplication is stored in one location on the second deduplication device;
delete the data from the first deduplication device after the data has moved to the second deduplication device; and
add a pointer to the metadata, after the metadata has moved to the disk, indicating where the data is located in the second deduplication device.
2. The method of
receiving at the first deduplication device a request to read the data from a host;
discovering the pointer to the second deduplication device;
retrieving the data, using the pointer, from the second duplication device;
sending the data to the first deduplication device;
writing the data to the first deduplication device; and
sending the data to the host.
3. The method of
4. The method of
6. The method of
receiving a write request to a location of the data;
writing data from the write request to the first deduplication device; and
changing the pointer to point to the data from the write request in the first deduplication device.
7. The method of
9. The apparatus of
receive at the first deduplication device a request to read the data from a host;
discover the pointer to the second deduplication device;
retrieve the data, using the pointer, from the second duplication device;
send the data to the first deduplication device;
write the data to the first deduplication device; and
send the data to the host.
10. The apparatus of
11. The apparatus of
12. The apparatus of
13. The apparatus of
receive a write request to a location of the data;
write data from the write request to the first deduplication device; and
change the pointer to point to the data from the write request in the first deduplication device.
14. The apparatus of
16. The article of
receive at the first deduplication device a request to read the data from a host;
discover the pointer to the second deduplication device;
retrieve the data, using the pointer, from the second duplication device;
send the data to the first deduplication device;
write the data to the first deduplication device; and
send the data to the host.
17. The article of
18. The article of
wherein the instructions causing the machine to add the pointer to the metadata comprises instructions causing the machine to add a hash value.
19. The article of
receive a write request to a location of the data;
write data from the write request to the first deduplication device; and
change the pointer to point to the data from the write request in the first deduplication device.
20. The article of
|
Storage systems in general, and block based storage systems specifically, are a key element in modern data centers and computing infrastructure. These systems are designed to store and retrieve large amounts of data, by providing data block address and data block content—for storing a block of data—and by providing a data block address for retrieval of the data block content that is stored at the specified address.
Storage solutions are typically partitioned into categories based on a use case and application within a computing infrastructure, and a key distinction exists between primary storage solutions and archiving storage solutions. Primary storage is typically used as the main storage pool for computing applications during application run-time. As such, the performance of primary storage systems is very often a key challenge and a major potential bottleneck in overall application performance, since storage and retrieval of data consumes time and delays the completion of application processing. Storage systems designed for archiving applications are much less sensitive to performance constraints, as they are not part of the run-time application processing.
In general computer systems grow over their lifetime and the data under management tends to grow over the system lifetime. Growth can be exponential, and in both primary and archiving storage systems, exponential capacity growth typical in modern computing environment presents a major challenge as it results in increased cost, space, and power consumption of the storage systems required to support ever increasing amounts of information.
Existing storage solutions, and especially primary storage solutions, rely on address-based mapping of data, as well as address-based functionality of the storage system's internal algorithms. This is only natural since the computing applications always rely on address-based mapping and identification of data they store and retrieve. However, a completely different scheme in which data, internally within the storage system, is mapped and managed based on its content instead of its address has many substantial advantages. For example, it improves storage capacity efficiency since any duplicate block data will only occupy actual capacity of a single instance of that block. As another example, it improves performance since duplicate block writes do not need to be executed internally in the storage system. Existing storage systems, either primary storage systems or archiving storage systems are incapable of supporting the combination of content based storage—with its numerous advantages—and ultra-high performance. This is a result of the fact that the implementation of content based storage scheme faces several challenges:
(a) intensive computational load which is not easily distributable or breakable into smaller tasks,
(b) an inherent need to break large blocks into smaller block sizes in order to achieve content addressing at fine granularity. This block fragmentation dramatically degrades the performance of existing storage solutions,
(c) inability to maintain sequential location of data blocks within the storage systems, since mapping is not address based any more, and such inability causes dramatic performance degradation with traditional spinning disk systems,
(d) the algorithmic and architectural difficulty in distributing the tasks associated with content based mapping over a large number of processing and storage elements while maintaining single content-addressing space over the full capacity range of the storage system.
A number of issues arise with respect to such devices, and it is necessary to consider such issues as performance, lifetime and resilience to failure of individual devices, overall speed of response and the like.
Such devices may be used in highly demanding circumstances where failure to process data correctly can be extremely serious, or where large scales are involved, and where the system has to be able to cope with sudden surges in demand.
In one aspect, a method includes moving metadata from a memory of a first deduplication device to a disk of the first deduplication device, moving the data related to the metadata or portions of the data to a second deduplication device, deleting the data from the first deduplication device and adding a pointer to the metadata at the first deduplication device indicating where the data is in the second deduplication device.
In another aspect, an apparatus includes electronic hardware circuitry configured to move metadata from a memory of a first deduplication device to a disk of the first deduplication device; move the data related to the metadata or portions of the data to a second deduplication device; delete the data from the first deduplication device; and add a pointer to the metadata at the first deduplication device indicating where the data is in the second deduplication device. The circuitry comprises at least one of a processor, a memory, a programmable logic device or a logic gate.
In a further aspect, an article includes a non-transitory computer-readable medium that stores computer-executable instructions. The instructions cause a machine to move metadata from a memory of a first deduplication device to a disk of the first deduplication device; move the data related to the metadata or portions of the data to a second deduplication device; delete the data from the first deduplication device; and add a pointer to the metadata at the first deduplication device indicating where the data is in the second deduplication device.
In a Content Addressable Storage (CAS) array, data is stored in blocks, for example of 4 KB, where each block has a unique large hash signature, for example of 20 bytes, saved on Flash memory.
The examples described herein include a networked memory system. The networked memory system includes multiple memory storage units arranged for content addressable storage of data. The data is transferred to and from the storage units using separate data and control planes. Hashing is used for the content addressing, and the hashing produces evenly distributed results over the allowed input range. The hashing defines the physical addresses so that data storage makes even use of the system resources.
A relatively small granularity may be used, for example with a page size of 4 KB, although smaller or larger block sizes may be selected at the discretion of the skilled person. This enables the device to detach the incoming user access pattern from the internal access pattern. That is to say the incoming user access pattern may be larger than the 4 KB or other system-determined page size and may thus be converted to a plurality of write operations within the system, each one separately hashed and separately stored.
Content addressable data storage can be used to ensure that data appearing twice is stored at the same location. Hence unnecessary duplicate write operations can be identified and avoided. Such a feature may be included in the present system as data deduplication. As well as making the system more efficient overall, it also increases the lifetime of those storage units that are limited by the number of write/erase operations.
The separation of Control and Data may enable a substantially unlimited level of scalability, since control operations can be split over any number of processing elements, and data operations can be split over any number of data storage elements. This allows scalability in both capacity and performance, and may thus permit an operation to be effectively balanced between the different modules and nodes.
The separation may also help to speed the operation of the system. That is to say it may speed up Writes and Reads. Such may be due to:
(a) Parallel operation of certain Control and Data actions over multiple Nodes/Modules
(b) Use of optimal internal communication/networking technologies per the type of operation (Control or Data), designed to minimize the latency (delay) and maximize the throughput of each type of operation.
Also, separation of control and data paths may allow each Control or Data information unit to travel within the system between Nodes or Modules in the optimal way, meaning only to where it is needed and if/when it is needed. The set of optimal where and when coordinates is not the same for control and data units, and hence the separation of paths ensures the optimization of such data and control movements, in a way which is not otherwise possible. The separation is important in keeping the workloads and internal communications at the minimum necessary, and may translate into increased optimization of performance.
De-duplication of data, meaning ensuring that the same data is not stored twice in different places, is an inherent effect of using Content-Based mapping of data to D-Modules and within D-Modules.
Scalability is inherent to the architecture. Nothing in the architecture limits the number of the different R, C, D, and H modules which are described further herein. Hence any number of such modules can be assembled. The more modules added, the higher the performance of the system becomes and the larger the capacity it can handle. Hence scalability of performance and capacity is achieved.
The principles and operation of an apparatus and method according to the present invention may be better understood with reference to the drawings and accompanying description.
Reference is now made to
The control modules 14 may control execution of read and write commands. The data modules 16 are connected to the storage devices and, under control of a respective control module, pass data to or from the storage devices. Both the C and D modules may retain extracts of the data stored in the storage device, and the extracts may be used for the content addressing. Typically the extracts may be computed by cryptographic hashing of the data, as will be discussed in greater detail below, and hash modules (
Routing modules 18 may terminate storage and retrieval operations and distribute command parts of any operations to control modules that are explicitly selected for the operation in such a way as to retain balanced usage within the system 10.
The routing modules may use hash values, calculated from data associated with the operations, to select the control module for the distribution. More particularly, selection of the control module may use hash values, but typically relies on the user address and not on the content (hash). The hash value is, however, typically used for selecting the Data (D) module, and for setting the physical location for data storage within a D module.
The storage devices may be solid state random access storage devices, as opposed to spinning disk devices; however disk devices may be used instead or in addition.
A deduplication feature may be provided. The routing modules and/or data modules may compare the extracts or hash values of write data with hash values of already stored data, and where a match is found, simply point to the matched data and avoid rewriting.
The modules are combined into nodes 20 on the network, and the nodes are connected over the network by a switch 22.
The use of content addressing with multiple data modules selected on the basis of the content hashing, and a finely-grained mapping of user addresses to Control Modules allow for a scalable distributed architecture.
A glossary is now given of terms used in the following description:
X-PAGE—A predetermined-size aligned chunk as the base unit for memory and disk operations. Throughout the present description the X-Page size is referred to as having 4 KB, however other smaller or larger values can be used as well and nothing in the design is limited to a specific value.
LUN or LOGICAL UNIT NUMBER is a common name in the industry for designating a volume of data, or a group of data blocks being named with the LUN. Each data block is referred to, by the external user of the storage system, according to its LUN, and its address within this LUN
LOGICAL X-PAGE ADDRESS—Logical address of an X-Page. The address contains a LUN identifier as well as the offset of the X-Page within the LUN.
LOGICAL BLOCK—512 bytes (sector) aligned chunk, which is the SCSI base unit for disk operations.
LOGICAL BLOCK ADDRESS—Logical address of a Logical Block. The logical block address contains a LUN identifier as well as the offset of the logical block within the LUN.
SUB-LUN—Division of a LUN to smaller logical areas, to balance the load between C modules. Each such small logical area is called a sub-LUN.
SUB-LUN UNIT SIZE—The fixed size of a sub-LUN. X-Page Data—Specific sequence of user data values that resides in an X-Page. Each such X-Page Data is uniquely represented in the system by its hash digest.
D PRIMARY—The D module responsible for storing an X-Page's Data
D BACKUP—The D module responsible for storing a backup for an X-Page Data. The backup is stored in a non-volatile way (NVRAM or UPS protected).
ACRONYMS:
LXA—Logical X-Page Address.
LB—Logical Block.
LBA—Logical Block Address.
AUS—Atomic Unit Size.
SL—Sub-LUN.
SLUS—Sub-LUN Unit Size.
MBE—Management Back End.
The examples described herein to a block-level storage system, offering basic and advanced storage functionality. The design may be based on a distributed architecture, where computational, Storage Area Networking (SAN), and storage elements are distributed over multiple physical Nodes, with all such Nodes being inter-connected over an internal network through a switch device. The distributed architecture enables the scaling of the system's capabilities in multiple aspects, including overall storage capacity, performance characteristics in bandwidth and I/O operations per second (IOPS), computational resources, internal and external networking bandwidth, and other. While being based on a distributed architecture, the system presents, externally, a unified storage system entity with scalable capabilities.
The system's architecture and internal algorithms implementing the basic and advanced storage functions are optimized for improved utilization of the capabilities of random-access memory/storage media, as opposed to contrast with mechanical-magnetic spinning disk storage media. The optimizations are implemented in the design itself, and may, for example, include the ability to break incoming writes into smaller blocks and distribute the operation over different Nodes. Such an adaptation is particularly suitable for random access memory/storage media but is less suitable in a spinning-disk environment, as it would degrade performance to extremely low levels. The adaptation includes the content/hash based mapping of data distributes the data over different D Nodes in general and within D Nodes over different SSD devices. Again, such a scheme is more suitable for random access memory/storage media than for a spinning-disk media because such spread of data blocks would result in very poor performance in the spinning disk case. That is to say, the described elements of the present architecture are designed to work well with random access media, and achieve benefits in performance, scalability, and functionality such as inline deduplication. Such random-access memory media can be based on any or a combination of flash memory, DRAM, phase change memory, or other memory technology, whether persistent or non-persistent, and is typically characterized by random seek/access times and random read/write speeds substantially higher than those exhibited by spinning disk media. The system's internal data block mapping, the algorithms implementing advanced storage functions, and the algorithms for protecting data stored in the system are designed to provide storage performance and advanced storage functionality at substantially higher performance, speed, and flexibility than those available with alternative storage systems.
Data mapping within the system is designed not only to improve performance, but also to improve the life span and reliability of the electronic memory media, in cases where the memory technology used has limitations on write/erase cycles, as is the case with flash memory. Lifetime maximization may be achieved by avoiding unnecessary write operations as will be explained in greater detail below. For the purpose of further performance optimization, life span maximization, and cost optimization, the system may employ more than a single type of memory technology, including a mix of more than one Flash technology (e.g., single level cell—SLC flash and multilevel cell—MLC flash), and a mix of Flash and DRAM technologies. The data mapping optimizes performance and life span by taking advantage of the different access speeds and different write/erase cycle limitations of the various memory technologies.
The core method for mapping blocks of data internally within the system is based on Content Addressing, and is implemented through a distributed Content Addressable Storage (CAS) algorithm.
This scheme maps blocks of data internally according to their content, resulting in mapping of identical block to the same unique internal location. The distributed CAS algorithm allows for scaling of the CAS domain as overall system capacity grows, effectively utilizing and balancing the available computational and storage elements in order to improve overall system performance at any scale and with any number of computational and storage elements.
The system supports advanced In-line block level deduplication, which may improve performance and save capacity.
Elements of the system's functionality are: Write (store) data block at a specified user address; Trim data block at a specified user address; Read data block from a specified user address; and In-line block level deduplication.
The following features may be provided: (1) A distributed CAS based storage optimized for electronic random-access storage media; The optimization includes utilizing storage algorithms, mainly the content-based uniformly-distributed mapping of data, that inherently spread data in a random way across all storage devices. Such randomization of storage locations within the system while maintaining a very high level of performance is preferably achievable with storage media with a high random access speed; (2) A distributed storage architecture with separate control and data planes; Data mapping that maximizes write-endurance of storage media; System scalability; (3) System resiliency to fault and/or failure of any of its components; (4) Use of multi-technology media to maximize write-endurance of storage media; and (5) In-line deduplication in ultrahigh performance storage using electronic random-access storage media.
The examples described herein implement block storage in a distributed and scalable architecture, efficiently aggregating performance from a large number of ultra-fast storage media elements (SSDs or other), preferably with no performance bottlenecks, while providing in-line, highly granular block-level deduplication with no or little performance degradation.
One challenge is to avoid performance bottlenecks and allow performance scalability that is independent of user data access patterns.
The examples described herein may overcome the scalability challenge by providing data flow (Write, Read) that is distributed among an arbitrary and scalable number of physical and logical nodes. The distribution is implemented by (a) separating the control and data paths (the “C” and “D” modules), (b) maintaining optimal load balancing between all Data modules, based on the content of the blocks (through the CAS/hashing mechanisms), hence ensuring always balanced load sharing regardless of user access patterns, (c) maintaining optimal load balancing between all Control modules, based on the user address of the blocks at fine granularity, hence ensuring always balanced load sharing regardless of user access patterns, and (d) performing all internal data path operations using small granularity block size, hence detaching the incoming user access pattern from the internal access pattern, since the user pattern is generally larger than the block size.
A second challenge is to support inline, highly granular block level deduplication without degrading storage (read/write speed) performance. The result should be scalable in both capacity—which is deduplicated over the full capacity space—and performance.
The solution involves distributing computation-intensive tasks, such as calculating cryptographic hash values, among an arbitrary number of nodes. In addition, CAS metadata and its access may be distributed among an arbitrary number of nodes. Furthermore, data flow algorithms may partition read/write operations in an optimally-balanced way, over an arbitrary and scalable number of Nodes, while guaranteeing consistency and inline deduplication effect over the complete storage space.
In detaching the data from the incoming pattern, the R-Module breaks up any incoming block which is larger than the granularity size across sub-LUNs, sending the relevant parts to the appropriate C-Modules. Each C-module is predefined to handle a range or set of Sub-LUN logical addresses. The C-Module breaks up the block it receives for distribution to D-Modules, at a pre-determined granularity, which is the granularity for which a Hash is now calculated. Hence the end result is that a request to write a certain block (for example of size 64 KB) ends up being broken up into for example 16 internal writes, each write comprising a 4 KB block.
The specific numbers for granularity can be set based on various design tradeoffs, and the specific number used herein of 4 KB is merely an example. The broken down blocks are then distributed to the D modules in accordance with the corresponding hash values.
A further challenge is to address flash-based SSD write/erase cycle limitations, in which the devices have a lifetime dependent on the number of write/erase cycles.
The solution may involve Inline deduplication to avoid writing in all cases of duplicate data blocks. Secondly, content (hash) based mapping to different data modules and SSDs results in optimal wear-leveling, ensuring equal spread of write operations to all data modules and SSDs independently of the user data/address access patterns.
In the following a system is considered from a functional point of view. As described above with respect to
Reference is now made to
A function of the R Module 202 is to terminate SAN Read/Write commands and route them to appropriate C and D Modules for execution by these Modules. By doing so, the R Module can distribute workload over multiple C and D Modules, and at the same time create complete separation of the Control and Data planes, that is to say provide separate control and data paths.
A function of the C Module 204 is to control the execution of a Read/Write command, as well as other storage functions implemented by the system. It may maintain and manage key metadata elements.
A function of the D Module 206 is to perform the actual Read/Write operation by accessing the storage devices 208 (designated SSDs) attached to it. The D module 206 may maintain metadata related with the physical location of data blocks.
A function of the H Module is to calculate the Hash function value for a given block of data.
Reference is now made to
In
All Nodes include a switch interface 308, to allow interconnecting with a switch in a multi-Node system configuration. A Node that contains a SAN function includes at least one SAN Interface module 310 and at least one R Module. A Node that contains a Store function includes at least one SSD Driver Module 312 and at least one D Module. Hence, Compute+SAN and Compute+SAN+STORE Nodes contain a SAN Interface, to interface with the external SAN. The interface may typically use a SCSI-based protocol running on any of a number of interfaces including Fiber Channel, Ethernet, and others, through which Read/Write and other storage function commands are being sent to the system. Compute+Store and Compute+SAN+Store Nodes contain an SSD driver 312 to interface with SSDs 208 attached to that specific Node, where data is stored and accessed.
Reference is now made to
The interconnections between each Node and the Switch may include redundancy, so as to achieve high system availability with no single point of failure. In such a case, each Node may contain two or more Switch Interface modules 406, and the Switch may contain two or more ports per physical Node.
As an example
A four node system configuration is shown in
A system that is built from multiple physical Nodes can inherently support a high availability construction, where there is no single point of failure. This means that any Node or sub-Node failure can be compensated for by redundant Nodes, having a complete copy of the system's meta-data, and a complete redundant copy of stored data (or parity information allowing recovery of stored data). The distributed and flexible architecture allows for seamless support of failure conditions by simply directing actions to alternate Nodes.
The R module is responsible for: routing SCSI I/O requests to the C modules, guarantee execution and return the result; and balancing the work load between the C modules for the requests it is routing.
An A→C table indicates which C module is responsible for each logical X-page address (LXA). Each C module is responsible for a list of Sub LUNs (SLs).
The R module receives requests for I/Os from the SAN INTERFACE routes them to the designated C modules and returns the result to the SAN INTERFACE.
If an I/O operation spans across multiple SLs, and perhaps multiple C modules, then the R module has the responsibility of breaking the big I/O operation into multiple smaller independent operations according to the sub LUN unit size (SLUS). Since the atomic unit size (AUS) is never larger than the SLUS, as explained in greater detail below, each such I/O is treated as an independent operation throughout the system. The results may then be aggregated before returning to the SAN INTERFACE.
The R module is responsible for maintaining an up-to-date A→C table coordinated with the MBE. The A→C table is expected to balance the range of all possible LXAs between the available C modules.
For write operations, the R module instructs the calculation of the hash digest for each X-Page by requesting such calculation from a Hash calculation module.
The C module is responsible for: receiving an I/O request from an R module on a certain SL, guaranteeing its atomic execution and returning the result; communicating with D modules to execute the I/O requests; monitoring the disk content of its SLs' logical space by associating each LXA with its hash digest; and balancing the work load between the D modules for the SLs it is maintaining.
An H→D table maps each range of hash digests to the corresponding D module responsible for this range.
An A→H table maps each LXA that belongs to the SLs C is responsible for, to the hash digest representing the X-Page Data that currently resides in this address.
The C module receives I/O requests from R modules, distributes the work to the D modules, aggregates the results and guarantees an atomic operation. The result is returned to the R module.
The C module maintains an up-to-date H→D table coordinated with the MBE. The table is expected to balance the range of all possible hash digests between the available D modules.
The C module maintains an A→H table in a persistent way. The C module may initiate 110 requests to D modules in order to save table pages to disk, and read them from disk. To avoid frequent disk operations, a Journal of the latest table operations may be maintained.
Data is balanced between the C modules based on the logical address, at the granularity of sub-LUNs.
The D module is responsible for: maintaining a set of LUNs which are attached locally and performing all I/O operations on these LUN; managing the physical layout of the attached LUNs; managing the mapping between X-Page Data hash digests and their physical location in a persistent way; managing deduplication of X-Page Data in a persistent way; and receiving disk I/O requests from C modules, perform them and returning a result.
The D module is also responsible for, for each write operation, backing up the X-Page Data in the designated D backup module and performing read-modify operations for writes that are smaller than X-Page size (This process also involves computing a hash digest for these X-Pages).
The D module is further responsible for maintaining an up-to-date H→(D, Dbackup) table coordinated with the MBE. The H→(D, Dbackup) table is expected to balance the range of all possible hash digests between the available D modules.
The D module does not communicate directly with R modules. The only interaction with R modules involves RDMA read/write operations of X-Page Data.
Balancing between the D modules is based on hashing of the content.
The D module makes use of a hash digest metadata table. The hash digest metadata table maps each in use hash digest, that represents actual X-Page Data, to its meta data information including its physical page on the storage media (SSD), its memory copy (if exists), a mapping to any backup memory copy and a reference count for the purpose of deduplication.
A further structure used is the H→(D, Dbackup) table. The H→(D, Dbackup) table maps each range of hash digests to the corresponding D module responsible for the range as well as the Dbackup module responsible for the range.
The D modules allocate a physical page for each X-Page. The D modules also manage the memory for the physical storage. They allocate memory pages for read/write operations and perform background destaging from memory to storage media when necessary, for example, when running low on memory.
The D modules manage a separate nonvolatile memory pool (NVRAM or UPS protected) for X-Page Data backup purposes. The backup holds X-Pages that are held in memory of the D primary and have not yet been destaged. When re-balancing between D modules occur (due to a D module failure for example), the D module may communicate with other D modules in order to create new backup copies or move a primary ownership as required.
The D modules allow deduplication per X-Page Data by maintaining a persistent reference count that guarantees only one copy per X-Page Data. The D modules manage the hash digest metadata table in a persistent way. The table is coordinated with the physical layout for physical pages allocation, with the memory pointer, memory backup pointer and deduplication reference count.
The D modules receive I/O requests from C modules, perform the requests while supporting deduplication and return the result. The D modules may perform RDMA read/write operations on memory that resides in other modules, such as R modules as mentioned above, as part of the I/O operation.
When a write operation smaller than the size of an X-Page is received, the D module may read the entire X-Page to memory and perform partial X-Page modification on that memory. In this case race conditions may occur, for example when two small writes to the same X-Page occur in parallel, and the D module may be required to compute the hash digest of the resulting X-Page. This is discussed in greater detail below.
The H-Module calculates the Hash function of a given block of data, effectively mapping an input value to a unique output value. The Hash function may be based on standards based hash functions such as SHA-1 and MDS, or based on a proprietary function. The hash function is selected to generate a uniformly distributed output over the range of potential input values.
The H modules usually share nodes with an R module but more generally, the H modules can reside in certain nodes, in all nodes, together with R modules, or together with C or D modules.
The following discussion provides high level I/O flows for read, write and trim.
Throughout these flows, unless noted otherwise, control commands are passed between modules using standard RPC messaging, while data “pull” operations may use RDMA read. Data push (as well as Journal) operations may use RDMA write.
The read flow of one X-Page may consist of one R module which receives the read request from the application, one C module in charge of the address requested and one D module which holds the X-Page to be read. Larger, or unaligned, requests may span several X-Pages and thus may involve several D modules. These requests may also span several SLs, in which case they may involve several C modules as well.
Reference is now made to
The C module, when receiving the request, consults the A→H component, from which it obtains a hash digest representing the X-Page to be read; consults the H→D component to determine which D module holds the X-Page in question; and sends this D module a read request which includes parameters that include a request ID (as received from the R module), the hash digest, a pointer to the buffer to read to, as received from the R module; and an identifier of the R module.
The D module, when receiving the request, reads the data of the requested X-Page from SSD and performs an RDMA write to the requesting R module, specifically to the pointer passed to it by the C module.
Finally the D module returns success or error to the requesting C module.
The C module in turn propagates success or error back to the requesting R module, which may then propagate it further to answer the application.
Reference is now made to
The rest of the R module's treatment is identical to the aligned one X-Page scenario previously described herein.
The C module, when receiving the request divides the logical address space to LXAs. For each LXA the C module consults the A→H component to determine the corresponding hash digest; consults the H→D table to determine which D module is responsible for the current LXA; sends each D module a read command containing all the hashes that the respective D module is responsible for. The parameters of the read command include a request ID (as received from the R module); a list of respective hash-pointer pairs; and the identifier of the R module.
Each D module, when receiving the request, acts per hash-pointer pair in the same manner as described above for one X-Page. Aggregated success or error is then sent to the requesting C module.
The C module aggregates all the results given to it by the D modules and return success or error back to the requesting R module, which may then answer the application.
In the case that a read request spans multiple SLs, the R module splits the request and sends several C modules read requests. Each C module may receive one request per SL. The flow may continue as in the simpler case above, except that now the R module aggregates the responses before it answers the application.
Read requests smaller than 4 KB, as well as requests not aligned to 4 KB, may be dealt with at the R module level. For each such parcel of data, the R module may request to read the encompassing X-Page. Upon successful completion of the read command, the R module may crop the non-relevant sections and return only the requested data to the application.
The write flow of one X-Page may consist of one R module which receives the write request from the application, one C module in charge of the address requested and three D modules: Dtarget which is in charge of the X-Page Data to be written (according to its appropriate hash digest), Dold which was in charge of the X-Page Data this address contained previously (“old” hash digest), and Dbackup in charge of storing a backup copy of the X-Page Data to be written.
Reference is now made to
When an R module receives a write request from the application, the R module allocates a request ID for this operation; translates the LBA to an LXA; computes a hash digest on the data to be written; consults its A→C component to determine which C module is in charge of the current LXA; and sends the designated C module a write command with parameters that include a request ID; an LXA; a hash digest; and a pointer to the buffer containing the data to be written.
The C module, when receiving the request consults its H→D component to understand which D module is in charge of the X-Page to be written (Dtarget); and sends Dtarget a write request with parameters that include the request ID (as received from the R module); the hash digest (as received from the R module); the pointer to the data to write (as received from the R module); and the identifier of the R module.
The D module receiving the write command, Dtarget, may first check if it already holds an X-Page corresponding to this hash. There are two options here:
First, Dtarget does not have the X-Page. In this case Dtarget fetches the data from the R module using RDMA read and stores it in its memory; consults the H→D component to determine which D module is in charge of storing a backup copy of this X-Page (Dbackup); performs an RDMA write of the X-Page Data to the Dbackup backup memory space; and returns success (or failure) to the C module.
Second, Dtarget has the X-Page. In this case Dtarget increases the reference count, returns success (or failure) to the C module.
The C module waits for a response from Dtarget. If a success is returned, the C module updates the A→H table to indicate that the LXA in question should point to the new hash and returns a response to the requesting R module.
If this is not a new entry in the A→H table, the C module asynchronously sends a decrease reference count command to Dold (the D module responsible for the hash digest of the previous X-Page Data). These commands may be aggregated at the C module and sent to the D modules in batches.
The R module may answer the application once it receives a response from the C module.
Reference is now made to
In the case that the write request spans a range of addresses which include more than one X-Page but only one SL, the R module sends the designated C module a write command with parameters that include a request ID; a first LXA; a size of the requested write in LXAs-n; and HBIG which is a unique identifier of the entire chunk of data to be written. HBIG may be a computed hash digest and thus equal for two identical chunks of data.
Additional parameters sent with the write command are n pointers that point to the buffers which hold the data to be written.
The rest of the R module treatment is the same as for the aligned one X-Page scenario.
The C module, when receiving the request, consults its H→D component to understand which D module is in charge of HBIG (Dtarget) and generates a hash digest per pointer by replacing one byte of HBIG with the offset of that pointer. It is noted that this byte must not collide with the bytes used by the H→D table distribution.
It may send Dtarget a write request with the parameters that include the request ID (as received from the R module); a list of respective hash-pointer pairs; and the Identifier of the R module.
The D module, when receiving the request, acts per hash-pointer pair in the same manner as described above for one X-Page. Aggregated success or error is then sent to the requesting C module.
The C module waits for a response from Dtarget. If the response indicates success, the C module updates its A→H table to indicate that the LXAs in question should point to the new hashes. Updating of entries in the A→H table may be done as an atomic operation, to ensure the write request is atomic. Note that all requests aligned to 4 KB (or another predefined block size) that fall within a SL may be atomic. The C module returns a response to the requesting R module. The C module adds the list of old hashes to the “decrease reference” batch if needed.
The R module answers the application once it receives a response from the C module.
In the case in which a write request spans multiple SLs, the R module splits the request and sends smaller write requests to several C modules. Each C module receives one request per SL (with a unique request ID). The flow continues as in the simpler case above, except that now the R module aggregates the responses before it answers the application.
Referring to
The first deduplication device 1010 includes a disk 1014 (e.g., a flash disk) and memory 1016. The memory 1016 includes metadata 1022a, which is associated with data 1026a stored on the first deduplication device 1010. The disk 1014 includes metadata 1022b, which is associated with data 1026b stored on the second deduplication device 1012 and includes a pointer 1028 indicating the location in the second deduplication device 1012 where the data 1026b is stored. In one example, the pointer 1028 is a hash value.
The metadata (e.g., metadata 1022a, 1022b) is the information which maps between logical unit and the physical layout of the data and may include address to hash information and hash to physical storage location information.
Referring to
Process 1100 moves metadata from memory to disk (1102). For example, the tiering module 1004 moves (i.e., destages) metadata 1022b from the memory 1016 to the disk 1016. Moving the metadata from memory 1016 to disk 1016 increases the amount of metadata stored by the deduplication device overall, and thus allows an increase of the address space supported. However accessing metadata which was destaged to flash (i.e., to the first deduplication storage device), is slower and thus the read/write locations which have been destaged will be slower. In one example, the metadata is destaged from the memory to the disk if the data associated with the metadata has not been accessed in a predetermined amount of time.
Process 1100 moves the data related to the metadata to a second deduplication device (1106) and deletes the data from the first deduplication device (1110). For example, the tiering module 1004 moves the data 1026b related to the metadata 1022b to the second deduplication device 1012 and deletes the data from the first deduplication device 1010, which increases the amount of data stored in the first deduplication device 1010. Since the data is accessed rarely, then the fact that it is on a slower second deduplication device has relatively low effect on the overall system performance.
In some embodiments, all the data related to the metadata being flushed to first deduplication device is tiered to the second deduplication device, but in some cases only portions of the data are destaged. The destaging process is asynchronous and does not delay any I/Os.
Process 1100 adds a pointer to the metadata indicating where the data is in the second deduplication device (1114). For example, the pointer 1028 is added to the metadata 1022b. The pointer 1028 may include the hash value of the data moved, as the second deduplication device 1012 can allow access to the data by its hash key as it is a deduplication device.
Referring to
Process 1200 reads metadata for the location to be read and discovers that the metadata points to the second deduplication device. The metadata in this case may be stored on the flash disk 1014 of the first deduplication device as it has been destaged from the memory 1016.
Process 1200 retrieves data, using a pointer at the first deduplication device, from the second duplication device (1206) and sends the data to the first deduplication device (1210). For example, if the data 1026b is to be read, the tiering module 1004 reads the pointer 1028 to determine the location of data 1026b and sends the data 1026b from the second deduplication device 1012 to the first deduplication device 1010. Process 1200 writes the data to the first deduplication device so that the next time the data can be retrieved faster (1214) and sends the data to the host (1216).
In many embodiments, if data is read from the second deduplication device not only is the data read retrieved to the first duplication device, but also data which is located near the data (i.e., close offsets) are moved from the second de-duplication device to the first deduplication device. Once data is retrieved from the second deduplication device, it is erased from there.
If the first deduplication device receives a write to a location with data that has been destaged to the second deduplication device, the write data is written to the first deduplication device and the metadata pointer is changed to point to the first deduplication device.
Referring to
The processes described herein (e.g., the processes 1100 and 1200) are not limited to use with the hardware and software of
The system may be implemented, at least in part, via a computer program product, (e.g., in a non-transitory machine-readable storage medium such as, for example, a non-transitory computer-readable medium), for execution by, or to control the operation of, data processing apparatus (e.g., a programmable processor, a computer, or multiple computers)). Each such program may be implemented in a high level procedural or object-oriented programming language to communicate with a computer system. However, the programs may be implemented in assembly or machine language. The language may be a compiled or an interpreted language and it may be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program may be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a communication network. A computer program may be stored on a non-transitory machine-readable medium that is readable by a general or special purpose programmable computer for configuring and operating the computer when the non-transitory machine-readable medium is read by the computer to perform the processes described herein. For example, the processes described herein may also be implemented as a non-transitory machine-readable storage medium, configured with a computer program, where upon execution, instructions in the computer program cause the computer to operate in accordance with the processes. A non-transitory machine-readable medium may include but is not limited to a hard drive, compact disc, flash memory, non-volatile memory, volatile memory, magnetic diskette and so forth but does not include a transitory signal per se.
The processes described herein are not limited to the specific examples described. For example, the processes 1100 and 1200 are not limited to the specific processing order of
The processing blocks (for example, in the processes 1100 and 1200) associated with implementing the system may be performed by one or more programmable processors executing one or more computer programs to perform the functions of the system. All or part of the system may be implemented as, special purpose logic circuitry (e.g., an FPGA (field-programmable gate array) and/or an ASIC (application-specific integrated circuit)). All or part of the system may be implemented using electronic hardware circuitry that include electronic devices such as, for example, at least one of a processor, a memory, a programmable logic device or a logic gate.
Elements of different embodiments described herein may be combined to form other embodiments not specifically set forth above. Other embodiments not specifically described herein are also within the scope of the following claims.
Natanzon, Assaf, Panidis, Anestis, Shoikhet, Kirill, Halevi, Ido
Patent | Priority | Assignee | Title |
10019194, | Sep 23 2016 | EMC IP Holding Company, LLC | Eventually consistent synchronous data replication in a storage system |
10031703, | Dec 31 2013 | EMC Corporation | Extent-based tiering for virtual storage using full LUNs |
10042751, | Sep 30 2015 | EMC IP HOLDING COMPANY LLC | Method and system for multi-tier all-flash array |
10055148, | Dec 22 2015 | EMC IP HOLDING COMPANY LLC | Storing application data as an enhanced copy |
10061666, | Dec 30 2011 | EMC International Company | Method and apparatus for adding a director to storage with network-based replication without data resynchronization |
10067837, | Dec 28 2015 | EMC IP HOLDING COMPANY LLC | Continuous data protection with cloud resources |
10078459, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Ransomware detection using I/O patterns |
10082980, | Jun 20 2014 | EMC IP HOLDING COMPANY LLC | Migration of snapshot in replication system using a log |
10101943, | Sep 25 2014 | EMC IP HOLDING COMPANY LLC | Realigning data in replication system |
10108356, | Mar 25 2016 | EMC IP HOLDING COMPANY LLC | Determining data to store in retention storage |
10114581, | Dec 27 2016 | EMC IP HOLDING COMPANY LLC | Creating a virtual access point in time on an object based journal replication |
10133874, | Dec 28 2015 | EMC IP HOLDING COMPANY LLC | Performing snapshot replication on a storage system not configured to support snapshot replication |
10140039, | Dec 15 2016 | EMC IP HOLDING COMPANY LLC | I/O alignment for continuous replication in a storage system |
10146961, | Sep 23 2016 | EMC IP Holding Company, LLC | Encrypting replication journals in a storage system |
10152267, | Mar 30 2016 | EMC Corporation | Replication data pull |
10191687, | Dec 15 2016 | EMC IP HOLDING COMPANY LLC | Adaptive snap-based replication in a storage system |
10210073, | Sep 23 2016 | EMC IP Holding Company, LLC | Real time debugging of production replicated data with data obfuscation in a storage system |
10223023, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Bandwidth reduction for multi-level data replication |
10229006, | Dec 28 2015 | EMC IP HOLDING COMPANY LLC | Providing continuous data protection on a storage array configured to generate snapshots |
10235060, | Apr 14 2016 | EMC IP Holding Company, LLC | Multilevel snapshot replication for hot and cold regions of a storage system |
10235061, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Granular virtual machine snapshots |
10235064, | Dec 27 2016 | EMC IP HOLDING COMPANY LLC | Optimized data replication using special NVME protocol and running in a friendly zone of storage array |
10235087, | Mar 30 2016 | EMC IP HOLDING COMPANY LLC | Distributing journal data over multiple journals |
10235088, | Mar 30 2016 | EMC IP HOLDING COMPANY LLC | Global replication policy for multi-copy replication |
10235090, | Sep 23 2016 | EMC IP Holding Company, LLC | Validating replication copy consistency using a hash function in a storage system |
10235091, | Sep 23 2016 | EMC IP HOLDING COMPANY LLC | Full sweep disk synchronization in a storage system |
10235092, | Dec 15 2016 | EMC IP HOLDING COMPANY LLC | Independent parallel on demand recovery of data replicas in a storage system |
10235145, | Sep 13 2012 | EMC International Company | Distributed scale-out replication |
10235196, | Dec 28 2015 | EMC IP HOLDING COMPANY LLC | Virtual machine joining or separating |
10235247, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Compressing memory snapshots |
10296419, | Mar 27 2015 | EMC IP HOLDING COMPANY LLC | Accessing a virtual device using a kernel |
10324637, | Dec 13 2016 | EMC IP HOLDING COMPANY LLC | Dual-splitter for high performance replication |
10324798, | Sep 25 2014 | EMC IP HOLDING COMPANY LLC | Restoring active areas of a logical unit |
10353603, | Dec 27 2016 | EMC IP HOLDING COMPANY LLC | Storage container based replication services |
10366011, | May 03 2018 | EMC IP HOLDING COMPANY LLC | Content-based deduplicated storage having multilevel data cache |
10372349, | Dec 30 2015 | EMC IP HOLDING COMPANY LLC | Method and apparatus for facilitating wear leveling of solid state disk |
10409629, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Automated host data protection configuration |
10409787, | Dec 22 2015 | EMC IP HOLDING COMPANY LLC | Database migration |
10409986, | Sep 26 2016 | EMC IP HOLDING COMPANY LLC | Ransomware detection in a continuous data protection environment |
10423634, | Dec 27 2016 | EMC IP HOLDING COMPANY LLC | Temporal queries on secondary storage |
10437783, | Sep 25 2014 | EMC IP HOLDING COMPANY LLC | Recover storage array using remote deduplication device |
10467102, | Dec 15 2016 | EMC IP HOLDING COMPANY LLC | I/O score-based hybrid replication in a storage system |
10489321, | Jul 31 2018 | EMC IP HOLDING COMPANY LLC | Performance improvement for an active-active distributed non-ALUA system with address ownerships |
10496487, | Dec 03 2014 | EMC IP HOLDING COMPANY LLC | Storing snapshot changes with snapshots |
10572172, | Apr 20 2018 | EMC IP HOLDING COMPANY LLC | Multi-granular data reduction for remote data replication |
10579282, | Mar 30 2016 | EMC IP HOLDING COMPANY LLC | Distributed copy in multi-copy replication where offset and size of I/O requests to replication site is half offset and size of I/O request to production volume |
10592166, | Aug 01 2018 | EMC IP HOLDING COMPANY LLC | Fast input/output in a content-addressable storage architecture with paged metadata |
10613770, | Dec 29 2015 | EMC IP HOLDING COMPANY LLC | Method and apparatus for controlling access to a disk array |
10628268, | Dec 15 2016 | EMC IP HOLDING COMPANY LLC | Proof of data replication consistency using blockchain |
10713221, | Jul 30 2018 | EMC IP HOLDING COMPANY LLC | Dual layer deduplication for a file system running over a deduplicated block storage |
10747458, | Nov 21 2017 | International Business Machines Corporation | Methods and systems for improving efficiency in cloud-as-backup tier |
10747606, | Dec 21 2016 | EMC IP HOLDING COMPANY LLC | Risk based analysis of adverse event impact on system availability |
10747667, | Nov 02 2018 | EMC IP HOLDING COMPANY LLC | Memory management of multi-level metadata cache for content-based deduplicated storage |
10776029, | Dec 21 2018 | DELL PRODUCTS, L.P. | System and method for dynamic optimal block size deduplication |
10776211, | Dec 27 2016 | EMC IP HOLDING COMPANY LLC | Methods, systems, and apparatuses to update point in time journal using map reduce to create a highly parallel update |
10853181, | Jun 29 2015 | EMC IP HOLDING COMPANY LLC | Backing up volumes using fragment files |
10853286, | Jul 31 2018 | EMC IP HOLDING COMPANY LLC | Performance improvement for an active-active distributed non-ALUA system with address ownerships |
10860555, | Aug 27 2018 | DELL PRODUCTS, L.P.; Dell Products, LP | Method and apparatus for two tier data deduplication using weighted graphs |
10956052, | Jul 21 2017 | EMC IP HOLDING COMPANY LLC | Online address to hash (A2H) metadata scanner |
11016677, | Dec 13 2016 | EMC IP HOLDING COMPANY LLC | Dual-splitter for high performance replication |
11093158, | Jan 29 2019 | EMC IP HOLDING COMPANY LLC | Sub-lun non-deduplicated tier in a CAS storage to reduce mapping information and improve memory efficiency |
11144247, | Aug 01 2018 | EMC IP HOLDING COMPANY LLC | Fast input/output in a content-addressable storage architecture with paged metadata |
11194496, | Nov 14 2017 | Samsung Electronics Co., Ltd. | Data deduplication using KVSSD |
11216388, | Apr 30 2019 | EMC IP HOLDING COMPANY LLC | Tiering between storage media in a content aware storage system |
9910735, | Mar 30 2016 | EMC IP HOLDING COMPANY LLC | Generating an application-consistent snapshot |
Patent | Priority | Assignee | Title |
7203741, | Oct 12 2000 | ZEPHYRTEL, INC | Method and system for accelerating receipt of data in a client-to-client network |
7719443, | Jun 27 2008 | EMC BENELUX B V , S A R L | Compressing data in a continuous data protection environment |
7840536, | Dec 26 2007 | EMC BENELUX B V , S A R L | Methods and apparatus for dynamic journal expansion |
7840662, | Mar 28 2008 | EMC BENELUX B V , S A R L | Dynamically managing a network cluster |
7844856, | Dec 26 2007 | EMC BENELUX B V , S A R L | Methods and apparatus for bottleneck processing in a continuous data protection system having journaling |
7860836, | Dec 26 2007 | EMC BENELUX B V , S A R L | Method and apparatus to recover data in a continuous data protection environment using a journal |
7882286, | Sep 26 2008 | EMC BENELUX B V , S A R L | Synchronizing volumes for replication |
7934262, | Dec 26 2007 | EMC BENELUX B V , S A R L | Methods and apparatus for virus detection using journal data |
7958372, | Dec 26 2007 | EMC BENELUX B V , S A R L | Method and apparatus to convert a logical unit from a first encryption state to a second encryption state using a journal in a continuous data protection environment |
8037162, | Oct 12 2000 | ZEPHYRTEL, INC | Method and system for accelerating receipt of data in a client to client network |
8041940, | Dec 26 2007 | EMC BENELUX B V , S A R L | Offloading encryption processing in a storage area network |
8060713, | Dec 21 2005 | EMC BENELUX B V , S A R L | Consolidating snapshots in a continuous data protection system using journaling |
8060714, | Sep 26 2008 | EMC BENELUX B V , S A R L | Initializing volumes in a replication system |
8103937, | Mar 31 2010 | EMC IP HOLDING COMPANY LLC | Cas command network replication |
8108634, | Jun 27 2008 | EMC B.V., S.A.R.L. | Replicating a thin logical unit |
8214612, | Sep 28 2009 | EMC IP HOLDING COMPANY LLC | Ensuring consistency of replicated volumes |
8250149, | Oct 12 2000 | ZEPHYRTEL, INC | Method and system for accelerating receipt of data in a client to client network |
8271441, | Dec 26 2009 | EMC IP HOLDING COMPANY LLC | Virtualized CG |
8271447, | Jun 18 2010 | EMC International Company | Mirroring metadata in a continuous data protection environment |
8332687, | Jun 23 2010 | EMC IP HOLDING COMPANY LLC | Splitter used in a continuous data protection environment |
8335761, | Dec 02 2010 | EMC International Company | Replicating in a multi-copy environment |
8335771, | Sep 29 2010 | EMC IP HOLDING COMPANY LLC | Storage array snapshots for logged access replication in a continuous data protection system |
8341115, | Dec 26 2009 | EMC International Company | Dynamically switching between synchronous and asynchronous replication |
8370648, | Mar 15 2010 | EMC International Company | Writing and reading encrypted data using time-based encryption keys |
8380885, | Jun 30 2011 | EMC IP HOLDING COMPANY LLC | Handling abort commands in replication |
8392680, | Mar 30 2010 | EMC International Company | Accessing a volume in a distributed environment |
8429362, | Mar 31 2011 | EMC IP HOLDING COMPANY LLC | Journal based replication with a virtual service layer |
8433869, | Sep 27 2010 | EMC International Company | Virtualized consistency group using an enhanced splitter |
8438135, | Jun 18 2010 | EMC International Company | Mirroring metadata in a continuous data protection environment |
8464101, | Mar 31 2010 | EMC IP HOLDING COMPANY LLC | CAS command network replication |
8478955, | Sep 27 2010 | EMC International Company | Virtualized consistency group using more than one data protection appliance |
8495304, | Dec 23 2010 | EMC IP HOLDING COMPANY LLC | Multi source wire deduplication |
8510279, | Mar 15 2012 | EMC International Company | Using read signature command in file system to backup data |
8521691, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Seamless migration between replication technologies |
8521694, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Leveraging array snapshots for immediate continuous data protection |
8543609, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Snapshots in deduplication |
8583885, | Dec 01 2009 | EMC IP HOLDING COMPANY LLC | Energy efficient sync and async replication |
8600945, | Mar 29 2012 | EMC IP HOLDING COMPANY LLC | Continuous data replication |
8601085, | Mar 28 2011 | EMC IP HOLDING COMPANY LLC | Techniques for preferred path determination |
8627012, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for improving cache performance |
8683592, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | Associating network and storage activities for forensic analysis |
8694700, | Sep 29 2010 | EMC IP HOLDING COMPANY LLC | Using I/O track information for continuous push with splitter for storage device |
8706700, | Dec 23 2010 | EMC IP HOLDING COMPANY LLC | Creating consistent snapshots across several storage arrays or file systems |
8712962, | Dec 01 2011 | EMC IP HOLDING COMPANY LLC | Snapshots in de-duplication |
8719497, | Sep 21 2011 | EMC IP HOLDING COMPANY LLC | Using device spoofing to improve recovery time in a continuous data protection environment |
8725691, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Dynamic LUN resizing in a replication environment |
8725692, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Replication of xcopy command |
8726066, | Mar 31 2011 | EMC IP HOLDING COMPANY LLC | Journal based replication with enhance failover |
8738813, | Dec 27 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for round trip synchronous replication using SCSI reads |
8745004, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Reverting an old snapshot on a production volume without a full sweep |
8751828, | Dec 23 2010 | EMC IP HOLDING COMPANY LLC | Sharing encryption-related metadata between multiple layers in a storage I/O stack |
8769336, | Dec 27 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for preventing journal loss on failover in symmetric continuous data protection replication |
8805786, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Replicating selected snapshots from one storage array to another, with minimal data transmission |
8806161, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Mirroring splitter meta data |
8825848, | Mar 20 2012 | EMC IP HOLDING COMPANY LLC | Ordering of event records in an electronic system for forensic analysis |
8832399, | Sep 27 2010 | EMC International Company | Virtualized consistency group using an enhanced splitter |
8850143, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Point in time access in a replication environment with LUN resizing |
8850144, | Mar 29 2012 | EMC IP HOLDING COMPANY LLC | Active replication switch |
8862546, | Jun 30 2011 | EMC IP HOLDING COMPANY LLC | Virtual access roll |
8892835, | Jun 07 2012 | EMC IP HOLDING COMPANY LLC | Insertion of a virtualization layer into a replication environment |
8898112, | Sep 07 2011 | EMC IP HOLDING COMPANY LLC | Write signature command |
8898409, | Jun 27 2012 | EMC International Company | Journal-based replication without journal loss |
8898515, | Jun 28 2012 | EMC International Company | Synchronous replication using multiple data protection appliances across multiple storage arrays |
8898519, | Mar 30 2012 | EMC IP HOLDING COMPANY LLC | Method and apparatus for an asynchronous splitter |
8914595, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Snapshots in deduplication |
8924668, | Dec 23 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for an application- and object-level I/O splitter |
8930500, | Oct 12 2000 | ZEPHYRTEL, INC | Method and system for accelerating receipt of data in a client to client network |
8930947, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for live migration of a virtual machine with dedicated cache |
8935498, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Splitter based hot migration |
8949180, | Jun 28 2012 | EMC International Company | Replicating key-value pairs in a continuous data protection system |
8954673, | Mar 20 2012 | EMC International Company | Using a conditional read request and a hash to determine synchronization of data in a cache at a host with data in storage array |
8954796, | Jun 26 2012 | EMC International Company | Recovery of a logical unit in a consistency group while replicating other logical units in the consistency group |
8959054, | Mar 25 2010 | EMC IP HOLDING COMPANY LLC | Methods and apparatus for optimal journaling for continuous data replication |
8977593, | Dec 26 2009 | EMC IP HOLDING COMPANY LLC | Virtualized CG |
8977826, | Dec 28 2011 | EMC IP HOLDING COMPANY LLC | Extent commands in replication |
8996460, | Mar 14 2013 | EMC IP HOLDING COMPANY LLC | Accessing an image in a continuous data protection using deduplication-based storage |
8996461, | Mar 28 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for replicating the punch command |
8996827, | Dec 27 2011 | EMC IP HOLDING COMPANY LLC | Creating and maintaining clones in continuous data protection |
9003138, | Jun 30 2011 | EMC IP HOLDING COMPANY LLC | Read signature command |
9026696, | Sep 29 2010 | EMC IP HOLDING COMPANY LLC | Using I/O track information for continuous push with splitter for storage device |
9031913, | Dec 28 2011 | EMC IP HOLDING COMPANY LLC | File replication |
9032160, | Dec 29 2011 | EMC IP HOLDING COMPANY LLC | Continuous data replication |
9037818, | Mar 29 2012 | EMC IP HOLDING COMPANY LLC | Active replication switch |
9063994, | Mar 31 2011 | EMC IP HOLDING COMPANY LLC | Networked based replication of distributed volumes |
9069479, | Sep 29 2011 | EMC IP HOLDING COMPANY LLC | Snapshots in deduplication |
9069709, | Jun 24 2013 | EMC International Company | Dynamic granularity in data replication |
9081754, | Mar 30 2012 | EMC IP HOLDING COMPANY LLC | Method and apparatus for cascaded replication using a multi splitter |
9081842, | Mar 15 2013 | EMC IP HOLDING COMPANY LLC | Synchronous and asymmetric asynchronous active-active-active data access |
9087008, | Jun 24 2013 | EMC International Company | Replicating a volume using snapshots |
9087112, | Jun 24 2013 | EMC International Company | Consistency across snapshot shipping and continuous replication |
9104529, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for copying a cache system |
9110914, | Mar 14 2013 | EMC IP HOLDING COMPANY LLC | Continuous data protection using deduplication-based storage |
9116811, | Jun 30 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9128628, | Mar 13 2013 | EMC IP HOLDING COMPANY LLC | Dynamic replication mode switching |
9128855, | Jun 26 2013 | EMC IP HOLDING COMPANY LLC | Flash cache partitioning |
9134914, | Dec 27 2012 | EMC Corporation | Deduplication |
9135119, | Sep 28 2012 | EMC IP HOLDING COMPANY LLC | System and method for data management |
9135120, | Jun 27 2012 | EMC IP HOLDING COMPANY LLC | Consistency group moving |
9146878, | Jun 25 2013 | EMC IP HOLDING COMPANY LLC | Storage recovery from total cache loss using journal-based replication |
9152339, | Mar 15 2013 | EMC IP HOLDING COMPANY LLC | Synchronization of asymmetric active-active, asynchronously-protected storage |
9152578, | Mar 12 2013 | EMC IP HOLDING COMPANY LLC | Securing data replication, backup and mobility in cloud storage |
9152814, | Mar 15 2010 | EMC International Company | Writing and reading encrypted data using time-based encryption keys |
9158578, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for migrating virtual machines |
9158630, | Dec 19 2013 | EMC IP HOLDING COMPANY LLC | Testing integrity of replicated storage |
9160526, | Dec 23 2010 | EMC IP HOLDING COMPANY LLC | Sharing encryption-related metadata between a host and an external intermediate device |
9177670, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Method and apparatus for flash cache management |
9189339, | Mar 28 2014 | EMC IP HOLDING COMPANY LLC | Replication of a virtual distributed volume with virtual machine granualarity |
9189341, | Mar 30 2012 | EMC IP HOLDING COMPANY LLC | Method and apparatus for multi-copy replication using a multi-splitter |
9201736, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Methods and apparatus for recovery of complex assets in distributed information processing systems |
9223659, | Jun 28 2012 | EMC International Company | Generating and accessing a virtual volume snapshot in a continuous data protection system |
9225529, | Dec 10 2013 | EMC IP HOLDING COMPANY LLC | Encrypted virtual machines in a cloud |
9235481, | Dec 29 2011 | EMC IP HOLDING COMPANY LLC | Continuous data replication |
9235524, | Dec 30 2011 | EMC IP HOLDING COMPANY LLC | System and method for improving cache performance |
9235632, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Synchronization of replication |
9244997, | Mar 15 2013 | EMC IP HOLDING COMPANY LLC | Asymmetric active-active access of asynchronously-protected data storage |
9256605, | Aug 03 2011 | EMC International Company | Reading and writing to an unexposed device |
9274718, | Jun 20 2014 | EMC IP HOLDING COMPANY LLC | Migration in replication system |
9275063, | Sep 30 2011 | EMC IP HOLDING COMPANY LLC | Replication optimizated IO |
9286052, | Sep 15 2011 | EMC International Company | Upgrading software on a pair of nodes in a clustered environment |
9305009, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Synchronous replication of virtualized storage processors |
9323750, | Sep 29 2010 | EMC IP HOLDING COMPANY LLC | Storage array snapshots for logged access replication in a continuous data protection system |
9330155, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Unified management of sync and async replication for block and file objects |
9336094, | Sep 13 2012 | EMC International Company | Scaleout replication of an application |
9336230, | Dec 28 2011 | EMC Corporation | File replication |
9367260, | Dec 13 2013 | EMC IP HOLDING COMPANY LLC | Dynamic replication system |
9378096, | Jun 30 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9378219, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Metro-cluster based on synchronous replication of virtualized storage processors |
9378261, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Unified synchronous replication for block and file objects |
9383937, | Mar 14 2013 | EMC IP HOLDING COMPANY LLC | Journal tiering in a continuous data protection system using deduplication-based storage |
9389800, | Mar 27 2014 | EMC IP HOLDING COMPANY LLC | Synthesizing virtual machine disk backups |
9405481, | Dec 17 2014 | EMC IP HOLDING COMPANY LLC | Replicating using volume multiplexing with consistency group file |
9405684, | Sep 28 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9405765, | Dec 17 2013 | EMC IP HOLDING COMPANY LLC | Replication of virtual machines |
9411535, | Mar 27 2015 | EMC IP HOLDING COMPANY LLC | Accessing multiple virtual devices |
9459804, | Mar 29 2012 | EMC IP HOLDING COMPANY LLC | Active replication switch |
9460028, | Dec 27 2012 | EMC IP HOLDING COMPANY LLC | Non-disruptive and minimally disruptive data migration in active-active clusters |
9471579, | Jun 24 2011 | EMC IP HOLDING COMPANY LLC | Replicating selected snapshots from one storage array to another, with minimal data transmission |
9477407, | Jun 28 2013 | EMC IP HOLDING COMPANY LLC | Intelligent migration of a virtual storage unit to another data storage system |
9501542, | Mar 11 2008 | EMC Corporation | Methods and apparatus for volume synchronization |
9507732, | Sep 28 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9507845, | Mar 27 2014 | EMC IP HOLDING COMPANY LLC | Virtual splitter |
9514138, | Mar 15 2012 | EMC International Company | Using read signature command in file system to backup data |
9524218, | Sep 09 2013 | EMC IP HOLDING COMPANY LLC | Leverage fast VP extent-level statistics within CDP environments |
9529885, | Sep 29 2014 | EMC IP HOLDING COMPANY LLC | Maintaining consistent point-in-time in asynchronous replication during virtual machine relocation |
9535800, | Sep 30 2014 | EMC IP HOLDING COMPANY LLC | Concurrent data recovery and input/output processing |
9535801, | Jun 30 2011 | EMC IP HOLDING COMPANY LLC | Xcopy in journal based replication |
9547459, | Jun 27 2013 | EMC IP HOLDING COMPANY LLC | Techniques for data relocation based on access patterns |
9547591, | Sep 28 2012 | EMC IP HOLDING COMPANY LLC | System and method for cache management |
9552405, | Sep 30 2013 | EMC IP HOLDING COMPANY LLC | Methods and apparatus for recovery of complex assets in distributed information processing systems |
9557921, | Mar 26 2015 | EMC IP HOLDING COMPANY LLC | Virtual volume converter |
9557925, | Dec 29 2014 | EMC IP HOLDING COMPANY LLC | Thin replication |
9563517, | Dec 30 2013 | EMC IP HOLDING COMPANY LLC | Cloud snapshots |
9563684, | Dec 31 2013 | EMC IP HOLDING COMPANY LLC | Replication cookie |
9575851, | Jun 27 2012 | EMC IP HOLDING COMPANY LLC | Volume hot migration |
9575857, | Jun 27 2012 | EMC IP HOLDING COMPANY LLC | Active/active replication |
9575894, | Mar 27 2015 | EMC IP HOLDING COMPANY LLC | Application aware cache coherency |
9582382, | Dec 16 2010 | EMC IP HOLDING COMPANY LLC | Snapshot hardening |
9588703, | Mar 28 2011 | EMC IP HOLDING COMPANY LLC | Method and apparatus for replicating the punch command |
9588847, | Mar 25 2014 | EMC IP HOLDING COMPANY LLC | Recovering corrupt virtual machine disks |
20140188805, |
Date | Maintenance Fee Events |
Oct 21 2020 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Oct 23 2024 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Date | Maintenance Schedule |
May 30 2020 | 4 years fee payment window open |
Nov 30 2020 | 6 months grace period start (w surcharge) |
May 30 2021 | patent expiry (for year 4) |
May 30 2023 | 2 years to revive unintentionally abandoned end. (for year 4) |
May 30 2024 | 8 years fee payment window open |
Nov 30 2024 | 6 months grace period start (w surcharge) |
May 30 2025 | patent expiry (for year 8) |
May 30 2027 | 2 years to revive unintentionally abandoned end. (for year 8) |
May 30 2028 | 12 years fee payment window open |
Nov 30 2028 | 6 months grace period start (w surcharge) |
May 30 2029 | patent expiry (for year 12) |
May 30 2031 | 2 years to revive unintentionally abandoned end. (for year 12) |