Aspects of the invention relate to creating a snapshot of a file system on an OAS system. A snapshot of a file system is an image of a file system at a particular point in time. A snapshot of a file system may be used to restore the file system to its state at the time of creation of the snapshot in, for example, situations where the file system has become corrupted.

Patent
   7752492
Priority
May 25 2007
Filed
May 25 2007
Issued
Jul 06 2010
Expiry
Aug 04 2028
Extension
437 days
Assg.orig
Entity
Large
70
5
all paid
13. A computer system comprising:
at least one storage device that stores a plurality of content units each identified by an object identifier; and
at least one controller that:
in response to a failure in the computer system that results in a loss of at least some of the plurality of content units caused, at least in part, by a loss of metadata identifying a storage location on the least one storage device for the lost content units, provides the object identifier for each of the lost content units.
1. A method of processing content in a computer system comprising at least one storage device that stores a plurality of content units each identified by an object identifier, the method comprising an act of:
(A) in response to a failure in the computer system that results in a loss of at least some of the plurality of content units caused, at least in part, by a loss of metadata identifying a storage location on the least one storage device for the lost content units, providing the object identifier for each of the lost content units.
7. At least one computer readable medium encoded with instructions that, when executed, perform a method of processing content in a computer system comprising at least one storage device that stores a plurality of content units each identified by an object identifier, the method comprising an act of:
(A) in response to a failure in the computer system that results in a loss of at least some of the plurality of content units caused, at least in part, by a loss of metadata identifying a storage location on the least one storage device for the lost content units, providing the object identifier for each of the lost content units.
2. The method of claim 1, wherein the computer system comprises at least one object addressable storage (OAS) system, and wherein the at least one storage device is disposed in the at least one (OAS) system, the at least one OAS system having an OAS layer that receives OAS access requests from an application program.
3. The method of claim 1, wherein the object identifier for each content unit is a content address that is generated, at least in part, based on at least a portion of the content of the content unit.
4. The method of claim 2, wherein the OAS system includes a file system, and wherein the failure that results in a loss of the at least some of the plurality of content units comprises a corruption of the file system.
5. The method of claim 4, further comprising an act of:
(B) restoring the file system from a previously-created snapshot of the file system.
6. The method of claim 5, wherein lost content units are content units stored on the least one storage device after creation of the previously-created snapshot.
8. The at least one computer readable medium of claim 7, wherein the computer system comprises at least one object addressable storage (OAS) system, and wherein the at least one storage device is disposed in that at least one (OAS) system, the at least one OAS system having an OAS layer that receives OAS access requests from an application program.
9. The at least one computer readable medium of claim 7, wherein the object identifier for each content unit is a content address that is generated, at least in part, based on at least a portion of the content of the content unit.
10. The at least one computer readable medium of claim 8, wherein the OAS system includes a file system, and wherein the failure that results in a loss of the at least some of the plurality of content units comprises a corruption of the file system.
11. The at least one computer readable medium of claim 10, wherein the method further comprises an act of:
(B) restoring the file system from a previously-created snapshot of the file system.
12. The at least one computer readable medium of claim 11, wherein lost content units are content units stored on the least one storage device after creation of the previously-created snapshot.
14. The computer system of claim 13, wherein the computer system further comprises at least one object addressable storage (OAS) system, and wherein the at least one storage device and the at least one controller are disposed in that at least one (OAS) system, the at least one OAS system having an OAS layer that receives OAS access requests from an application program.
15. The computer system of claim 13, wherein the object identifier for each content unit is a content address that is generated, at least in part, based on at least a portion of the content of the content unit.
16. The computer system of claim 14, wherein the OAS system includes a file system, and wherein the failure that results in a loss of the at least some of the plurality of content units comprises a corruption of the file system.
17. The computer system of claim 16, wherein the at least one controller restores the file system from a previously-created snapshot of the file system.

The present application relates to computer systems employing object addressable storage.

Virtually all computer application programs rely on storage that may be used to store computer code and data manipulated by the computer code. A typical computer system includes one or more host computers that execute such application programs and one or more storage systems that provide storage.

The host computers may access data by sending access requests to the one or more storage systems. Some storage systems require that the access requests identify units of data to be accessed using logical volume and block addresses that define where the units of data are stored on the storage system. Such storage systems are known as “block I/O” storage systems. In some block I/O storage systems, the logical volumes presented by the storage system to the host correspond directly to physical storage devices (e.g., disk drives) on the storage system, so that the specification of a logical volume and block address specifies where the data is physically stored within the storage system. In other block I/O storage systems (referred to as intelligent storage systems), internal mapping techniques may be employed so that the logical volumes presented by the storage system do not necessarily map in a one-to-one manner to physical storage devices within the storage system. Nevertheless, the specification of a logical volume and a block address used with an intelligent storage system specifies where associated content is logically stored within the storage system, and from the perspective of devices outside of the storage system (e.g., a host) is perceived as specifying where the data is physically stored.

In contrast to block I/O storage systems, some storage systems receive and process access requests that identify a data unit or other content unit (also referenced to as an object) using an object identifier, rather than an address that specifies where the data unit is physically or logically stored in the storage system. Such storage systems are referred to as object addressable storage (OAS) systems. In object addressable storage, a content unit may be identified (e.g., by host computers requesting access to the content unit) using its object identifier and the object identifier may be independent of both the physical and logical location(s) at which the content unit is stored (although it is not required to be because in some embodiments the storage system may use the object identifier to inform where a content unit is stored in a storage system). From the perspective of the host computer (or user) accessing a content unit on an OAS system, the object identifier does not control where the content unit is logically (or physically) stored. Thus, in an OAS system, if the physical or logical location at which the unit of content is stored changes, the identifier by which host computer(s) access the unit of content may remain the same. In contrast, in a block I/O storage system, if the location at which the unit of content is stored changes in a manner that impacts the logical volume and block address used to access it, any host computer accessing the unit of content must be made aware of the location change and then use the new location of the unit of content for future accesses.

One example of an OAS system is a content addressable storage (CAS) system. In a CAS system, the object identifiers that identify content units are content addresses. A content address is an identifier that is computed, at least in part, from at least a portion of the content (which can be data and/or metadata) of its corresponding unit of content. For example, a content address for a unit of content may be computed by hashing the unit of content and using the resulting hash value as the content address. Storage systems that identify content by a content address are referred to as content addressable storage (CAS) systems.

Some OAS system employ file systems to manage storage of objects on one or more storage devices. A file system is a logical construct that translates physical blocks of storage on a storage device into logical files and directories. In this way, the file system aids in organizing content stored on a disk. For example, an application program having ten logically related blocks of content to store on disk may store the content in a single file in the file system. Thus, the application program may simply track the name and/or location of the file, rather than tracking the block addresses of each of the ten blocks on disk that store the content.

File systems maintain metadata for each file that, inter alia, indicates the physical disk locations of the content logically stored in the file. For example, in UNIX file systems an inode is associated with each file and stores metadata about the file. The metadata includes information such as access permissions, time of last access of the file, time of last modification of the file, and which blocks on the physical storage devices store its content. The file system may also maintain a map, referred to as a free map in UNIX file systems, of all the blocks on the physical storage system at which the file system may store content. The file system tracks which blocks in the map are currently in use to store file content and which are available to store file content.

When an application program requests that the file system store content in a file, the file system may use the map to select available blocks and send a request to the physical storage devices to store the file content at the selected blocks. The file system may then store metadata (e.g., in an inode) that associates the filename for the file with the physical location of the content on the storage device(s). When the file system receives a subsequent request to access the file, the file system may access the metadata, use it to determine the blocks on the physical storage device at which the file's content is physically stored, request the content from the physical storage device(s), and return the content in response to the request.

As mentioned above, some OAS systems may store content in a file system. FIG. 1 shows an example of such an OAS system 101 that includes an OAS interface 103, a file system 105, and one or more storage devices 107. When OAS Interface 103 receives a request (e.g., from an application program) to store a content unit, the OAS Interface may assign an object identifier to the content unit (which may be generated either by the OAS system, the entity that issued the request, or some other entity), and may issue a request to file system 105 to store the content unit in one or more files. The file system may store the content unit on physical storage device(s) 107, and may store metadata associating the file(s) in which the content of the content unit is stored with the physical location(s) of the content on the physical storage device(s).

When a request to access the content unit (that identifies the content unit using its object identifier) is subsequently received by OAS Interface 103, the OAS Interface determines the file or files in file system 105 that logically store the content of the content unit in any suitable way.

In some OAS systems, when the OAS Interface receives a request to store a content unit and stores the content unit in one or more files in the file system, the OAS Interface may store metadata that associates the object identifier for the content unit with the filename(s) and/or file system location(s) of the file. The OAS Interface may use this metadata to determine the file(s) that store the content of the content unit. In some OAS systems, when the OAS Interface, in response to a request to store a content unit, stores the content of the content unit in one or more file(s) in the file system, the OAS Interface may instruct the file system to give the one or more file(s) a file name that includes all or a portion of the object identifier for the content unit. When a subsequent access request for the content unit (that identifies the content unit using its object identifier) is received, the OAS Interface may determine the file(s) that store the content of the content unit by locating the file(s) that have the object identifier (or a portion thereof) in their filename.

Once the OAS Interface determines the file(s) in file system 105 that store(s) the content of the content unit, the OAS Interface made send a request to the file system to access the file(s). In response, the file system may determine the physical storage location(s) of the content unit on the physical storage device(s), and request the content stored and the determined physical storage location(s) from the physical storage device. Upon receiving the requested content, the file system may return the content to the OAS Interface, which may return it to the requesting entity.

The simplified block diagram of OAS system 101 shows file system 105 directly accessing storage device(s) 107. However, this is provided merely as a simplified example, as file system 105 may access the storage device(s) in any suitable way. For example, in some embodiments file system 105 may access the storage device(s) via a device driver that provides an interface to the storage device(s) or via an operating system that interfaces with the device driver for the storage device(s).

One embodiment is directed to a method of processing content in a computer system comprising at least one computer that executes an application program that uses an object addressable storage (OAS) interface to access a content unit using an object identifier, the computer system further comprising a file system having a non-OAS storage interface that makes available a plurality of files. The method comprises acts of: (A) providing an OAS layer that receives OAS access requests from the application program for content units in a format compatible with the OAS interface and converts the OAS access requests into file system access requests that are in a format compatible with the non-OAS storage interface provided by the file system; and (B) using a snapshot facility to create a snapshot of the file system at least one point in time to facilitate data recovery in the event that the file system becomes corrupted subsequent to the creating of the snapshot. Another embodiment is directed to at least one computer readable medium encoded with instructions that, when executed, perform the above-described method.

A further embodiment is directed to a computer system for use with an application program that uses an object addressable storage (OAS) interface to access a content unit stored in the computer system using an object identifier. The computer system comprises: a file system having a non-OAS storage interface that makes available a plurality of files; and at least one controller that: provides an OAS layer that receives OAS access requests from the application program for content units in a format compatible with the OAS interface and converts the OAS access requests into file system access requests that are in a format compatible with the non-OAS storage interface provided by the file system; and uses a snapshot facility to create a snapshot of the file system at least one point in time to facilitate data recovery in the event that the file system becomes corrupted subsequent to the creating of the snapshot.

Another embodiment is directed to a method of processing content in a computer system comprising at least one storage device that stores a plurality of content units each identified by an object identifier. The method comprises an act of: (A) in response to a failure in the computer system that results in a loss of at least some of the plurality of content units, providing the object identifier for each of the lost content units. A further embodiment is directed to at least one computer readable medium encoded with instructions that, when executed, perform the above-described method.

Another embodiment is directed to a computer system comprising: at least one storage device that stores a plurality of content units each identified by an object identifier; and at least one controller that: in response to a failure in the computer system that results in a loss of at least some of the plurality of content units, provides the object identifier for each of the lost content units.

A further embodiment is directed to a method of processing content in a computer system comprising at least one computer that executes an application program that uses an object addressable storage (OAS) interface to access a content unit using an object identifier, the computer system further comprising a first file system and a second file system, the first and second file systems each having a non-OAS storage interface that makes available a plurality of files. The method comprises acts of: (A) creating a snapshot of the first file system at least one point in time; (B) receiving, subsequent to creating the snapshot, an OAS access request from the application program to store a content unit, wherein the OAS access request is in a format compatible with the OAS interface; (C) converting the OAS access request into at least one file system access request that is in a format compatible with the non-OAS storage interface provided by the first and second file systems; (D) transmitting the at least one file system access request to store the content unit to the first file system; and (E) transmitting the at least one file system access request to store the content unit to the second file system so that the content unit is stored in both the first and second file systems. Another embodiment is directed to at least one computer readable medium encoded with instructions that, when executed, perform the above-described method.

A further embodiment is directed to a computer system comprising: at least one computer that executes an application program that uses an object addressable storage (OAS) interface to access a content unit using an object identifier, the at last one computer having a first file system and a second file system, the first and second file systems each having a non-OAS storage interface that makes available a plurality of files; and at least one controller that: creates a snapshot of the first file system at least one point in time; receives, subsequent to creating the snapshot, an OAS access request from the application program to store a content unit, wherein the OAS access request is in a format compatible with the OAS interface; converts the OAS access request into at least one file system access request that is in a format compatible with the non-OAS storage interface provided by the first and second file systems; transmits the at least one file system access request to store the content unit to the first file system; and transmits the at least one file system access request to store the content unit to the second file system so that the content unit is stored in both the first and second file systems.

Another embodiment is directed to a method of processing content in a computer system comprising a file system that stores a plurality of files. The method comprises: employing a snapshot utility to create a snapshot of the file system at given point in time; and ensuring that, when data is deleted from the file system after creation of the snapshot of the file system, the deleted data exists in the snapshot only in encrypted form. A further embodiment is directed to at least one computer readable medium encoded with instructions that, when executed, perform the above-described method.

Another embodiment is directed to a computer system comprising: a file system that stores a plurality of files; and at least one controller that: employs a snapshot utility to create a snapshot of the file system at given point in time; and ensures that, when data is deleted from the file system after creation of the snapshot of the file system, the deleted data exists in the snapshot only in encrypted form.

FIG. 1 is a block diagram of an illustrative object addressable storage (OAS) system, on which some embodiments of the invention may be implemented;

FIG. 2 is a block diagram of an OAS system that includes a snapshot facility, in accordance with one embodiment;

FIG. 3 is a diagram of a timeline illustrating the periodic creation of file system snapshots, in accordance with one embodiment;

FIG. 4 is a block diagram of an illustrative OAS system in which an OAS Interface maintains a list of object identifiers of content units stored to the OAS system in between the creation of snapshots, in accordance with one embodiment;

FIG. 5 is a block diagram of an illustrative OAS system in which an OAS Interface mirrors write requests to two file systems, in accordance with one embodiment; and

FIG. 6 is a block diagram of illustrative storage system that includes a utility for encrypting and decrypting snapshots.

Applicants have appreciated that in any OAS system of the type described above, content stored on storage device(s) 107 may be lost due to corruption of the stored content, hardware failure of a storage device, or other reasons. Various techniques may be used to protect against such losses of content. One such technique is the use of a redundant array of independent disks (RAID), which is a data storage scheme in which content is shared and/or replicated across multiple independent disks. Thus, for example, storage devices 107 may be configured as a RAID array.

In another technique, instead of or in addition to configuring storage devices 107 as a RAID array, content stored on storage devices 107 may be replicated to a different storage system (e.g., at a remote location). This may be done in any suitable way, as the invention is not limited in this respect.

As discussed above, some OAS systems employ a file system that maintains metadata that maps logical files to physical storage locations at which the content logically stored in the files is physically stored on the physical storage devices, and may also indicate which blocks on the physical storage device(s) are being used to store file content and which are available to store file content.

The file system may be too large to store entirely in memory. Thus, the file system may keep only a portion of this metadata in memory, with the remainder being stored elsewhere (e.g., on one of the physical storage devices of the OAS system). The file system may load portions of the metadata from the physical storage devices (e.g., a disk drive) into memory as they are needed, and may page unneeded portions out to disk. Applicants have appreciated that, in conventional OAS systems of the type described above, while the file system metadata stored on disk may be protected from corruption or loss, for example, by RAID, replication, or in some other way, the file system metadata stored in memory is not. Thus, the file system metadata may become corrupted, which may result in incorrect or bad metadata being stored to disk and/or the file system being unable to locate the physical storage location(s) associated with a file because the metadata does not properly identify them. Embodiments of the invention are directed to protecting against corruption of a file system in an OAS system.

The techniques described herein are not limited to protecting against any particular types of corruptions, as numerous types are possible. For example, the file system metadata that indicates which blocks on the physical storage device are available and which blocks are in use may become corrupted such that blocks that are in use are indicated as available. This may result in the file system allocating blocks that its metadata indicates as available (but which are in actuality in use by an existing file) to a new file and overwriting the content of the existing file with the content of the new file. In addition, metadata that indicates at which physical blocks the content of a file is stored may be corrupted. This may result in the file system being unable to locate the content of the file.

File system metadata can become corrupted or inaccurate for any of a number of reasons. For example, if a power outage or system crash occurs before file system metadata stored and modified in memory is written to disk, the updated file system metadata may be lost, and the version of the file system metadata stored on disk may be inaccurate. In addition, two programs attempting to write to or modify a file at the same time may cause file system metadata corruption. Metadata corruption may also occur in situations where a program that is modifying a file crashes before the file system operation is complete. Bugs in the file system may be another cause of metadata corruption.

In some embodiments, the file system used by an OAS system may be protected using a snapshot facility that creates a snapshot of the file system. FIG. 2 is an example of a OAS system 201 in which a file system is protected using a snapshot facility. OAS system 201 includes an OAS Interface 103 and a file system 105. OAS Interface receives access requests (e.g., read and write requests) originated by application programs or other accessing entities for content units in a format compatible with OAS interface (e.g., identifying a content unit via an object identifier) and converts the access requests to requests that are in a format compatible with the interface presented by file system 105 (e.g., identifying files that store requested content by their filenames). File system 105 responds to the access requests by mapping logical file system names and locations to the corresponding physical locations on physical storage devices 107 where the requested content is logically stored.

OAS system 201 also includes a snapshot facility 203 to create one or more snapshots of file system 105. A file system snapshot is an image of the file system as it existed at a particular point in time. In embodiments of the invention that employ a snapshot facility to protect against file system metadata loss and/or corruption, the snapshot facility may create the snapshot in any suitable way, using any suitable technique. For example, the snapshot facility may create a snapshot of the file system at a particular point in time by creating a copy of the file system data (e.g., including the file system metadata and content stored in the files). Alternatively, the snapshot facility may use a copy-on-first-write technique to create the snapshot. In the copy-on-first-write technique, a point in time of the snapshot is selected. Each time a block of data in the file system is modified for the first time after the selected point in time, a copy of the block as stored prior to the modification is stored. Thus, the file system as it existed at the selected point in time may be recovered by replacing the blocks in the current version of the file system that were modified after the selected point in time with the copied blocks.

In some embodiments, prior to creating a snapshot, snapshot facility 203 may send a request to OAS Interface 103 to complete processing of any in-process write operations before creation of the snapshot and suspend new write operations (e.g., operations to store content units) until after the snapshot is created. In this manner, no write operations will be ongoing at the time of creation of the snapshot. This enables the file system to have a definite state at the time of creation of the snapshot.

In embodiments in which the snapshot facility creates a full copy of the file system at a selected snapshot time, the snapshot facility need not track modifications to the file system or make additional copies once the snapshot copy of the file system is created. When using the copy-on-first-write technique, the snapshot facility monitors modifications of the file system and creates copies of modified blocks after the selected snapshot time.

Snapshot facility 203 (as well as the snapshot facilities described in connection with other embodiments) may operate at the block level, the file system level, or any other suitable level, as the invention is not limited in this respect. In embodiments in which the snapshot facility operates at the block level, the snapshot facility may be unaware of the mapping between blocks of storage on the physical storage device and the logical files of the storage system. Thus, when a snapshot facility creates a snapshot, the snapshot may include blocks of physical storage that are not part of the file system. For example, in embodiments in which the snapshot facility creates a full copy of the file system, the snapshot facility may include in the snapshot all blocks of storage that are available to it, regardless of whether those blocks are used by the file system to store file data or file system metadata. In embodiments wherein the snapshot facility uses a copy-on-first-write technique, the snapshot facility may monitor any write requests sent to the physical storage device(s) and copy blocks that are modified for the first time since the creation of the snapshot (though these blocks may not store file data or file system metadata). Thus, in some embodiments, “a file system snapshot” may include data that is not associated with the file system (e.g., not file data or file system metadata).

In embodiments in which the snapshot facility operates at the file system level, the snapshot facility may intercept requests to the file system to modify and/or delete files and/or directories. When the snapshot facility intercepts a request to modify or delete a file, the snapshot facility may create a copy of the file and store the copy along with metadata (e.g., metadata indicating the time that the copy was created) and then forward the request to the file system.

In some embodiments OAS Interface 103 may control the snapshot facility. Thus, for example, the OAS Interface may select the times at which a snapshot is to be created and, at each selected time, may complete any in-progress write operations, suspend new write operations, and instruct the snapshot facility to create a snapshot of the file system.

Regardless of which of these (or any other) snapshot techniques are used, the time of creation of the snapshot, as used herein, refers to the selected snapshot time. Thus, for example, in embodiments in which the snapshot facility creates a snapshot by creating a full copy of the file system, if the selected snapshot time is time T1 (i.e., the snapshot stores the state of the file system as it existed at time T1), then the snapshot facility stores a copy of the file system as it existed at time T1 and the time of creation of the snapshot is time T1. In embodiments in which the snapshot facility uses a copy-on-first-write technique, if the selected snapshot time is T1 (i.e., the snapshot stores the state of the file system as it existed at time T1), then the time of creation of the snapshot is time T1, even though the snapshot facility may continue to monitor modifications to the file system and make copies of modified blocks for the snapshot.

In some embodiments, snapshot facility 203 may periodically create snapshots of the file system in OAS system 201. For example, snapshots of the file system may be created at some interval (e.g., either a regular interval or an irregular interval). Any suitable interval may be used, as the invention is not limited in this respect. The interval may be on the order of seconds, minutes, hours, days, weeks, months, or any other suitable period of time.

As shown in the illustrative timeline of FIG. 3, a snapshot of file system 105 may be created at times T0, T1, T2, T3, T4, and T5. Thus, for example, if a file system corruption occurs in between times T1 and T2, the file system can be restored to its state at time T1 using the snapshot.

In some embodiments, when a snapshot of a file system is created, a file system check (e.g., a check performed by a utility such as the UNIX fsck utility) of the snapshot may be performed to verify that the snapshot of the file system is internally consistent. Any suitable check of the file system may be performed, as the invention is not limited in this respect. For example, the file system check may include checking that each allocated block is allocated to no more than one file, checking that each block that is currently in use by a file is indicated as an allocated block by the file system metadata, checking that each block that is indicated as available by the file system is not currently in use by any of the files, and/or any other suitable check. However, the invention is not limited to performing a file system check of the snapshot, as in some embodiments, when a snapshot of the file system is created it may be assumed that the snapshot stores a file system in a consistent state.

In embodiments in which a file system check is performed, if a snapshot is determined to be internally consistent by the file system check, then it may be treated as the most recent valid snapshot of the file system. Alternatively, if the snapshot is determined to be internally inconsistent, any inconsistencies may be corrected, and/or a new snapshot of the file system may be created. Performing a file system check of the snapshot is useful in verifying that the snapshot is valid and can be used to restore the file system, if desired.

Thus, in some embodiments, periodically creating a snapshot of the file system allows the system to be restored to a known state in the event a failure. Applicants have appreciated that when restoring the system to its state at particular point in time, modifications made to the system after that point in time (e.g., content units being stored, modified, and/or deleted), may be lost. Thus, for example, if a file system corruption occurs between time T3 and time T4, the file system may be restored to its state at time T3. However, any modifications made to the file system (e.g., creation of new files, modification of files, deletion of files) after time T3 may be lost. Thus, even though the content of a file created in between times T3 and T4 may be physically stored on the physical storage device(s), the file system may be unable to locate the content because the metadata that indicates which physical storage location(s) store the content is gone.

The window of time in between snapshots (e.g., the amount of time between times T1 and T2) affects the risk that modifications will be lost in the event that the system is restored from a snapshot. For example, if the amount of time between snapshots is large, there is a greater chance that one or more modifications were made in between snapshots. If the amount of time between snapshots is large, there is a smaller chance that modifications were made in between snapshots.

Applicants have appreciated that in some embodiments, it may be advantageous to identify any content lost as a result of a file system failure. Identifying lost content may enable a determination of the scope of the loss and/or enable re-creation of any lost content. Thus, in some embodiments if the OAS system experiences a failure (e.g., a file system corruption) that results in the loss of one or more content units, the OAS system may identify each content unit that was lost as a result of the failure. Though the content of the content units themselves may be still be lost, the OAS system can identify (e.g., via their object identifiers) those content units that were lost as a result of the failure, which can be advantageous for any of numerous reasons, examples of which are described above.

The identification of content units lost as a result of a failure in the OAS system may be accomplished in any suitable way, as the invention is not limited in this respect. In some embodiments, a snapshot of the file system may be periodically created (e.g., as shown in FIG. 3). The OAS system may track the object identifiers of content units stored to the OAS system in between snapshots. If a failure occurs that results in restoring the file system from the most recent valid snapshot, the OAS system may provide the object identifiers of the content units stored to the OAS system after the creation of the most recent valid snapshot.

For example, as shown in FIG. 4, illustrative OAS system 401 includes an OAS Interface 403 that stores content units in a file system 105 in response to requests generated by application program 409. As discussed above, file system 105 stores content received from OAS Interface 403 on physical storage device(s) 107. OAS system 401 further includes a snapshot facility 405 that periodically creates a snapshot of file system 105. As discussed above with respect to snapshot facility 203, snapshot facility 405 may create snapshots of the file system in any suitable way, using any suitable technique, and at any suitable interval.

When snapshot facility 405 creates a snapshot, it may request OAS Interface 403 to maintain a list 407 of object identifiers of content units stored after creation of the snapshot (or to identify such content units in any other suitable way). A file system check may then be performed on the snapshot to verify that it is internally consistent. If the snapshot is internally consistent, then the snapshot may be treated as the most recent valid snapshot and establish a new time period at which the file system is known to be good. If not, a new snapshot may be created for the relevant time period that will be internally consistent.

Thus, for example, snapshot facility 405 may create a snapshot of the file system at time T1 (FIG. 3). As discussed above, in some embodiments, snapshot facility 405 may request OAS Interface 403 to complete any in-progress write operations and suspend new write operations prior to creating the snapshot. Once snapshot facility 405 creates the snapshot, it may request that OAS Interface 403 maintain a list 407 of object identifiers written after creation of the most recent snapshot until creation of new more recent valid snapshot. In embodiments in which snapshot facility requests OAS Interface 403 to complete in-progress write operations and suspend new write operations, the snapshot facility may also instruct OAS Interface to resume processing write operations. The snapshot facility may then perform a file system check of snapshot to verify its consistency.

As discussed above, in some embodiments, the OAS Interface may control operation of the snapshot facility. Thus, for example, the OAS Interface may select the times at which a snapshot is to be created and, at each selected time, may complete any in-progress write operations, suspend new write operations, and instruct the snapshot facility to create a snapshot of the file system.

Further, it should be appreciated that performing a file system check of a snapshot to verify its consistency is optional and the invention is not limited in this respect. For example, in some embodiments, when a snapshot of the file system is created it may be assumed that the snapshot stores a file system in a consistent state.

OAS Interface 403 may store the object identifiers of contents stored to the OAS system after time T1 in list 407. If a failure occurs after time T1 but before time T2, the file system may be restored from the snapshot created at time T1 and the OAS system may provide the object identifiers in list 407 (e.g., to application program 409 or any other suitable entity) to identify the content units lost as a result of the failure.

Assuming that no failure occurs between times T1 and T2, snapshot facility 405 may create a new snapshot of the file system at time T2 and request the OAS Interface to create a new list of object identifiers of content units stored after time T2. After the snapshot is created, the snapshot facility may request the OAS system to create a new list of object identifiers of content units stored after time T2. In some embodiments, once it is determined, e.g. via a file system check, that the snapshot created at time T2 is internally consistent, the previous list of object identifiers (e.g., the list created between times T1 and T2) may be deleted. In other embodiments, the previous list of object identifiers may be kept and stored.

In the example of FIG. 4, application program 409 accesses OAS system 403 directly. However, the invention is not limited in this respect, as application program 409 may access OAS system 403 in any suitable way, including via one or more software layers. For example, in some embodiments, application program 409 may access OAS system 403 via an application program interface (API) that presents an interface that the application program may use to send access requests to OAS system 403, but shields the application program from the details of communicating with the OAS system, such as, for example, the protocols used to communicate with the OAS system.

It should be appreciated from the discussion above that maintaining information about which content units are stored after the creation of a snapshot provides an additional degree of protection over simply periodically creating a snapshot (as this information may assist in assessing the scope of lost data resulting from restoring from a snapshot or re-creating content units lost as a result of restoring from a snapshot). However, in some embodiments, there may be a trade off between this additional degree of protection and conserving processing and/or storage resources. For example, maintaining the information identifying content units stored after the creation of snapshot may use additional processing and storage resources. Thus, in some situations where it is desired to conserve processing and/or storage resources, it may be desired to simply periodically create snapshots of the file system without maintaining information about which content units are stored after the creation of a snapshot.

Applicants have appreciated that thought it may be useful to identify content units lost as a result of a failure, it may be desirable, in some situations, to be able to recover the lost content units, rather than simply identify them. Thus, in some embodiments of the invention, content units that are lost as a result of a failure (e.g., a file system corruption) of an OAS system may be recovered. This may be done in any suitable way, as the invention is not limited in this respect.

For example, in some embodiments, a snapshot of the file system may be periodically created, in a manner similar to that discussed above in connection with FIGS. 2 and 4. In addition, a new file system (referred to herein as a file system mirror) may be created and content units stored to the OAS system may be stored in both the original file system and in the file system mirror. Thus, if a failure occurs sometime after the creation of the most recent valid snapshot and the file system is restored using the most recent valid snapshot, content units stored to the OAS system after creation of the snapshot may be recovered from the file system mirror.

That is, as shown in FIG. 5, illustrative OAS system 501 includes an OAS Interface 503 that stores content units in two file systems (i.e., file system 105 and file system mirror 507). File system 105 and file system mirror 507 store the content of the content units as blocks on physical storage device(s) 107. In some embodiments, file system mirror 507 stores only those files stored in file system 105 after the creation of the snapshot, but does not store files stored in file system 105 before the creation of the snapshot. OAS Interface also includes a snapshot facility 505 that periodically creates a snapshot of file system 105 (e.g., in a manner similar to that discussed above in connection with snapshot facility 405 and snapshot facility 203). Because a file system check may be performed on the most recent file system snapshot to verify that the file system can be restored from this snapshot, it may be determined that the file system in its state at the time this snapshot was created can be recovered. Thus, in some embodiments, only content units stored after the creation of the most recent valid snapshot need be stored in the file system mirror.

As discussed above, it should be appreciated that performing a file system check of a snapshot to verify its consistency is optional and the invention is not limited in this respect. For example, in some embodiments, when a snapshot of the file system is created it may be assumed that the snapshot stores a file system in a consistent state.

Once snapshot facility 505 creates a snapshot of file system 105, it may send a request to OAS Interface 503 to mirror any writes to file system mirror 507 during the window between the creation of the most recent valid snapshot until the time at which the next snapshot is created and verified as internally consistent. That is, for any content unit that OAS Interface 503 stores in file system 105, OAS Interface 503 may also store the content unit in file system mirror 507. If a failure occurs resulting in restoring the file system from a snapshot, content units stored to the OAS system after the snapshot was created may be recovered from file system mirror 507.

OAS Interface 503 may mirror writes to file system mirror 507 in any suitable way, as the invention is not limited in this respect. For example, in some embodiments, when OAS Interface 503 receives a request to store a content unit that includes the content of the content unit, OAS Interface 503 may temporarily store the content of the content unit in a buffer and transmit a request to file system 105 to store the content in the buffer in a file. OAS Interface may then transmit a request to file system mirror 507 to store the content in the buffer in a file in file system mirror 507. In other embodiments, when OAS Interface 503 receives a request to store a content unit that includes the content of the content unit, the OAS Interface may temporarily store the content in a first buffer and then copy the content from the first buffer into a second buffer. The OAS Interface may then send a request to file system 105 to store the content of the first buffer in a file in file system 105 and may send a request to file system mirror 507 to store the content of the second buffer in a file in file system mirror 507.

File system mirror 507 may have any suitable directory structure, as the invention is not limited in this respect. For example, in some embodiments, file system mirror 507 may have a parallel directory structure to file system 105. That is, file system mirror 507 may have a replica of the directory structure of file system 105, such that every directory in file system 105 has a parallel directory in file system mirror 507. If additional directories are created in file system 105 after a snapshot is created, those same directories may be created in file system mirror 507. Thus, when the OAS Interface stores a content unit as a file, having a particular filename, in a particular directory in file system 105, the OAS Interface may store the content units as a file in the same directory in file system mirror 507 and give the file the same filename as the corresponding file in file system 105.

In some embodiments, file system mirror 507 need not have the same directory structure as file system 105. For example, in some embodiments, file system mirror 507 may have a flat (i.e., as opposed to hierarchical directory structure) directory structure and all content units stored in file system mirror 507 may be stored in a single directory. When the OAS Interface stores the content of a content unit in a file in file system mirror 507, OAS Interface may give the file a filename that identifies the filename and file system location of the corresponding file in file system 105.

In some embodiments OAS Interface 503 may control the snapshot facility. Thus, for example, the OAS Interface may select the times at which a snapshot is to be created and, at each selected time, may complete any in-progress write operations, suspend new write operations, and instruct the snapshot facility to create a snapshot of the file system.

As discussed above, a failure may occur after creation of a snapshot that results in the OAS system restoring file system 105 using the snapshot and file system mirror 507. This restoration may be performed in any suitable way, as the invention is not limited in this respect. For example, if a failure occurs in between time T1 and time T2 (FIG. 3), the OAS system may restore file system 105 using the snapshot created at time T1 (e.g., by copying the blocks from the snapshot to the file system blocks, or in any other suitable way). As a result, file system 105 is restored to the same state that it was in at time T1, but does not have any of the files stored in file system 105 between time T1 and time T2. However, any files stored in file system 105 between time T1 and time T2 are stored in file system mirror 507. Thus, any such files in file system mirror 507 may be copied from the file system mirror into the restored file system. This may be done in any suitable way, as the invention is not limited in this respect.

For example, in some embodiments in which file system mirror 507 includes a parallel directory structure to file system 105, files stored in file system mirror 507 may be located and copied to the parallel location in file system 105. In embodiments in which the filename and location of a file in file system 105 is encoded in the filename of the corresponding file in file system mirror 507, each file in file system mirror 507 may be copied to the location in file system 105 encoded in its filename and may be stored in a file at this location with the filename encoded in the filename of the corresponding file in file system 105.

When a file system mirror is no longer needed, any suitable action may be taken with respect to it. For example, if a snapshot of file system 105 is created at time T1, file system mirror 507 may store files stored in file system 105 starting at time T1 up until the next valid snapshot of file system 105 is created (e.g., at time T2). At time T2, a new file system mirror may be created in which the OAS Interface stores any files stored in file system 105 between times T2 and T3 (when a new snapshot is created). Thus, the previous file system mirror is no longer needed. Thus, in some embodiments, the previous file system mirror may discarded. In other embodiments, the previous file system mirror may be maintained.

Applicants have appreciated that in some situations, when a file is deleted from a file system it is desired that the content of the file no longer exist on the storage system, particularly in situations where the file is “shredded” upon deletion. Shredding of a file refers to deleting the logical file from the file system, but also deleting the physical content of the file on the underlying physical storage device(s) by overwriting the blocks at which the physical content is stored with other data (e.g., all zeros or one or more patterns of zeros and ones designed to hinder data recovery).

Applicants have further appreciated that when a snapshot of a file system is created and a file in the file system is later shredded, the content of the file may still exist in the snapshot, even though it may be desired that the content of the file no longer exist anywhere on the storage system.

Thus, some embodiments of the invention are directed to inhibiting recovery of deleted content from a snapshot. In some embodiments this is achieved by encrypting content stored in a snapshot. Only the encrypted snapshot is stored, such that the content of any file that is shredded after creation of the content unit exists on the storage system (i.e., in the snapshot) only in encrypted form.

For example, as shown in FIG. 6, storage system 601 includes a file system 105, a snapshot facility 603, and encryption/decryption utility 605. Storage system 601 may be any suitable type of storage system, as the invention is not limited in this respect. For example, storage system 601 may be a block I/O storage system, an OAS system, or any other suitable type of storage system.

In some embodiments, a user and/or administrator of storage system 601 may configure the storage system to shred any content that is deleted. That is, for example, if a file in file system 105 is deleted from the file system, the underlying data of the file stored on the physical storage device(s) of the storage system may be deleted. In some embodiments, the user and/or administrator may configure the storage system to shred all deleted content or only particular subsets of the content. For example, a user or administrator might configure the storage system to shred deleted files that store word processing documents, but not to shred deleted files that include e-mails.

In embodiments in which the storage system is configured to shred deleted content, the snapshot facility may use encryption/decryption utility 605 to encrypt any content that is moved or copied from file system 105 and onto a snapshot location (e.g., the physical storage device(s) of storage system 601).

When snapshot facility 603 creates a snapshot of file system 105, snapshot facility 603 may pass the snapshot to encryption/decryption utility 605. Utility 605 may encrypt the snapshot and store it on the storage system. Utility 605 may encrypt the snapshot in any suitable way. For example, utility 605 may encrypt the snapshot using an encryption key that it generated or that it received from a user or administrator of the storage system.

In embodiments in which the snapshot facility creates a full copy of the file system, encryption/decryption utility 605 may encrypt the entire copy using the encryption key. In embodiments in which the snapshot facility uses a copy-on-first-write technique, only those blocks copied on a write command may be encrypted (e.g., using the encryption key). In this manner, data deleted from file system 105 is not duplicated in the snapshot.

In some embodiments, when an encrypted file system snapshot is created using encryption/decryption utility 605, a file system check utility 607 may be internally run within storage system 601. This may be done in any suitable way, as the invention is not limited in this respect. For example, the file system may access the physical disks via snapshot facility 603. That is, in response to a request from the file system, snapshot facility 603 may employ encryption/decryption utility 605 to access the physical storage and decrypt the data. Once the encryption/decryption utility 605 has decrypted the content on the snapshot version of the file system, file system check utility 607 can then verify the integrity of the snapshot version of the file system. It should be appreciated, however, that the invention is not limited to performing a file system check on the snapshot of the file system as, in some embodiments, it may be assumed that the snapshot is a valid snapshot.

In some situations, it may be desired to restore file system 105 from the encrypted snapshot version. This may be done in any suitable way, as the invention is not limited in this respect. For example, in some embodiments, utility 605 may retrieve the snapshot from the physical storage device(s). That is, snapshot facility 603 may be unable to directly read the physical blocks that store the snapshot version of the file system because the content of the snapshot is encrypted. Thus, the snapshot facility may employ encryption/decryption utility 605 to decrypt the content using the encryption key and return the unencrypted data to snapshot facility 603. Snapshot facility 603 may use the unencrypted snapshot to restore the file system.

In some embodiments, encryption/decryption utility 605 may be used exclusively by the internal file system check utility 607 and snapshot facility 603, and any attempt by another entity (such as an application program) may result in the entity being unable to access decrypted content. This may be done to protect the encrypted content from unauthorized access, and can be accomplished in any number of ways. For example, this may be accomplished by blocking application program access to the snapshot version of the file system and/or returning encrypted content to any entity other than the file system check utility and/or snapshot facility. Further, because the content stored thereon is encrypted, it may be difficult to access the content by physically removing the physical storage device(s) and reading the content directly from the physical devices.

The above-described embodiments of the present invention can be implemented on any suitable computer, and a system employing any suitable type of storage system.

Examples of suitable computers and/or storage systems are described in the patent applications listed below in Table 1 (collectively “the OAS applications”), each of which is incorporated herein by reference. It should be appreciated that the computers and storage systems described in these applications are only examples of computers and storage systems on which the embodiments of the present invention may be implemented, as the aspects of the invention described herein are not limited to being implemented in any particular way.

TABLE 1
Title Ser. No. Filing Date
Content Addressable 09/236,366 Jan. 21, 1999
Information, Encapsulation,
Representation, And
Transfer
Access To Content 09/235,146 Jan. 21, 1999
Addressable Data Over A
Network
System And Method For 09/391,360 Sep. 7, 1999
Secure Storage Transfer
And Retrieval Of Content
Addressable Information
Method And Apparatus For 10/731,790 Dec. 9, 2003
Data Retention In A
Storage System
Methods And Apparatus 10/731,613 Dec. 9, 2003
For Facilitating Access To
Content In A Data Storage
System
Methods And Apparatus 10/731,796 Dec. 9, 2003
For Caching A Location
Index In A Data Storage
System
Methods And Apparatus 10/731,603 Dec. 9, 2003
For Parsing A Content
Address To Facilitate
Selection Of A Physical
Storage Location In A Data
Storage System
Methods And Apparatus 10/731,845 Dec. 9, 2003
For Generating A Content
Address To Indicate Data
Units Written To A Storage
System Proximate In Time
Methods And Apparatus 10/762,044 Jan. 21, 2004
For Modifying A Retention
Period For Data In A
Storage System
Methods And Apparatus 10/761,826 Jan. 21, 2004
For Extending A Retention
Period For Data In A
Storage System
Methods And Apparatus 10/762,036 Jan. 21, 2004
For Indirectly Identifying A
Retention Period For Data
In A Storage System
Methods And Apparatus 10/762,043 Jan. 21, 2004
For Indirectly Identifying A
Retention Period For Data
In A Storage System
Methods And Apparatus 10/787,337 Feb. 26, 2004
For Increasing Data Storage
Capacity
Methods And Apparatus 10/787,670 Feb. 26, 2004
For Storing Data In A
Storage Environment
Methods And Apparatus 10/910,985 Aug. 4, 2004
For Segregating A Content
Addressable Computer
System
Methods And Apparatus 10/911,330 Aug. 4, 2004
For Accessing Content In A
Virtual Pool On A Content
Addressable Storage
System
Methods and Apparatus For 10/911,248 Aug. 4, 2004
Including Storage System
Capability Information In
An Access Request To A
Content Addressable
Storage System
Methods And Apparatus 10/911,247 Aug. 4, 2004
For Tracking Content
Storage In A Content
Addressable Storage
System
Methods and Apparatus For 10/911,360 Aug. 4, 2004
Storing Information
Identifying A Source Of A
Content Unit Stored On A
Content Addressable
System
Software System For 11/021,892 Dec. 23, 2004
Providing Storage System
Functionality
Software System For 11/022,022 Dec. 23, 2004
Providing Content
Addressable Storage
System Functionality
Methods And Apparatus 11/022,077 Dec. 23, 2004
For Providing Data
Retention Capability Via A
Network Attached Storage
Device
Methods And Apparatus 11/021,756 Dec. 23, 2004
For Managing Storage In A
Computer System
Methods And Apparatus 11/021,012 Dec. 23, 2004
For Processing Access
Requests In A Computer
System
Methods And Apparatus 11/021,378 Dec. 23, 2004
For Accessing Information
In A Hierarchical File
System
Methods And Apparatus 11/034,613 Jan. 12, 2005
For Storing A Reflection
On A Storage System
Method And Apparatus For 11/034,737 Jan. 12, 2005
Modifying A Retention
Period
Methods And Apparatus 11/034,732 Jan. 12, 2005
For Managing Deletion of
Data
Methods And Apparatus 11/107,520 Apr. 15, 2005
For Managing The Storage
Of Content
Methods And Apparatus 11/107,063 Apr. 15, 2005
For Retrieval Of Content
Units In A Time-Based
Directory Structure
Methods And Apparatus 11/107,194 Apr. 15, 2005
For Managing The
Replication Of Content
Methods And Apparatus 11/165,104 Jun. 23, 2005
For Managing the Storage
Of Content In A File
System
Methods And Apparatus 11/165,103 Jun. 23, 2005
For Accessing Content
Stored In A File System
Methods And Apparatus 11/165,102 Jun. 23, 2005
For Storing Content In A
File System
Methods And Apparatus 11/212,898 Aug. 26, 2005
For Managing the Storage
of Content
Methods And Apparatus 11/213,565 Aug. 26, 2005
For Scheduling An Action
on a Computer
Methods And Apparatus 11/213,233 Aug. 26, 2005
For Deleting Content From
A Storage System
Method and Apparatus For 11/324,615 Jan. 3, 2006
Managing The Storage Of
Content
Method and Apparatus For 11/324,639 Jan. 3, 2006
Providing An Interface To
A Storage System
Methods And Apparatus 11/324,533 Jan. 3, 2006
For Managing A File
System On A Content
Addressable Storage
System
Methods And Apparatus 11/324,637 Jan. 3, 2006
For Creating A File System
Methods And Apparatus 11/324,726 Jan. 3, 2006
For Mounting A File
System
Methods And Apparatus 11/324,642 Jan. 3, 2006
For Allowing Access To
Content
Methods And Apparatus 11/324,727 Jan. 3, 2006
For Implementing A File
System That Stores Files
On A Content Addressable
Storage System
Methods And Apparatus 11/324,728 Jan. 3, 2006
For Reconfiguring A
Storage System
Methods And Apparatus 11/324,646 Jan. 3, 2006
For Increasing The Storage
Capacity Of A Storage
System
Methods And Apparatus 11/324,644 Jan. 3, 2006
For Accessing Content On
A Storage System
Methods And Apparatus 11/392,969 Mar. 28, 2006
For Transferring Content
From A Storage System
Methods And Apparatus 11/391,654 Mar. 28, 2006
For Requesting Content
From A Storage System
Methods And Apparatus 11/392,981 Mar. 28, 2006
For Transferring Content
To Multiple Destinations
Methods And Apparatus 11/390,878 Mar. 28, 2006
For Receiving Content
From A Storage System At
Multiple Servers
Methods And Apparatus 11/390,564 Mar. 28, 2006
For Transferring Content
From An Object
Addressable Storage
System
Methods And Apparatus 11/391,636 Mar. 28, 2006
For Requesting Content
From An Object
Addressable Storage
System
Methods And Apparatus 11/438,770 May 23, 2006
For Conversion Of Content
Methods And Apparatus 11/439,025 May 23, 2006
For Selecting A Data
Format For A Content Unit
Methods And Apparatus 11/439,022 May 23, 2006
For Accessing A Content
Unit On A Storage System
Methods And Apparatus 11/438,817 May 23, 2006
For Enabling Selection Of
A Content Unit Data
Format
Methods And Apparatus 11/474,658 Jun. 26, 2006
For Accessing Content
Methods And Apparatus 11/474,846 Jun. 26, 2006
For Providing Access To
Content
Methods And Apparatus 11/474,655 Jun. 26, 2006
For Retrieving Stored
Content
Methods And Apparatus 11/474,661 Jun. 26, 2006
For Accessing Content
Through Multiple Nodes
Methods And Apparatus 11/474,719 Jun. 26, 2006
For Receiving Content
Methods And Apparatus 11/474,749 Jun. 26, 2006
For Processing Access
Requests
Methods And Apparatus 11/474,802 Jun. 26, 2006
For Providing Content
Methods And Apparatus 11/483,465 Jul. 10, 2006
For Managing Content
Methods And Apparatus 11/483,799 Jul. 10, 2006
For Moving Content
Methods And Apparatus 11/483,494 Jul. 10, 2006
For Storing Content
Methods And Apparatus 11/519,374 Sep. 12, 2006
For Caching Content In A
Computer System
Employing Object
Addressable Storage
Methods And Apparatus 11/644,430 Dec. 22, 2006
For Selection Of A Storage
Location For A Content
Unit
Methods And Apparatus 11/644,423 Dec. 22, 2006
For Modifying An Object
Identifier For A Content
Unit
Methods And Apparatus 11/644,174 Dec. 22, 2006
For Storing Content On A
Storage System
Methods And Apparatus 11/644,857 Dec. 22, 2006
For Increasing The Storage
Capacity Of A Zone Of A
Storage System
Methods And Apparatus 11/644,428 Dec. 22, 2006
For Selecting A Storage
Zone For A Content Unit

The above-described embodiments of the present invention can be implemented in any of numerous ways. For example, the embodiments may be implemented using hardware, software or a combination thereof. When implemented in software, the software code can be executed on any suitable processor or collection of processors, whether provided in a single computer or distributed among multiple computers. It should be appreciated that any component or collection of components that perform the functions described above can be generically considered as one or more controllers that control the above-discussed functions. The one or more controllers can be implemented in numerous ways, such as with dedicated hardware, or with general purpose hardware (e.g., one or more processors) that is programmed using microcode or software to perform the functions recited above.

In this respect, it should be appreciated that one implementation of the embodiments of the present invention comprises at least one computer-readable medium (e.g., a computer memory, a floppy disk, a compact disk, a tape, etc.) encoded with a computer program (i.e., a plurality of instructions), which, when executed on a processor, performs the above-discussed functions of the embodiments of the present invention. The computer-readable medium can be transportable such that the program stored thereon can be loaded onto any computer environment resource to implement the aspects of the present invention discussed herein. In addition, it should be appreciated that the reference to a computer program which, when executed, performs the above-discussed functions, is not limited to an application program running on a host computer. Rather, the term computer program is used herein in a generic sense to reference any type of computer code (e.g., software or microcode) that can be employed to program a processor to implement the above-discussed aspects of the present invention.

It should be appreciated that in accordance with several embodiments of the present invention wherein processes are implemented in a computer readable medium, the computer implemented processes may, during the course of their execution, receive input manually (e.g., from a user).

The phraseology and terminology used herein is for the purpose of description and should not be regarded as limiting. The use of “including,” “comprising,” “having,” “containing”, “involving”, and variations thereof, is meant to encompass the items listed thereafter and additional items.

Having described several embodiments of the invention in detail, various modifications and improvements will readily occur to those skilled in the art. Such modifications and improvements are intended to be within the spirit and scope of the invention. Accordingly, the foregoing description is by way of example only, and is not intended as limiting. The invention is limited only as defined by the following claims and the equivalents thereto.

Todd, Stephen J., Armangau, Philippe, Kilian, Michael

Patent Priority Assignee Title
10311153, Nov 28 2014 Nasuni Corporation Versioned file system with global lock
10417193, May 24 2016 VMware LLC Distributed file system consistency check
10481894, Feb 12 2016 Nutanix, Inc. Virtualized file server upgrade
10481895, Feb 12 2016 Nutanix, Inc. Virtualized file server rolling upgrade
10489140, Feb 12 2016 Nutanix, Inc. Virtualized file server high availability
10496390, Feb 12 2016 Nutanix, Inc. Virtualized file server resilience
10496391, Feb 12 2016 Nutanix, Inc. Virtualized file server block awareness
10540164, Feb 12 2016 Nutanix, Inc. Virtualized file server upgrade
10540165, Feb 12 2016 Nutanix, Inc. Virtualized file server rolling upgrade
10540166, Feb 12 2016 Nutanix, Inc. Virtualized file server high availability
10719305, Feb 12 2016 Nutanix, Inc. Virtualized file server tiers
10719306, Feb 12 2016 Nutanix, Inc. Virtualized file server resilience
10719307, Feb 12 2016 Nutanix, Inc. Virtualized file server block awareness
10728090, Dec 02 2016 Nutanix, Inc Configuring network segmentation for a virtualization environment
10809998, Feb 12 2016 Nutanix, Inc. Virtualized file server splitting and merging
10824455, Dec 02 2016 Nutanix, Inc Virtualized server systems and methods including load balancing for virtualized file servers
10831465, Feb 12 2016 Nutanix, Inc Virtualized file server distribution across clusters
10838708, Feb 12 2016 Nutanix, Inc. Virtualized file server backup to cloud
10949192, Feb 12 2016 Nutanix, Inc. Virtualized file server data sharing
11048595, Dec 05 2016 Nutanix, Inc Disaster recovery for distributed file servers, including metadata fixers
11086826, Apr 30 2018 Nutanix, Inc Virtualized server systems and methods including domain joining techniques
11106447, Feb 12 2016 Nutanix, Inc. Virtualized file server user views
11194680, Jul 20 2018 Nutanix, Inc.; Nutanix, Inc Two node clusters recovery on a failure
11218418, May 20 2016 Nutanix, Inc. Scalable leadership election in a multi-processing computing environment
11281484, Dec 06 2016 Nutanix, Inc Virtualized server systems and methods including scaling of file system virtual machines
11288239, Dec 06 2016 Nutanix, Inc Cloning virtualized file servers
11294777, Dec 05 2016 Nutanix, Inc Disaster recovery for distributed file servers, including metadata fixers
11310286, May 09 2014 Nutanix, Inc. Mechanism for providing external access to a secured networked virtualization environment
11409510, Feb 12 2016 Nutanix, Inc. Virtualized file server
11537384, Feb 12 2016 Nutanix, Inc. Virtualized file server distribution across clusters
11544049, Feb 12 2016 Nutanix, Inc. Virtualized file server disaster recovery
11550557, Feb 12 2016 Nutanix, Inc. Virtualized file server
11550558, Feb 12 2016 Nutanix, Inc. Virtualized file server deployment
11550559, Feb 12 2016 Nutanix, Inc. Virtualized file server rolling upgrade
11562034, Dec 02 2016 Nutanix, Inc Transparent referrals for distributed file servers
11568073, Dec 02 2016 Nutanix, Inc Handling permissions for virtualized file servers
11579861, Feb 12 2016 Nutanix, Inc. Virtualized file server smart data ingestion
11645065, Feb 12 2016 Nutanix, Inc. Virtualized file server user views
11669320, Feb 12 2016 Nutanix, Inc. Self-healing virtualized file server
11675746, Apr 30 2018 Nutanix, Inc. Virtualized server systems and methods including domain joining techniques
11768809, May 08 2020 Nutanix, Inc. Managing incremental snapshots for fast leader node bring-up
11770447, Oct 31 2018 Nutanix, Inc Managing high-availability file servers
11775397, Dec 05 2016 Nutanix, Inc. Disaster recovery for distributed file servers, including metadata fixers
11875176, Dec 06 2016 Virtualized server systems and methods including scaling of file system virtual machines
11888599, May 20 2016 Nutanix, Inc. Scalable leadership election in a multi-processing computing environment
11922157, Feb 12 2016 Nutanix, Inc. Virtualized file server
11922203, Dec 06 2016 Nutanix, Inc. Virtualized server systems and methods including scaling of file system virtual machines
11947952, Feb 12 2016 Nutanix, Inc. Virtualized file server disaster recovery
11954078, Dec 06 2016 Nutanix, Inc. Cloning virtualized file servers
11966729, Feb 12 2016 Nutanix, Inc. Virtualized file server
11966730, Feb 12 2016 Nutanix, Inc. Virtualized file server smart data ingestion
12072770, Aug 19 2021 Nutanix, Inc; Nutanix, Inc. Share-based file server replication for disaster recovery
12079621, Feb 12 2016 Virtualized file server distribution across clusters
12117972, Aug 19 2021 Nutanix, Inc File server managers and systems for managing virtualized file servers
12131192, Mar 18 2021 Nutanix, Inc. Scope-based distributed lock infrastructure for virtualized file server
12135963, Feb 12 2016 Nutanix, Inc. Virtualized file server distribution across clusters
12153690, Jan 24 2022 Nutanix, Inc Consistent access control lists across file servers for local users in a distributed file server environment
12153913, Feb 12 2016 Nutanix, Inc. Virtualized file server deployment
12164383, Aug 19 2021 Nutanix, Inc; Nutanix, Inc. Failover and failback of distributed file servers
8555342, Dec 23 2009 EMC IP HOLDING COMPANY LLC Providing secure access to a set of credentials within a data security mechanism of a data storage system
8621276, Dec 17 2010 Microsoft Technology Licensing, LLC File system resiliency management
8661063, Oct 12 2010 Nasuni Corporation Versioned file system with sharing
8880474, Jan 23 2009 Nasuni Corporation Method and system for interfacing to cloud storage
8990272, Jul 30 2010 Nasuni Corporation Versioned file system with pruning
9043614, Apr 30 2008 NetApp, Inc Discarding sensitive data from persistent point-in-time image
9152578, Mar 12 2013 EMC IP HOLDING COMPANY LLC Securing data replication, backup and mobility in cloud storage
9235596, Oct 12 2010 Nasuni Corporation Versioned file system with sharing
9274896, Aug 30 2010 Nasuni Corporation Versioned file system with fast restore
9575841, Jan 23 2009 Nasuni Corporation Method and system for interfacing to cloud storage
ER89,
Patent Priority Assignee Title
7254748, Oct 14 2003 AVAGO TECHNOLOGIES INTERNATIONAL SALES PTE LIMITED Error correcting content addressable memory
20050060631,
20070168777,
20070174662,
20070271482,
///////////////////////////////////////////////////////////////////////////////////////////
Executed onAssignorAssigneeConveyanceFrameReelDoc
May 21 2007TODD, STEPHEN J EMC CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0194110665 pdf
May 22 2007ARMANGAU, PHILIPPEEMC CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0194110665 pdf
May 24 2007KILIAN, MICHAELEMC CorporationASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0194110665 pdf
May 25 2007EMC Corporation(assignment on the face of the patent)
Sep 06 2016EMC CorporationEMC IP HOLDING COMPANY LLCASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0402030001 pdf
Sep 07 2016DELL SYSTEMS CORPORATIONTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016WYSE TECHNOLOGY L L C THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016WYSE TECHNOLOGY L L C CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016Spanning Cloud Apps LLCCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016SCALEIO LLCCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016EMC CorporationCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016MOZY, INC CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016ASAP SOFTWARE EXPRESS, INC CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016Aventail LLCCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016FORCE10 NETWORKS, INC CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016EMC IP HOLDING COMPANY LLCCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016Maginatics LLCCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016DELL SYSTEMS CORPORATIONCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016DELL SOFTWARE INC CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016Dell Products L PCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016DELL MARKETING L P CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016DELL INTERNATIONAL L L C CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016Dell USA L PCREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016CREDANT TECHNOLOGIES, INC CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENTSECURITY AGREEMENT0401340001 pdf
Sep 07 2016Spanning Cloud Apps LLCTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016SCALEIO LLCTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016ASAP SOFTWARE EXPRESS, INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016Aventail LLCTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016Dell USA L PTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016DELL INTERNATIONAL L L C THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016DELL MARKETING L P THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016Dell Products L PTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016DELL SOFTWARE INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016MOZY, INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016Maginatics LLCTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016FORCE10 NETWORKS, INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016EMC IP HOLDING COMPANY LLCTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016EMC CorporationTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Sep 07 2016CREDANT TECHNOLOGIES, INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSECURITY AGREEMENT0401360001 pdf
Mar 20 2019DELL INTERNATIONAL L L C THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Mar 20 2019DELL MARKETING L P THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Mar 20 2019Dell Products L PTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Mar 20 2019Dell USA L PTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Mar 20 2019FORCE10 NETWORKS, INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Mar 20 2019WYSE TECHNOLOGY L L C THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Mar 20 2019EMC IP HOLDING COMPANY LLCTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Mar 20 2019EMC CorporationTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Mar 20 2019CREDANT TECHNOLOGIES, INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0494520223 pdf
Apr 09 2020WYSE TECHNOLOGY L L C THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Apr 09 2020DELL MARKETING L P THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Apr 09 2020DELL INTERNATIONAL L L C THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Apr 09 2020CREDANT TECHNOLOGIES INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Apr 09 2020Dell Products L PTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Apr 09 2020Dell USA L PTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Apr 09 2020EMC CorporationTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Apr 09 2020FORCE10 NETWORKS, INC THE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Apr 09 2020EMC IP HOLDING COMPANY LLCTHE BANK OF NEW YORK MELLON TRUST COMPANY, N A SECURITY AGREEMENT0535460001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchDell Products L PRELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchDELL SOFTWARE INC RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchDELL SYSTEMS CORPORATIONRELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchEMC CorporationRELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchEMC IP HOLDING COMPANY LLCRELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchFORCE10 NETWORKS, INC RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchMaginatics LLCRELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchMOZY, INC RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchWYSE TECHNOLOGY L L C RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchDELL MARKETING L P RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchSCALEIO LLCRELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchDELL INTERNATIONAL, L L C RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchDell USA L PRELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchCREDANT TECHNOLOGIES, INC RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchAventail LLCRELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Nov 01 2021Credit Suisse AG, Cayman Islands BranchASAP SOFTWARE EXPRESS, INC RELEASE BY SECURED PARTY SEE DOCUMENT FOR DETAILS 0582160001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDell Products L PRELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDELL MARKETING CORPORATION SUCCESSOR-IN-INTEREST TO FORCE10 NETWORKS, INC AND WYSE TECHNOLOGY L L C RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTEMC CORPORATION ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO MAGINATICS LLC RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTEMC IP HOLDING COMPANY LLC ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO MOZY, INC RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSCALEIO LLCRELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDELL MARKETING L P ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO CREDANT TECHNOLOGIES, INC RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDELL MARKETING CORPORATION SUCCESSOR-IN-INTEREST TO ASAP SOFTWARE EXPRESS, INC RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDELL INTERNATIONAL L L C RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDell USA L PRELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDELL MARKETING L P ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO CREDANT TECHNOLOGIES, INC RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDell USA L PRELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDELL INTERNATIONAL L L C RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDell Products L PRELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDELL MARKETING CORPORATION SUCCESSOR-IN-INTEREST TO FORCE10 NETWORKS, INC AND WYSE TECHNOLOGY L L C RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTEMC CORPORATION ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO MAGINATICS LLC RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTEMC IP HOLDING COMPANY LLC ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO MOZY, INC RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTSCALEIO LLCRELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 045455 0001 0617530001 pdf
Mar 29 2022THE BANK OF NEW YORK MELLON TRUST COMPANY, N A , AS NOTES COLLATERAL AGENTDELL MARKETING CORPORATION SUCCESSOR-IN-INTEREST TO ASAP SOFTWARE EXPRESS, INC RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL FRAME 040136 0001 0613240001 pdf
Date Maintenance Fee Events
Jan 06 2014M1551: Payment of Maintenance Fee, 4th Year, Large Entity.
Dec 26 2017M1552: Payment of Maintenance Fee, 8th Year, Large Entity.
Dec 15 2021M1553: Payment of Maintenance Fee, 12th Year, Large Entity.


Date Maintenance Schedule
Jul 06 20134 years fee payment window open
Jan 06 20146 months grace period start (w surcharge)
Jul 06 2014patent expiry (for year 4)
Jul 06 20162 years to revive unintentionally abandoned end. (for year 4)
Jul 06 20178 years fee payment window open
Jan 06 20186 months grace period start (w surcharge)
Jul 06 2018patent expiry (for year 8)
Jul 06 20202 years to revive unintentionally abandoned end. (for year 8)
Jul 06 202112 years fee payment window open
Jan 06 20226 months grace period start (w surcharge)
Jul 06 2022patent expiry (for year 12)
Jul 06 20242 years to revive unintentionally abandoned end. (for year 12)