Abstract:
Certain embodiments described herein relate to an improved block-level replication system. One or more components in an information management system may receive a request to perform a block-level replication between a source storage device and a destination storage device, and depending on the specific replication mode requested, (i) store block-level changes directly to the destination storage device or (ii) first to a recovery point store and then later to the destination storage device.
Abstract:
An illustrative ISCSI server computing device provides user computing devices with “private writable snapshots” of a desired volume of data and/or further provides “private writable backup copies.” The ISCSI service is provided without invoking snapshot limits imposed by storage arrays and further without specialized backup software and pseudo-disk drivers installed on the user computing devices. A user can browse as well as edit personal versions of any number and/or versions of block-level backup copies—the “private writable backup copies.” Likewise, a user can browse and edit personal versions of any number of snapshots of one or more versions of one or more desired data volumes—the “private writable snapshots.” A user can have any number of co-existing private writable snapshots and private writable backup copies. Sparse files, extent-files, software snapshots, and/or media agents co-residing on the ISCSI server are used in the illustrative embodiments.
Abstract:
Systems and methods for performing file-level restore operations for block-level data volumes are described. In some embodiments, the systems and methods restore data from a block-level data volume contained in secondary storage by receiving a request to restore one or more files from the block-level data volume, mounting a virtual GUID Partition Table (GPT) disk to the block-level data volume, accessing one or more mount paths established by the virtual GPT disk between the data agent and the block-level data volume, and browsing data from one or more files within the block-level data volume via the established one or more mount paths provided by the virtual GPT disk.
Abstract:
Stubbing systems and methods are provided for intelligent data management in a replication environment, such as by reducing the space occupied by replication data on a destination system. In certain examples, stub files or like objects replace migrated, de-duplicated or otherwise copied data that has been moved from the destination system to secondary storage. Access is further provided to the replication data in a manner that is transparent to the user and/or without substantially impacting the base replication process. In order to distinguish stub files representing migrated replication data from replicated stub files, priority tags or like identifiers can be used. Thus, when accessing a stub file on the destination system, such as to modify replication data or perform a restore process, the tagged stub files can be used to recall archived data prior to performing the requested operation so that an accurate copy of the source data is generated.
Abstract:
A portable data protection system is described for protecting, transferring or copying data using continuous data protection (CDP) over intermittent or occasional connections between a computer system or mobile device containing the data to be protected, transferred or copied, called a data source, and one or more computer systems that receive the data, called a data target. CDP can be broken down logically into two phases: 1) detecting changes to data on a data source and 2) replicating the changes to a data target. The portable data protection system uses a method that performs the first phase continuously or near continuously on the data source, and the second phase when a connection is available between the data source and the data target.
Abstract:
Systems and methods for performing file-level restore operations for block-level data volumes are described. In some embodiments, the systems and methods restore data from a block-level data volume contained in secondary storage by receiving a request to restore one or more files from the block-level data volume, mounting a virtual disk to the block-level data volume, accessing one or more mount paths established by the virtual disk between the data agent and the block-level data volume, and browsing data from one or more files within the block-level data volume via the established one or more mount paths provided by the virtual disk.
Abstract:
An illustrative ISCSI server computing device provides user computing devices with “private writable snapshots” of a desired volume of data and/or further provides “private writable backup copies.” The ISCSI service is provided without invoking snapshot limits imposed by storage arrays and further without specialized backup software and pseudo-disk drivers installed on the user computing devices. A user can browse as well as edit personal versions of any number and/or versions of block-level backup copies—the “private writable backup copies.” Likewise, a user can browse and edit personal versions of any number of snapshots of one or more versions of one or more desired data volumes—the “private writable snapshots.” A user can have any number of co-existing private writable snapshots and private writable backup copies. Sparse files, extent-files, software snapshots, and/or media agents co-residing on the ISCSI server are used in the illustrative embodiments.
Abstract:
Systems and methods for performing file-level restore operations for block-level data volumes are described. In some embodiments, the systems and methods restore data from a block-level data volume contained in secondary storage by receiving a request to restore one or more files from the block-level data volume, mounting a virtual GUID Partition Table (GPT) disk to the block-level data volume, accessing one or more mount paths established by the virtual GPT disk between the data agent and the block-level data volume, and browsing data from one or more files within the block-level data volume via the established one or more mount paths provided by the virtual GPT disk.
Abstract:
Systems and methods for performing file-level restore operations for block-level data volumes are described. In some embodiments, the systems and methods restore data from a block-level data volume contained in secondary storage by receiving a request to restore one or more files from the block-level data volume, mounting a virtual GUID Partition Table (GPT) disk to the block-level data volume, accessing one or more mount paths established by the virtual GPT disk between the data agent and the block-level data volume, and browsing data from one or more files within the block-level data volume via the established one or more mount paths provided by the virtual GPT disk.
Abstract:
Systems and methods disclosed herein improve on current technology for block-level data replication to cloud computing environments. A new system architecture deploys one or more replication tail proxies in a cloud computing environment, locally (at the cloud) tracks replicated data and determines which replicated data meet criteria for reconstructing a desired point-in-time in the cloud, and persists data blocks received at the replication tail proxy until they are processed as recovery points. The disclosed approach presents resiliency and performance advantages. First, the resiliency of block-level data replication to cloud is improved by deploying the replication tail proxy in the destination cloud. Second, a Recovery Time Objective (RTO) is reduced by enabling faster cloud deployment of virtual machines for disaster recovery, failover, and/or test purposes based on the replicated data.