Abstract:
An aspect of the invention is directed to a plurality of MDSs (metadata servers) in a distributed storage system which includes data servers storing file contents, each MDS having a processor and a memory and storing file system metadata. Directories of a file system namespace are distributed to the MDSs through referral directories referring to real directories using hash value of inode number of each of the referral directories. During a process to create a directory in the file system namespace, a first MDS dynamically creates a referral directory in the first MDS, and creates a real directory in a second MDS, the referral directory in the first MDS referring to the real directory in the second MDS by maintaining location information of the real directory in the second MDS, the real directory in the second MDS containing the file system metadata for the directory in the file system namespace.
Abstract:
A pool is formed based on a plurality of storage devices. This pool is constituted by a plurality of real pages. Real pages of different lengths are included in this plurality of real pages. Among a plurality of virtual pages which make up a virtual volume, a controller compresses a write data element for a write destination virtual page, selects a real page of a real page length based on the data length of a data unit including the compressed write data element, and allocates the selected real page to the write destination virtual page.
Abstract:
A computer system with a plurality of storage systems connected to each other via a network, each storage system including a virtual machine whose data is stored in hierarchized storage areas. When a virtual machine of a first storage system is migrated from the first storage system to a second storage system, the second storage system stores data of the virtual machine of the first storage system as well as data of its own virtual machine, in the hierarchized storage areas in the second storage system.
Abstract:
A pool is formed based on a plurality of storage devices. This pool is constituted by a plurality of real pages. Real pages of different lengths are included in this plurality of real pages. Among a plurality of virtual pages which make up a virtual volume, a controller compresses a write data element for a write destination virtual page, selects a real page of a real page length based on the data length of a data unit including the compressed write data element, and allocates the selected real page to the write destination virtual page.
Abstract:
The present invention provides a storage subsystem accessing a storage resource efficiently to thereby enhance throughput. The storage subsystem adopts a configuration where respective processor packages have ownership of the respective LDEVs, wherein an I/O command from a host specifying an address extending over a division position within a single LDEV is divided, and the I/O processing is executed in a shared manner by multiple processor packages. The I/O processing is executed by allocating a processor package that processes a specific storage area with priority.
Abstract:
First type metadata is associated with unstructured data included in an unstructured data source. A data processing system performs an extraction process. This extraction process includes: (a) creating, for each of a plurality of selected pieces of unstructured data in the unstructured data source, second type metadata, which is metadata including content information representing one or more content attributes of the piece of unstructured data; and (b) associating the created second type metadata with the first type metadata of the piece of unstructured data.
Abstract:
The present invention provides a storage subsystem accessing a storage resource efficiently to thereby enhance throughput. The storage subsystem adopts a configuration where respective processor packages have ownership of the respective LDEVs, wherein an I/O command from a host specifying an address extending over a division position within a single LDEV is divided, and the I/O processing is executed in a shared manner by multiple processor packages. The I/O processing is executed by allocating a processor package that processes a specific storage area with priority.
Abstract:
The present invention aims at improving the performance of a compression function in a storage system, and solves the prior art problem of having to decompress a whole compression unit even if a read request or a write request targets only a portion smaller than the compression unit, causing increase of overhead of decompression processing and elongation of processing time, and deteriorating performance. The present invention prevents unnecessary decompression processing and reduces the overhead of processing by suppressing the range of decompression processing to a minimum portion according to the relationship between the read/write request range and the compression unit.
Abstract:
A first storage system includes a plurality of first storage devices and is coupled to a computer. A second storage system includes a plurality of second storage devices and is coupled to the first storage system. A first controller provides a thin provisioning logical volume (LU) to the computer. A second controller provides an external thin provisioning LU to the first storage system. The first controller provides pool areas associated with the thin provisioning LU, including a first pool area mapped to the external thin provisioning LU, and allocates the first pool area to a first region in the thin provisioning LU to store a write data to the first region in the thin provisioning LU. The second controller allocates at least one of a plurality of pool areas to store the write data to the first region in the thin provisioning LU.
Abstract:
A first storage system includes a plurality of first storage devices and is coupled to a computer. A second storage system includes a plurality of second storage devices and is coupled to the first storage system. A first controller provides a thin provisioning logical volume (LU) to the computer. A second controller provides an external thin provisioning LU to the first storage system. The first controller provides pool areas associated with the thin provisioning LU, including a first pool area mapped to the external thin provisioning LU, and allocates the first pool area to a first region in the thin provisioning LU to store a write data to the first region in the thin provisioning LU. The second controller allocates at least one of a plurality of pool areas to store the write data to the first region in the thin provisioning LU.