Abstract:
An image encoding method that expresses an image using pixels or blocks, and colors assigned to the pixels or the blocks when encoding an image to be encoded, and encodes a block division tree and color information, the color information being assigned to all nodes including a root node, branch nodes, and leaf nodes of the block division tree, the method includes: a division tree generation step of dividing the image to be encoded using an arbitrary division tree and thereby generating the block division tree; a color information determination step of determining representative colors that represent interim blocks or final blocks and assigning the determined colors as color information corresponding to the respective nodes, the interim blocks or the final blocks corresponding to the nodes of the block division tree; a tree encoding step of encoding the block division tree; and a color information encoding step of encoding the color information.
Abstract:
When multiview images are encoded, a first viewpoint-synthesized image for an encoding target region is generated by using a reference viewpoint image from a viewpoint other than that of the encoding target image and a reference depth map for the reference viewpoint image. A second viewpoint-synthesized image for reference pixels is generated by using the first viewpoint-synthesized image, where the reference pixels are a set of already-encoded pixels, which are referred to in intra prediction of the encoding target region. An intra predicted image for the encoding target region is generated by using a decoded image for the reference pixels and the second viewpoint-synthesized image.
Abstract:
A video encoding apparatus is a video encoding apparatus which, when encoding an encoding target picture which is one frame of a multi-view video including videos of a plurality of different views, performs encoding while performing prediction between different views, for each of encoding target areas which are areas into which the encoding target picture is divided, using a reference view picture which is a picture for a reference view different from a view of the encoding target picture and a depth map for an object in the multi-view video, and includes a representative depth setting unit which sets a representative depth from the depth map, a transformation matrix setting unit which sets a transformation matrix that transforms a position on the encoding target picture into a position on the reference view picture based on the representative depth, a representative position setting unit which sets a representative position from a position within each of the encoding target areas, a disparity information setting unit which sets disparity information between the view of the encoding target and the reference view for each of the encoding target areas using the representative position and the transformation matrix, and a prediction picture generation unit which generates a prediction picture for each of the encoding target areas using the disparity information.
Abstract:
A video encoding apparatus that predictive-encodes an encoding target image included in an encoding target video. The apparatus includes a prediction device that predicts the encoding target image with reference to a previously-encoded picture as a reference picture and determines first reference information which indicates a first reference region as a reference destination; a second reference information determination device that determines, from a depth map corresponding to the first reference region, second reference information which indicates a second reference region as another reference destination for the encoding target image; and a predicted image generation device that generates a predicted image based on the second reference information, or both the first reference information and the second reference information.
Abstract:
A video encoding apparatus that predictive-encodes an encoding target image included in an encoding target video. The apparatus includes a prediction device that predicts the encoding target image with reference to a previously-encoded picture as a reference picture and determines first reference information which indicates a first reference region as a reference destination; a second reference information determination device that determines, from reference information used when the first reference region was predictive-encoded, second reference information which indicates a second reference region as another reference destination for the encoding target image; and a predicted image generation device that generates a predicted image based on the second reference information, or both the first reference information and the second reference information.
Abstract:
An image encoding device and an image decoding device encodes with a low overall output size while preventing encoding-efficiency degradation in occlusion regions. When encoding a multiview image comprising a plurality of images from different perspectives, this device, using a reference image from a different perspective from a target image being encoded and a reference depth map for a subject in said reference image, encodes while performing image prediction across different perspectives. The device is provided with: a combined-perspective-image generation unit that uses the reference image and reference depth map to generate a combined-perspective image for the target image; a usability determination unit that, for each encoding region determines whether or not the combined-perspective image is usable; and an image encoding unit that performs predictive encoding on the target image while selecting predicted-image generation methods for encoding regions for which the combined-perspective image was determined to be unusable.
Abstract:
According to an aspect of the present invention, there is provided a noise reducing device including: a provisional image reconstructing unit that generates a provisional image by reconstructing an encoded image obtained through compressed spectral imaging; a similar patch search unit that acquires a plurality of similar provisional patches that are small regions including images similar to each other in the provisional image; a group generating unit that acquires a plurality of similar encoded patches that are small regions including images similar to each other in the encoded image based on information on positions where the similar provisional patches have been acquired; a low-rank approximation unit that performs low-rank approximation based on the plurality of similar encoded patches to acquire a patch in which noise on an image has been reduced in a region included in the similar encoded patches; and a patch integrating unit that generates an encoded image by integrating a plurality of patches in which noise has been reduced, depending on information on positions of the patches.
Abstract:
A coding device includes: a position information coder configured to convert position information indicating a position of a point included in point cloud data into a tree structure, divide the tree structure in a depth direction to generate layer groups, divide at least one of the layer groups on a basis of the position information to generate subgroups, and code the subgroup to generate a position data unit corresponding to a part of the position information; a levels of detail structure generator configured to generate, for each attribute data unit corresponding to a part of attribute information indicating an attribute of a point included in the point cloud data, a detail level up to a layer overlapping with a parent attribute data unit in the depth direction, in order from a leaf side to a root side of the tree structure; and a prediction reference point determiner configured to determine, for each of the attribute data units, in order from the root side to the leaf side of the tree structure, for each point not subsampled when the detail level is generated, a prediction reference point from a layer same as a layer including the point not subsampled when the detail level is generated or a layer closer to the root side than the layer including the point not subsampled when the detail level is generated.
Abstract:
An imaging device includes an optical element including a transparent substrate and a plurality of structures disposed on or in the transparent substrate in a plane direction of the transparent substrate, an imaging sensor in which a plurality of pixels each including a photoelectric conversion element are arranged, and a signal processing unit configured to generate an image signal based on an electric signal obtained from the imaging sensor, wherein the optical element outputs light with a different point spread function for each wavelength to form, on the imaging sensor, an image in which the point spread function of each wavelength is convoluted, the plurality of structures have the same height in a side view, and the signal processing unit reconstructs an image in which the point spread function of each wavelength is convoluted.
Abstract:
Based on a representative depth determined from a depth map corresponding to an object in a multi-viewpoint video, a transformation matrix is determined which transforms a position on an encoding target image, which is one frame of the multi-viewpoint video, into a position on a reference viewpoint image from a reference viewpoint which differs from the viewpoint of the encoding target image. A representative position is determined which belongs to an encoding target region obtained by dividing the encoding target image. A corresponding position which corresponds to the representative position and belongs to the reference viewpoint image is determined by using the representative position and the transformation matrix. Based on the corresponding position, synthesized motion information assigned to the encoding target region is generated from motion information for the reference viewpoint image, and a predicted image for the encoding target region is generated by using the synthesized motion information.