Abstract:
Techniques are described for encoding and decoding depth data for three-dimensional (3D) video data represented in a multiview plus depth format using depth coding modes that are different than high-efficiency video coding (HEVC) coding modes. Examples of additional depth intra coding modes available in a 3D-HEVC process include at least two of a Depth Modeling Mode (DMM), a Simplified Depth Coding (SDC) mode, and a Chain Coding Mode (CCM). In addition, an example of an additional depth inter coding mode includes an Inter SDC mode. In one example, the techniques include signaling depth intra coding modes used to code depth data for 3D video data in a depth modeling table that is separate from the HEVC syntax. In another example, the techniques of this disclosure include unifying signaling of residual information of depth data for 3D video data across two or more of the depth coding modes.
Abstract:
A method of coding video data includes upsampling at least a portion of a reference layer picture to an upsampled picture having an upsampled picture size. The upsampled picture size has a horizontal upsampled picture size and a vertical upsampled picture size. At least one of the horizontal or vertical upsampled picture sizes may be different than a horizontal picture size or vertical picture size, respectively, of an enhancement layer picture. In addition, position information associated with the upsampled picture may be signaled. An inter-layer reference picture may be generated based on the upsampled picture and the position information.
Abstract:
A video coder scales a motion vector of a current prediction unit (PU) of a current picture in order to compensate for a difference in temporal distance. In addition, the video coder determines a predictive block for the current PU, determines a disparity reference block based on samples of the disparity reference picture at a location indicated by a disparity vector of the current PU, and determines, based on samples of the fixed reference picture at a location indicated by the scaled motion vector, a temporal-disparity reference block for the current PU. The video coder then determines a residual predictor for the current PU. Each sample of the residual predictor for the current PU indicates a difference between a sample of the temporal-disparity reference block for the current PU and a corresponding sample of the disparity reference block.
Abstract:
In one example of the disclosure, a method of coding video data comprises coding video data using texture-first coding, and performing an NBDV derivation process for a block of the video data using a plurality of neighboring blocks. The NBDV derivation process comprises designating a motion vector associated with a neighboring block of the plurality of neighboring blocks coded with a block-based view synthesis prediction (BVSP) mode as an available disparity motion.
Abstract:
Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of multiview video data, wherein the track includes at least one depth view. The method further includes parsing information to determine a spatial resolution associated with the depth view, wherein decoding the spatial resolution does not require parsing of a sequence parameter set of the depth view. Another example method includes composing a track of multiview video data, wherein the track includes the one or more views. The example method further includes composing information to indicate a spatial resolution associated with the depth view, wherein decoding the spatial resolution does not require parsing of a sequence parameter set of the depth view.
Abstract:
In an example, a method of coding video data includes determining, for a first block of video data in a first layer of video data, a temporal motion vector and associated temporal reference picture for predicting the first block, where the temporal reference picture has a picture order count value. The method also includes determining a disparity reference block in a disparity reference picture indicated by a disparity vector associated with the first block, and determining whether a decoded picture buffer contains a temporal-disparity reference picture in the second view and having the picture order count value of the temporal reference picture. When the decoded picture buffer does not contain a temporal-disparity reference picture in the second view and having the picture order count value of the temporal reference picture, the method includes modifying an inter-view residual prediction process for predicting residual data of the first block.
Abstract:
A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include a dedicated array for parameter sets.
Abstract:
A video encoder generates, based on a reference picture set of a current view component, a reference picture list for the current view component. The reference picture set includes an inter-view reference picture set. The video encoder encodes the current view component based at least in part on one or more reference pictures in the reference picture list. In addition, the video encoder generates a bitstream that includes syntax elements indicating the reference picture set of the current view component. A video decoder parses, from the bitstream, syntax elements indicating the reference picture set of the current view component. The video decoder generates, based on the reference picture set, the reference picture list for the current view component. In addition, the video decoder decodes at least a portion of the current view component based on one or more reference pictures in the reference picture list.
Abstract:
The techniques of this disclosure may be generally related to using motion information for a corresponding block from a texture view component that corresponds with a block in a depth view component in coding the block in the depth view component. In some examples, for coding purposes, the techniques may use motion information when the spatial resolution of the texture view component is different than the spatial resolution of the depth view component.
Abstract:
In an example, a method of processing video data includes splitting a current block of video data into a plurality of sub-blocks for deriving motion information of the current block, where the motion information indicates motion of the current block relative to reference video data. The method also includes deriving, separately for each respective sub-block of the plurality of sub-blocks, motion information comprising performing a motion search for a first set of reference data that corresponds to a second set of reference data outside of each respective sub-block. The method also includes decoding the plurality of sub-blocks based on the derived motion information and without decoding syntax elements representative of the motion information.