Abstract:
A device for processing three-dimensional (3D) video data may determine, based on direct dependent layers signaled in a video parameter set, that the current texture layer of the video data is dependent on a depth layer of the video data; and process the current texture layer using the depth layer.
Abstract:
In general, the disclosure describes techniques for coding depth blocks in video data. A video coding device may determine a size of a sub-prediction unit (sub-PU) for motion parameter inheritance (MPI) in a depth prediction unit (depth PU). The depth PU may comprise one or more sub-PUs, each sub-PU having equal size. Distinct motion information can be inherited for each sub-PU from a texture block co-located to the sub-PU. The device may encode each sub-PU based on MPI corresponding to texture blocks co-located with the sub-PUs and generate a syntax element that indicates the size of each sub-PU for MPI. The device may further receive the syntax element, and, upon receiving an index value that indicates a sub-PU MPI mode, inherit MPI for each sub-PU based on the corresponding texture blocks.
Abstract:
A computing device generates a file that comprises a track box that contains metadata for a track in the file. Media data for the track comprises a sequence of samples. Each of the samples is a video access unit of multi-layer video data. As part of generating the file, the computing device generates, in the file, an additional box that documents all of the samples containing at least one Intra Random Access Point (IRAP) picture.
Abstract:
In one example, a device for processing video data includes a memory for storing an enhancement layer of video data coded according to an extension of a video coding standard, and one or more processors configured to decode a hierarchy extension descriptor for an elementary stream including the enhancement layer, wherein the hierarchy extension descriptor includes data representative of two or more reference layers on which the enhancement layer depends, wherein the two or more reference layers include a first enhancement layer, conforming to a first scalability dimension, and a second enhancement layer, conforming to a second scalability dimension, and wherein the first scalability dimension is different than the second scalability dimension, and to process the video data based at least in part on the data representative of the two or more reference layers.
Abstract:
Techniques are described determining a partition pattern for intra-prediction encoding or decoding a depth block from a partition pattern of one or more partition patterns associated with smaller sized blocks. A video encoder may intra-prediction encode the depth block based on the determined partition pattern, and a video decoder may intra-prediction decode the depth block based on the determine partition pattern.
Abstract:
A computing device may obtain, from a first bitstream that includes a coded representation of the video data, a Supplemental Enhancement Information (SEI) message that includes an indication of an extraction mode that was used to produce the first bitstream. If the extraction mode is the first extraction mode, the first bitstream includes one or more coded pictures not needed for correct decoding of the target output layer set. If the extraction mode is the second extraction mode, the first bitstream does not include the one or more coded pictures not needed for correct decoding of the target output layer set.
Abstract:
A video coder generates a list of merging candidates for coding a video block of the 3D video. A maximum number of merging candidates in the list of merging candidates may be equal to 6. As part of generating the list of merging candidates, the video coder determines whether a number of merging candidates in the list of merging candidates is less than 5. If so, the video coder derives one or more combined bi-predictive merging candidates. The video coder includes the one or more combined bi-predictive merging candidates in the list of merging candidates.
Abstract:
Techniques are described for sub-prediction unit (PU) based motion prediction for video coding in HEVC and 3D-HEVC. In one example, the techniques include an advanced temporal motion vector prediction (TMVP) mode to predict sub-PUs of a PU in single layer coding for which motion vector refinement may be allowed. The advanced TMVP mode includes determining motion vectors for the PU in at least two stages to derive motion information for the PU that includes different motion vectors and reference indices for each of the sub-PUs of the PU. In another example, the techniques include storing separate motion information derived for each sub-PU of a current PU predicted using a sub-PU backward view synthesis prediction (BVSP) mode even after motion compensation is performed. The additional motion information stored for the current PU may be used to predict subsequent PUs for which the current PU is a neighboring block.
Abstract:
Techniques for decoding video data include receiving residual data corresponding to a block of video data, wherein the block of video data is encoded using asymmetric motion partitioning, is uni-directionally predicted using backward view synthesis prediction (BVSP), and has a size of 16×12, 12×16, 16×4 or 4×16, partitioning the block of video data into sub-blocks, each sub-block having a size of 8×4 or 4×8, deriving a disparity motion vector for each of the sub-blocks from a corresponding depth block in a depth picture corresponding to a reference picture, synthesizing a respective reference block for each of the sub-blocks using the respective derived disparity motion vector, and decoding the block of video data by performing motion compensation on each of the sub-blocks using the residual data and the synthesized respective reference blocks.
Abstract:
This disclosure describes techniques for in-loop depth map filtering for 3D video coding processes. In one example, a method of decoding video data comprises decoding a depth block corresponding to a texture block, receiving a respective indication of one or more offset values for the decoded depth block, and performing a filtering process on edge pixels of the depth block using at least one of the one or more offset values to create a filtered depth block.