Abstract:
A video encoder signals, in a bitstream, a syntax element that indicates whether a current video unit is predicted from a VSP picture. The current video unit is a macroblock or a macroblock partition. The video encoder determines, based at least in part on whether the current video unit is predicted from the VSP picture, whether to signal, in the bitstream, motion information for the current video unit. A video decoder decodes the syntax element from the bitstream and determines, based at least in part on the syntax element, whether the bitstream includes the motion information.
Abstract:
The techniques of this disclosure may be generally related to using motion information for a corresponding block from a texture view component that corresponds with a block in a depth view component in coding the block in the depth view component. In some examples, for coding purposes, the techniques may use motion information when the spatial resolution of the texture view component is different than the spatial resolution of the depth view component.
Abstract:
This disclosure describes features and techniques applicable to three-dimensional (3D) video coding. In one example, a technique may include coding a texture view video block, and coding a depth view video block, wherein the depth view video block is associated with the texture view video block. Coding the depth view video block may include coding a syntax element to indicate whether or not motion information associated with the texture view video block is adopted as motion information associated with the depth view video block.
Abstract:
In an example, a method of decoding video data includes selecting a motion information derivation mode from a plurality of motion information derivation modes for determining motion information for a current block, where each motion information derivation mode of the plurality comprises performing a motion search for a first set of reference data that corresponds to a second set of reference data outside of the current block, and where the motion information indicates motion of the current block relative to reference video data. The method also includes determining the motion information for the current block using the selected motion information derivation mode. The method also includes decoding the current block using the determined motion information and without decoding syntax elements representative of the motion information.
Abstract:
Examples include a device for coding video data, the device including a memory configured to store video data, and one or more processors configured to obtain adaptive loop filtering (ALF) information for a current coding tree unit (CTU) from one or more of: (i) one or more spatial neighbor CTUs of the current CTU or (ii) one or more temporal neighbor CTUs of the current CTU, to form a candidate list based at least partially on the obtained ALF information for the current CTU, and to perform a filtering operation on the current CTU using ALF information associated with a candidate from the candidate list. Coding video data includes encoding video data, decoding video data, or both encoding and decoding video data.
Abstract:
When coding multiview video data, a video encoder and video decoder may select a candidate picture from one of one or more random access point view component (RAPVC) pictures and one or more pictures having a lowest temporal identification value. The video encoder and video decoder may determine whether a block in the selected candidate picture is inter-predicted with a disparity motion vector and determine a disparity vector for a current block of a current picture based on the disparity motion vector. The video encoder and video decoder may inter-prediction encode or decode, respectively, the current block based on the determined disparity vector.
Abstract:
A video coder signals, in a bitstream, a syntax element that indicates whether inter-view/layer reference pictures are ever included in a reference picture list for a current view component/layer representation. A video decoder obtains, from the bitstream, the syntax element that indicates whether inter-view/layer reference pictures are ever included in a reference picture list for a current view component/layer representation. The video decoder decodes the current view component/layer representation.
Abstract:
Techniques are described for deriving a disparity vector for a current block based on a disparity motion vector of a neighboring block in a 3D-AVC video coding process. The disparity vector derivation allows for texture-first coding where a depth view component of a dependent view is coded subsequent to the coding of the corresponding texture component of the dependent view.
Abstract:
A system and method for decoding video. A first syntax element for a block of video data is received, a value of the first syntax element indicating one of a plurality of mapping functions to be used to determine a magnitude of a scaling parameter for cross-component prediction. A second syntax element for the block of video data is received, a value of the second syntax element corresponding to the magnitude of the scaling parameter, wherein receiving the second syntax element includes decoding the value of the second syntax element with a specific binarization method. The magnitude of the scaling parameter is determined using the one of the plurality of mapping functions indicated by the first syntax element and the value of the second syntax element. Cross-component prediction is performed for at least one component of the block video data using the determined magnitude of the scaling parameter.
Abstract:
In general, this disclosure describes techniques for coding video blocks using a color-space conversion process. A video coder, such as a video encoder or a video decoder, may determine whether to use color-space conversion for encoding the video data. In response to determining to use color-space conversion, the video coder may quantize data of a first color component of the video data using a first offset of a first quantization parameter (QP) and quantize data of a second color component of the video data using a second offset of a second QP, wherein the second color component is different than the first color component, and the second QP is different than the first QP. The video coder may further inverse quantize data of the first color component using the first offset and inverse quantize data of the second color component using the second offset.