Abstract:
A video decoder receives a value for a first syntax element representing whether a dependency type syntax element for a current layer is signaled, wherein the dependency type syntax element identifies a type of dependency of a current layer relative to a reference layer; and in response to the value for the first syntax element indicating that the dependency type syntax element is not signaled determines that the type of dependency of the current layer relative to the reference layer is a predetermined type and decodes a block of the current layer using inter-layer prediction conforming to the predetermined type.
Abstract:
A device may determine, based on a value, whether all cross-layer random access skipped (CL-RAS) pictures of an intra random access point (IRAP) access unit are present in a video data bitstream. In addition, the device may reconstruct pictures of the video data based at least in part on syntax elements decoded from the video data bitstream.
Abstract:
Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving a non-entropy encoded layer dependency information at a position within a video parameter set (VPS) extension prior to syntax elements of the VPS extension that are entropy encoded. The systems, methods, and devices may encode or decode the non-entropy encoded layer dependency information before an entropy encoded syntax element. The systems, methods, and devices may encode or decode video data of one or more of the layers of video data based on the non-entropy encoded layer dependency information. The layer dependency information indicates whether one of the layers is a direct reference layer for another of the layers.
Abstract:
Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving multilayer video data. The systems, methods, and devices may transmit or receive a video parameter set (VPS) including information for a series of layers, each layer including visual signal information. The systems, methods, and devices may code (encode or decode) video data based on the visual signal information signaled per layer in the VPS.
Abstract:
In one example, a device for coding video data includes a video coder configured to code, for a tile of an enhancement layer picture, data indicating a number of tiles in a base layer picture that need to be coded before the tile of the enhancement layer picture can be coded, code tiles of the base layer picture corresponding to the number of tiles, and, after coding the tiles of the base layer picture, code the tile of the enhancement layer picture substantially in parallel with at least one other tile in the base layer picture.
Abstract:
Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of video data, wherein the track includes one or more views. The method further includes parsing information to determine whether the track includes only texture views, only depth views, or both texture and depth views. Another example method includes composing a track of video data, wherein the track includes one or more views and composing information that indicates whether the track includes only texture views, only depth views, or both texture and depth views.
Abstract:
In one example, a device includes a video coder configured to code a multilayer bitstream comprising a plurality of layers of video data, where the plurality of layers of video data are associated with a plurality of layer sets, and where each layer set contains one or more layers of video data of the plurality of layers, and to code on one or more syntax elements of the bitstream indicating one or more output operation points, where each output operation point is associated with a layer set of the plurality of layer sets and one or more target output layers of the plurality of layers.
Abstract:
A device for processing video data receives an indication of a maximum layer identification (ID) value for a bitstream; receives a flag for a first layer with a layer ID value less than the maximum layer ID value; and, based on a value of the flag, determines if the first layer is included in an operation point.
Abstract:
A video coding device, such as a video encoder or a video decoder, may be configured to code a sub-picture timing supplemental enhancement information (SEI) message associated with a first decoding unit (DU) of an access unit (AU). The video coding device may further code a duration between coded picture buffer (CPB) removal time of a second DU of the AU in decoding order and CPB removal time of the first DU in the sub-picture SEI message. The coding device may also derive a CPB removal time of the first DU based at least in part on the sub-picture timing SEI message.
Abstract:
A video coding device, such as a video encoder or a video decoder, may be configured to decode a duration between coded picture buffer (CPB) removal time of a first decoding unit (DU) in an access unit (AU) and CPB removal time of a second DU, wherein the first DU comprises a non-video coding layer (VCL) network abstraction layer (NAL) unit with nal_unit_type equal to UNSPEC0, EOS_NUT, EOB_NUT, in the range of RSV_NVCL44 to RSV_NVCL47 or in the range of UNSPEC48 to UNSPEC63. The video decoder determines a removal time of the first DU based at least in part on the decoded duration and decodes video data of the first DU based at least in part on the removal time.