Abstract:
Techniques described herein are related to coding layer identifiers for operation points in video coding. In one example, a method of decoding video data is provided. The method comprises decoding syntax elements in a video parameter set (VPS) within a conforming bitstream indicating a first operation point having a first set of content. The method further comprises decoding, if present, syntax elements in the VPS within the conforming bitstream indicating hypothetical reference decoder (HRD) parameter information having a second set of content for the first operation point, wherein the conforming bitstream does not include syntax elements in the VPS that duplicate at least one of the first or second set of content for a second operation point, and wherein decoding syntax elements comprises decoding the syntax elements indicating the first operation point and the HRD parameter information only within conforming bitstreams.
Abstract:
This disclosure describes techniques for selection of coded picture buffer (CPB) parameters used to define a CPB for a video coding device for clean random access (CRA) pictures and broken link access (BLA) pictures in a video bitstream. A video coding device receives a bitstream including one or more CRA pictures or BLA pictures, and also receives a message indicating whether to use an alternative set of CPB parameters for at least one of the CRA pictures or BLA pictures. The message may be received from an external means, such as a processing means included in a streaming server or network entity. The video coding device sets a variable defined to indicate the set of CPB parameters for a given one of the pictures based on the received message, and selects the set of CPB parameters for the given one of the pictures based on the variable for the picture.
Abstract:
In one example, a device for decoding video data includes a video decoder configured to decode a value representative of a difference between most significant bits (MSBs) of a reference picture order count (POC) value and MSBs of a long-term reference picture (LTRP) POC value, wherein the reference POC value corresponds to a picture for which data must have been received in order to properly decode a current picture, determine the MSBs of the LTRP POC value based on the decoded value and the reference POC value, and decode at least a portion of the current picture relative to the LTRP based at least in part on the LTRP POC value. The picture for which data must have been received in order to properly decode a current picture may correspond to the current picture itself or a most recent random access point (RAP) picture.
Abstract:
This disclosure described techniques for coding layer dependencies for a block of video data. According to these techniques, a video encoder generates layer dependencies associated with a given layer. The video encoder also generates a type of prediction associated with one or more of the layer dependencies. In some examples, the video encoder generates a first syntax element to signal layer dependencies and a second syntax element to signal a type of prediction associated with one or more of the layer dependencies. A video decoder may obtain the layer dependencies associated with a given layer and the type of prediction associated with one or more of the layer dependencies.
Abstract:
An apparatus configured to code video information in a bitstream includes a memory and a processor in communication with the memory. The memory is configured to store video information associated with a plurality of video layers in the bitstream, the plurality of video layers in the bitstream divided into a plurality of bitstream partitions, herein each bitstream partition contains at least one of the plurality of video layers. The processor is configured to process a bitstream conformance parameter associated with a first bitstream partition of the plurality of bitstream partitions, wherein the bitstream conformance parameter is applicable to the first bitstream partition but not to another portion of the bitstream not encompassed by the first bitstream partition. The processor may encode or decode the video information in the bitstream.
Abstract:
According to certain aspects, an apparatus for coding video information includes a memory and a processor configured to determine whether a first syntax element is present in a bitstream, the first syntax element associated with a sequence parameter set (SPS) and a first flag indicative of whether a temporal identifier (ID) of a reference picture for pictures that refer to the SPS can be nested; and in response to determining that the first syntax element is not present in the bitstream: obtain a second syntax element indicative of a maximum number of temporal sub-layers in a particular layer of the plurality of layers; and determine whether to set the first flag equal to a second flag indicative of whether a temporal ID of a reference picture for any pictures can be nested based at least in part on a value of the second syntax element.
Abstract:
An apparatus for coding video information according to certain aspects includes a memory and a processor. The memory is configured to store video information associated with one or more layers. The processor is configured to code a current access unit (AU) in a bitstream including a plurality of layers, the plurality of layers including a reference layer and at least one corresponding enhancement layer. The processor is further configured to code a first end of sequence (EOS) network abstraction layer (NAL) unit associated with the reference layer in the current AU, the first EOS NAL unit having the same layer identifier (ID) as the reference layer. The processor is also configured to code a second EOS NAL unit associated with the enhancement layer in the current AU, the second EOS NAL unit having the same layer ID as the enhancement layer.
Abstract:
An apparatus according to certain aspects includes a memory and a processor configured to: in response to determining that a reference layer intra random access point (IRAP) picture starts a new coded video sequence (CVS) at the current AU: determine whether the reference layer IRAP picture activates a new video parameter set (VPS) or defines a prediction boundary at the current AU; in response to determining that the reference layer IRAP picture activates a new VPS or defines a prediction boundary at the current AU, remove pictures in the decoded picture buffer (DPB) associated with each of a plurality of layers; and in response to determining that the reference layer IRAP picture does not activate a new VPS and does not define a prediction boundary at the current AU, remove only pictures in the DPB associated with the reference layer.
Abstract:
An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a video layer having a picture. The processor is configured to determine whether the picture is a non-picture-order-count (POC)-anchor picture, and based on the determination of whether the picture is a non-POC-anchor picture, perform one of (1) refraining from indicating a POC reset in connection with the picture, or (2) indicating the POC reset in connection with the picture. The processor may encode or decode the video information.
Abstract:
An apparatus for decoding video information according to certain aspects includes a memory unit and a processor operationally coupled to the memory unit. The memory unit is configured to store at least one reference picture list of an enhancement layer, the at least one reference picture list comprising residual prediction reference picture information. The processor is configured to: decode signaled information about residual prediction reference picture generation; generate a residual prediction reference picture based on an enhancement layer reference picture and the decoded signaled information such that the generated residual prediction reference picture has the same motion field and the same picture order count (POC) as the enhancement layer reference picture from which it is generated; and store the generated residual prediction reference picture in the at least one reference picture list of the enhancement layer.