Abstract:
A device for decoding video data receives the video data, determines a scaling parameter for a block of the video data; and scales the block in a video decoding loop using the scaling parameter to increase a dynamic range for luminance values of the block. A device for encoding video data partitions the video data into blocks; determines a scaling parameter for a block of the video data; and scales the block in a video encoding loop using the scaling parameter to decrease a dynamic range for luminance values of the block.
Abstract:
Systems and methods for encoding a decoding video data are disclosed. The method for decoding video data can include receiving a bitstream having a plurality of layers of video data. The plurality of layers of video data can include a base layer and a video parameter set (VPS). The VPS can indicate a plurality of output layer sets (OLS) and profile, tier, and level (PTL) information for each OLS of the plurality of OLSs, including the 0th OLS. The method for decoding can also include selecting an OLS from the one or more layers as a target OLS, and decoding the target OLS based on PTL information signaled in the VPS.
Abstract:
Methods for defining decoder capability for decoding multi-layer bitstreams containing video information, in which the decoder is implemented based on multiple single-layer decoder cores, are disclosed. In one aspect, the method may include identifying at least one allocation of layers of the bitstream into at least one set of layers. The method may further include detecting whether each set of layers is capable of being exclusively assigned to one of the decoder cores for the decoding of the bitstream. The method may also include determining whether the decoder is capable of decoding the bitstream based at least in part on detecting whether each set of layers is capable of being exclusively assigned to one of the decoder cores.
Abstract:
In an example, a method of decoding video data includes decoding, by a video decoder, a first picture of multi-layer video data, where the first picture has a first picture type. The method also includes, based on the first picture having the first picture type, decoding, by the video decoder, data representative of picture order count (POC) most significant bits (MSB) value for the first picture.
Abstract:
A computing device generates a file that comprises a track box that contains metadata for a track in the file. Media data for the track comprises a sequence of samples. Each of the samples is a video access unit of multi-layer video data. As part of generating the file, the computing device generates, in the file, an additional box that documents all of the samples containing at least one Intra Random Access Point (IRAP) picture.
Abstract:
An example method of decoding video data includes receiving a plurality of pictures, receiving, in a message associated with a first picture of the plurality of pictures, information indicating a picture order count (POC) value of a recovery point picture of a gradual decoder refresh (GDR) set, when a picture that follows the first picture in decoding order has the POC value that is equal to a POC value of the recovery point picture, identifying the picture having the POC value equal to the POC value of the recovery point picture, as a recovery point picture, and when none of the pictures that follow the first picture has the POC value that is equal to the POC value of the recovery point picture, identifying one of the pictures that has a POC value greater than the POC value of the recovery point picture as the recovery point picture.
Abstract:
A video processing device includes a memory storing video data and one or more processors configured to: receive a first network abstraction layer (NAL) unit comprising a first picture of an access unit; in response to determining the first NAL unit comprises an intra random access point (IRAP) picture and in response to a NAL unit type for the first NAL unit indicating the presence of an instantaneous decoding refresh (IDR) picture without any associated leading pictures for a second NAL unit of the access unit comprising another IRAP picture, determine a NAL unit type for the second NAL unit to be a NAL unit type indicating the presence of an IDR picture without any associated leading pictures; and, process the first NAL unit and the second NAL unit based on the NAL unit type for the second NAL unit.
Abstract:
This disclosure relates to processing video data, including processing video data to conform to a high dynamic range (HDR)/wide color gamut (WCG) color container. The techniques apply, on an encoding side, pre-processing of color values prior to application of a static transfer function and/or apply post-processing on the output from the application of the static transfer function. By applying pre-processing, the examples may generate color values that when compacted into a different dynamic range by application of the static transfer function linearize the output codewords. By applying post-processing, the examples may increase signal to quantization noise ratio. The examples may apply the inverse of the operations on the encoding side on the decoding side to reconstruct the color values.
Abstract:
A device may determine, based on data in a bitstream, a luma sample (Y) of a pixel, a Cb sample of the pixel, and the Cr sample of the pixel. Furthermore, the device may obtain, from the bitstream, a first scaling factor and a second scaling factor. Additionally, the device may determine, based on the first scaling factor, the Cb sample for the pixel, and Y, a converted B sample (B′) for the pixel. The device may determine, based on the second scaling factor, the Cr sample for the pixel, and Y, a converted R sample (R′) for the pixel. The device may apply an electro-optical transfer function (EOTF) to convert Y′, R′, and B′ to a luminance sample for the pixel, a R sample for the pixel, and a B sample for the pixel, respectively.
Abstract:
Systems, methods, and devices for coding multilayer video data are disclosed that may include, encoding, decoding, transmitting, or receiving a non-entropy encoded set of profile, tier, and level syntax structures, potentially at a position within a video parameter set (VPS) extension. The systems, methods, and devices may refer to one of the profile, tier, and level syntax structures for each of a plurality of output layer sets. The systems, methods, and devices may encode or decode video data of one of the output layer sets based on information from the profile, tier, and level syntax structure referred to for the output layer set.