Abstract:
Methods, apparatuses, and computer-readable medium are provided for a frame rate up-conversion coding mode, in which an affine motion model is applied when conducting bilateral matching. The frame rate up-conversion coding mode can include generated additional frames from frames provided in a bitstream. In various implementations, bilateral matching includes, for a current block in a frame that is being generated, identifying a first block in a first reference picture a second block in a second reference picture. Affine (e.g., non-linear) motion information can be determined as between the first block and the second block. The current block can be predicted using the affine motion information.
Abstract:
An example device includes a memory to store the video data, and processing circuitry in communication with the memory. The processing circuitry is configured to compare a value of a dimension of a current block of the stored video data to a value of a corresponding dimension of a neighboring block of the current block to obtain a relative dimension value. The processing circuitry is further configured to determine, based on the relative dimension value, that the current block is to be partitioned according to a prediction tree (PT) portion of a multi-type tree-based partitioning scheme. The PT portion comprises partitioning according to one of a binary tree structure or a center-side triple tree structure. The processing circuitry is further configured to partition, based on the determination, the current block according to the PT portion, to form a plurality of sub-blocks.
Abstract:
An apparatus for coding video data according to certain aspects includes a memory and a processor in communication with the memory. The memory is configured to store video information, such as base layer video information and enhancement layer video information. The processor is configured to determine a value of a current video unit of enhancement layer video information based at least on a weighted inter-layer predictor and a weighted intra-layer predictor of at least one color component of the current video unit.
Abstract:
Techniques are described for signaling information used to generate three-dimensional (3D) color lookup tables for color gamut scalability in multi-layer video coding. A lower layer of video data may include color data in a first color gamut and a higher layer of the video data may include color data in a second color gamut. To generate inter-layer reference pictures, a video encoder and/or video decoder performs color prediction to convert the color data of a reference picture in the first color gamut to the second color gamut. The video coder may perform color prediction using a 3D lookup table. According to the techniques, a video encoder may encode partition information and/or color values of a 3D lookup table generated for color gamut scalability. A video decoder may decode the partition information and/or color values to generate the 3D lookup table in order to perform color gamut scalability.
Abstract:
An example method includes obtaining, for a current block of video data, values of motion vectors (MVs) of an affine motion model of a neighboring block of video data; deriving, from the values of the MVs of the affine motion model of the neighboring block, values of predictors for MVs of an affine motion model of the current block; decoding, from a video bitstream, a representation of differences between the values of the MVs of the affine motion model for the current block and the values of the predictors; determining the values of the MVs of the affine motion model for the current block from the values of the predictors and the decoded differences; determining, based on the determined values of the MVs of the affine motion model for the current block, a predictor block of video data; and reconstructing the current block based on the predictor block.
Abstract:
A method of decoding video data including receiving a bitstream that includes a sequence of bits that forms a representation of a coded picture of the video data, partitioning the coded picture of the video data into a plurality of blocks using three or more different partition structures, and reconstructing the plurality of blocks of the coded picture of the video data. Partitioning the coded picture of the video data may include partitioning the coded picture of the video data into the plurality of blocks using the three or more different partition structures, wherein at least three of the three or more different partition structures may be used at each depth of a tree structure that represents how a particular block of the coded picture of the video data is partitioned.
Abstract:
In one embodiment, an apparatus configured to encode video information includes a memory unit and a processor. The memory unit is configured to store an inter-layer reference picture comprising video information. The processor is operationally coupled to the memory unit. In addition, the processor is configured to partition the inter-layer reference picture into a quadtree structure having a plurality of leafs; determine inter-component filter parameters specific to each individual leaf based upon the video information in each individual leaf and signal the inter-component filter parameters for each of the leafs.
Abstract:
As part of a process of encoding video data, a video encoder applies a transform to a non-square residual block to generate a block of transform coefficients. Furthermore, the video encoder modifies the transform coefficients such that each respective transform coefficient of the block of transform coefficients is based on the respective dequantized transform coefficient multiplied by a factor of √{square root over (2)} or an approximation of √{square root over (2)}. As part of a process of decoding video data, a video decoder applies, by the video decoder, a dequantization process to transform coefficients of a non-square transform block. The video decoder modifies the dequantized transform coefficients such that each respective dequantized transform coefficient of the dequantized transform coefficients is based on the respective dequantized transform coefficient divided by a factor of √{square root over (2)} or an approximation of √{square root over (2)}.
Abstract:
An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a reference layer (RL) and an enhancement layer, the RL comprising an RL picture having an output region that includes a portion of the RL picture. The processor is configured to determine whether a condition indicates that information outside of the output region is available to predict a current block in the enhancement layer. The processor may encode or decode the video information.
Abstract:
Techniques are described for identifying and reducing the incidence of artifacts in video using color gamut scalability (CGS) parameters and tables in scalable video coding (SVC). Derivation of CGS mapping tables are performed for each partition of pixel values in a color space. The pixel value domain is split into partitions and each is optimized independently. Color prediction techniques for CGS may be used by video encoders and/or video decoders to generate inter-layer reference pictures when a color gamut for a lower layer of video data is different than a color gamut for a higher layer of the video data. When mapped values are used as inter-layer predication references for the enhancement layer blocks, artifacts may appear in some frames of the sequences. A video encoder may identify blocks that potentially contain these artifacts and disable inter-layer prediction in those identified blocks.