Abstract:
In one example, an apparatus for context adaptive entropy coding may include a coder configured to determine one or more initialization parameters for a context adaptive entropy coding process based on one or more initialization parameter index values. The coder may be further configured to determine one or more initial context states for initializing one or more contexts of the context adaptive entropy coding process based on the initialization parameters. The coder may be still further configured to initialize the contexts based on the initial context states. In some examples, the initialization parameters may be included in one or more tables, wherein, to determine the initialization parameters, the coder may be configured to map the initialization parameter index values to the initialization parameters in the tables. Alternatively, the coder may be configured to calculate the initialization parameters using the initialization parameter index values and one or more formulas.
Abstract:
A video coder configured to perform sample adaptive offset filtering can determine a center value for a set of pixels based on values of pixels in the set, divide bands of pixels values into groups based on the center value, and determine offset values for the bands based on the groups.
Abstract:
A reduction in the number of binarizations and/or contexts used in context adaptive binary arithmetic coding (CABAC) for video coding is proposed. In particular, this disclosure proposes techniques that may lower the number contexts used in CABAC by up to 56.
Abstract:
In general, techniques are described for performing transform dependent de-blocking filtering, which may be implemented by a video encoding device. The video encoding device may apply a transform to a video data block to generate a block of transform coefficients, apply a quantization parameter to quantize the transform coefficients and reconstruct the block of video data from the quantized transform coefficients. The video encoding device may further determine at least one offset used in controlling de-blocking filtering based on the size of the applied transform, and perform de-blocking filtering on the reconstructed block of video data based on the determined offset. Additionally, the video encoder may specify a flag in a picture parameter set (PPS) that indicates whether the offset is specified in one or both of the PPS and a header of an independently decodable unit.
Abstract:
Techniques are described related to receiving first and second sub-sequences of video, wherein the first sub-sequence includes one or more frames each having a first resolution, and the second sub-sequence includes one or more frames each having a second resolution, receiving a first sequence parameter set and a second sequence parameter set for the coded video sequence, wherein the first sequence parameter set indicates the first resolution of the one or more frames of the first sub-sequence, and the second sequence parameter set indicates the second resolution of the one or more frames of the second sub-sequence, and wherein the first sequence parameter set is different than the second sequence parameter set, and using the first sequence parameter set and the second sequence parameter set to decode the coded video sequence.
Abstract:
A device for decoding video data can determine transform coefficients for a first block of a slice; derive a shift parameter for a second block of the slice based on the transform coefficients for the first block; receive one or more flags; determine a transform coefficient for the second block based on the one or more flags and based on the shift parameter; and output a decoded version of the second block based on the determined transform coefficient.
Abstract:
A method comprises: for each of a plurality of dimensions: identifying a reference position for the dimension, the reference position for the dimension being a position in a reference frame for the respective dimension, and the reference frame for the respective dimension and a reference frame for at least one other dimension in the plurality of dimensions being different reference frames in a plurality of reference frames; identifying an inter predictor for the respective dimension, wherein a predictor has a coordinate value in the respective dimension corresponding to a coordinate value in the respective dimension of the inter predictor for the respective dimension; and encoding or decoding the current point based on the predictor.
Abstract:
A device for encoding point cloud data, the device comprising: a memory to store the point cloud data; and one or more processors coupled to the memory and implemented in circuitry, the one or more processors configured to: determine a horizontal plane position of a node, wherein the horizontal plane position indicates a position of a single plane that is perpendicular to a first axis of a coordinate system, wherein the first axis is a horizontal axis; determine, from a plurality of contexts consisting of 8 contexts, a context for the horizontal plane position of the node; and perform arithmetic encoding on a syntax element indicating the horizontal plane position using the determined context.
Abstract:
Methods and devices for decoding video data are described. An example method includes in-loop filtering a current block of the video data using a neural network-based in-loop filter to generate an in-loop filtered current block, wherein the neural network-based in-loop filter is trained using an architecture comprising a U-Net architecture comprising one or more residual blocks and one or more transform blocks; and outputting the in-loop filtered current block.
Abstract:
A video coder may be configured to code a plurality of syntax elements that indicate an absolute value of a transform coefficients in a block, wherein the plurality of syntax elements include a significance flag and four or more greater_than_N flags. The video coder may further determine first respective contexts for each syntax element in a first set of the plurality of syntax elements from respective context tables for each syntax element in the first set of the plurality of syntax elements, and determine second respective contexts for each syntax element in a second set of the plurality of syntax elements from a shared context table.