Abstract:
An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a base layer, an enhancement layer, or both. The processor is operationally coupled to the memory unit and is configured to provide a first indicator configured to enable or disable inter-layer syntax prediction and to provide a second indicator configured to enable or disable inter-layer texture prediction. The first and second indicators can be provided separately from each other. In one embodiment, the processor is further configured to disable inter-layer syntax prediction based at least in part upon information indicative of a codec associated with the base layer.
Abstract:
An apparatus configured to code (e.g., encode or decode) video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a base layer and an enhancement layer. The processor is configured to up-sample a base layer reference block by using an up-sampling filter when the base and enhancement layers have different resolutions; perform motion compensation interpolation by filtering the up-sampled base layer reference block; determine base layer residual information based on the filtered up-sampled base layer reference block; determine weighted base layer residual information by applying a weighting factor to the base layer residual information; and determine an enhancement layer block based on the weighted base layer residual information. The processor may encode or decode the video information.
Abstract:
Systems, methods, and devices for coding video data are described herein. In some aspects, a memory is configured to store the video data associated with a base layer and an enhancement layer. The base layer may comprise a reference block and base layer motion information associated with the reference block. The enhancement layer may comprise a current block. A processor operationally coupled to the memory is configured to determine a position of the base layer motion information in a candidate list based on a prediction mode in a plurality of prediction modes used at the enhancement layer. The processor is further configured to perform a prediction of the current block based at least in part on the candidate list.
Abstract:
An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video information associated with a reference layer and a corresponding enhancement layer. The processor determines a value of a video unit positioned at a position within the enhancement layer based at least in part on an intra prediction value weighted by a first weighting factor, wherein the intra prediction value is determined based on at least one additional video unit in the enhancement layer, and a value of a co-located video unit in the reference layer weighted by a second weighting factor, wherein the co-located video unit is located at a position in the reference layer corresponding to the position of the video unit in the enhancement layer. In some embodiments, the at least one of the first and second weighting factors is between 0 and 1.
Abstract:
Techniques for encoding and decoding video data are described. A method of coding video may include determining a plurality of motion vector candidates for a block of video data for use in a motion vector prediction process, wherein each of the motion vector candidates points to a respective reference frame index, performing the motion vector prediction process using the motion vector candidates to determine a motion vector for the block of video data, and performing motion compensation for the block of video data using the motion vector and a common reference frame index, wherein the common reference frame index is used regardless of the respective reference frame index associated with the determined motion vector.
Abstract:
In an example, aspects of this disclosure relate to a method for decoding a reference index syntax element in a video decoding process that includes decoding at least one bin of a reference index value with a context coding mode of a context-adaptive binary arithmetic coding (CABAC) process. The method also includes decoding, when the reference index value comprises more bins than the at least one bin coded with the context coded mode, at least another bin of the reference index value with a bypass coding mode of the CABAC process, and binarizing the reference index value.
Abstract:
Systems and techniques are provided for overlapped block motion compensation (OBMC). A method can include determining an OBMC mode is enabled for a current subblock of video data; for a neighboring subblock(s) adjacent to the current subblock, determining whether a first, second and third condition are met, the first condition comprising that all reference picture lists for predicting the current subblock are used to predict the neighboring subblock; the second condition comprising that identical reference pictures are used to determine motion vectors associated with the current subblock and the neighboring subblock, and the third condition comprising that a difference between motion vectors of the current subblock and the neighboring subblock do not exceed a threshold; and based on determining that the OBMC mode is enabled and the first, second, and third conditions are met, determining not to use motion information of the neighboring subblock for motion compensation of the current subblock.
Abstract:
Systems, methods, and media are provided for loop filtering across raster scan slices. One example includes obtaining the video data comprising one or more pictures and a first block of a picture having a pixel subject to filtering. A second block is determined to be located in the first slice in a particular relation to the second block. A third block that includes pixels for filtering the pixel is determined to be in a second slice at a diagonal corner of the first block, with filtering across slice boundaries disabled. First one or more pixels of the second block are identified as available for performing loop filtering of the pixel and second one or more pixels of the third block identified as unavailable for performing the loop filtering of the pixel of the first block. The first one or more pixels and the second one or more pixels are padded.
Abstract:
Systems and techniques are described for processing video data. For example, an apparatus can obtain a block of video data encoded using inter-prediction. In some examples, the apparatus can determine a direction associated with the block of video data, for instance based on pixels of the block of video data, based on pixels of at least one neighboring block of the block of video data, or based on information associated with a geometric partitioning mode (GPM) associated with the block of video data. The apparatus can apply a non-separable transform to the block of video data, for instance to decode or encode the video data. In some examples, the apparatus can apply the non-separable transform to the block according to the direction associated with the block of video data.
Abstract:
Systems and techniques are provided for processing video data. For example, the systems and techniques can include obtaining a current picture of video data and obtaining reference pictures for the current picture from the video data. A merge mode candidate can be determined for the current picture. First and second motion vectors can be identified for the merge mode candidate. A motion vector search strategy can be selected for the merge mode candidate from a plurality of motion vector search strategies. The selected motion vector search strategy can be associated with one or more constraints corresponding to at least one of the first motion vector or the second motion vector. The selected motion vector search strategy can be used to determine refined motion vectors based on the first motion vector, the second motion vector, and the reference pictures. The merge mode candidate can be processed using the refined motion vectors.