Abstract:
A video coding device configured according to some aspects of this disclosure includes a memory configured to store a plurality of motion vector candidates. Each motion vector candidate can corresponding to at least one of a plurality of prediction units (PUs) partitioned in a parallel motion estimation region (MER). The video coding device also includes a processor in communication with the memory. The processor is configured to select a subset of the plurality of motion vector candidates to include in a merge candidate list. The selection can be based on a priority level of each motion vector candidate. The processor can be further configured to generate the merge candidate list to include the selected motion vector candidates.
Abstract:
This disclosure relates to techniques for performing sample adaptive offset (SAO) processes in a video coding process. A video coder may store sets of SAO information. The SAO information may include data indicative of offset values. The video coder may also store mapping information that maps at least some of the sets of SAO information for one or more sequence partitions of a frame of video data. Additionally, the video coder may perform the SAO processes for one of the partitions of the frame based on the stored SAO information and the stored mapping information.
Abstract:
A filter unit of a video encoder or video decoder can determine a first metric for a group of pixels within a block of pixels based on a comparison of a subset of the pixels in the block to other pixels in the block; determine a filter based on the first metric; and generate a filtered image by applying the filter to the group of pixels. The subset of pixels can be selected to not include pixels on the boundary of the block of pixels.
Abstract:
This disclosure describes techniques for performing sample adaptive offset signaling and coding in a video coding process. Techniques of the disclosure include both a merge-based and prediction-based signaling process for sample adaptive offset information (i.e., offset values and offset type). The techniques includes determining offset information for a current partition, comparing the offset information of the current partition with offset information of one or more neighbor partitions, coding a merge instruction in the case that the offset information of one of the one or more neighbor partitions is the same as the offset information of the current partition, and coding one of a plurality of prediction instructions in the case that the offset information of the one or more neighbor partitions is not the same as the offset information of the current partition.
Abstract:
In general, techniques are described for performing multiple passes of sample adaptive offset (SAO) filtering when coding video data. A video decoding device comprising one or more processors may perform the techniques. The processors may determine a first SAO pixel classification for a block of video data and determine a first offset value based on the first SAO pixel classification and one or more pixel values of the block. The one or more processors may also determine a second SAO pixel classification for the block and determine a second offset value based on the second SAO pixel classification and the one or more pixel values of block of video data. The processors may then apply the first offset value and the second offset value to the block of video data to generate a block of SAO filtered video data.
Abstract:
A video encoder determines reference blocks for each inter-predicted prediction unit (PU) of a tree block group such that each of the reference blocks is in a reference picture that is in a reference picture subset for the tree block group. The reference picture subset for the tree block group includes less than all reference pictures in a reference picture set of the current picture. The tree block group comprises a plurality of concurrently-coded tree blocks in the current picture. For each inter-predicted PU of the tree block group, the video encoder indicates, in a bitstream that includes a coded representation of video data, a reference picture that includes the reference block for the inter-predicted PU. A video decoder receives the bitstream, determines the reference pictures of the inter-predicted PUs of the tree block group, and generates decoded video blocks using the reference blocks of the inter-predicted PUs.
Abstract:
Techniques and systems are provided for compressing data in a neural network. For example, output data can be obtained from a node of the neural network. Re-arranged output data having a re-arranged scanning pattern can be generated. The re-arranged output data can be generated by re-arranging the output data into the re-arranged scanning pattern. One or more residual values can be determined for the re-arranged output data by applying a prediction mode to the re-arranged output data. The one or more residual values can then be compressed using a coding mode.
Abstract:
Techniques and systems are provided for compressing data in a neural network. For example, output data can be obtained from a node of the neural network. Re-arranged output data having a re-arranged scanning pattern can be generated. The re-arranged output data can be generated by re-arranging the output data into the re-arranged scanning pattern. One or more residual values can be determined for the re-arranged output data by applying a prediction mode to the re-arranged output data. The one or more residual values can then be compressed using a coding mode.
Abstract:
Certain aspects of the present disclosure provide a method of encoding data. The method generally includes receiving data comprising a fractional number comprising an exponential component and a fractional component, the exponential component being represented by an exponential bit sequence, the fractional component being represented by a fractional bit sequence. The method further includes determining if the fractional component is within a threshold of 0 or 1. The method further includes setting the fractional component to 0 when the fractional component is within the threshold of 0 or 1. The method further includes downscaling the fractional bit sequence based on a difference between the exponential component and a second threshold. The method further includes encoding the data. The method further includes transmitting the encoded data.
Abstract:
In one example, a device for accessing image data includes a memory configured to store image data, the memory comprising a first region and a second region; and one or more processing units configured to code most significant bits (MSBs) of a plurality of residuals of samples of a block of an image, each of the residuals representing a respective difference value between a respective raw sample value and a respective predicted value for the respective raw sample value, access the coded MSBs in the first region of the memory, and access least significant bits (LSBs) of the plurality of residuals of the samples in the second region of the memory.