Abstract:
Implementations of sample adaptive offset (SAO) processing a reconstructed picture in an image or video coding system are described. In one example implementation, a method may receive input data associated with the reconstructed picture. The method may also perform multiple stages of SAO filtering operations on a SAO processing unit of at least a portion of the reconstructed picture. Information related to a SAO parameter set that signals one or more SAO types, one or more SAO subtypes, one or more SAO offset values, or a combination thereof, used by the multiple stages of SAO filtering operations are encoded or decoded.
Abstract:
A method and apparatus for coding video data using Inter prediction mode or Merge mode in a video coding system are disclosed, where the video data is configured into a Base Layer (BL) and an Enhancement Layer (EL), and the EL has higher spatial resolution or better video quality than the BL. In one embodiment, at least one information piece of motion information associated with one or more BL blocks in the BL is identified. A motion vector prediction (MVP) candidate list or a Merge candidate list for the selected block in the EL is then determined, where said at least one information piece associated with said one or more BL blocks in the BL is included in the MVP-candidate list or a Merge candidate MVP candidate list or the Merge candidate list. The input data associated with the selected block is coded or decoded using the MVP candidate list or the Merge candidate list.
Abstract:
A method and apparatus for inter-layer prediction for scalable video coding are disclosed. Embodiments of the present invention utilize weighted prediction for scalable coding. The weighted prediction is based on the predicted texture data and the inter-layer Intra prediction data derived from BL reconstructed data. The inter-layer Intra prediction data corresponds to the BL reconstructed data or up-sampled BL reconstructed data. The predicted texture data corresponds to spatial Intra prediction data or motion-compensated prediction data based on the second EL video data in the same layer as the current EL picture. Embodiments of the present invention also utilize the reference picture list including an inter-layer reference picture (ILRP) corresponding to BL reconstructed texture frame or up-sampled BL reconstructed texture frame for Inter prediction of EL video data. The motion vector is limited to a range around (0,0) when the ILRP is selected as a reference picture.
Abstract:
A method for for signaling high-level information in image and video coding is provided. A video decoder receives data from a bitstream to be decoded as a current picture of a video. The video decoder parses a picture header of the current picture that includes a set of one or more slice information syntax elements in the picture header for indicating presence of one or more types of slices that are present in the current picture. The video decoder reconstructs the slices of the current picture by using the set of slice information syntax elements.
Abstract:
Video processing methods and apparatuses include receiving input video data associated with a current picture composed of multiple Coding Tree Units (CTUs) for encoding or decoding, determining a number of subpictures, partitioning the current picture into one or more subpictures, and encoding or decoding each subpicture in the current picture. Each subpicture contains multiple complete CTUs and boundaries of each subpicture are aligned with grids of the current picture in units of CTUs. The number of subpictures in the current picture is limited by an allowed maximum number of slices.
Abstract:
A video coder that constrains the total number of regular bins used for entropy coding syntax elements of a current block is provided. The video coder entropy encodes or decodes the syntax elements selectively as either regular bins using context modeling or as bypass bins without context modeling. A constraint is specified to limit a total number of regular bins used for entropy coding the syntax elements of the current block. There may be no constraint limiting a number of regular bins specific to an individual syntax element of the current block.
Abstract:
A method for performing transform skip mode (TSM) in a video decoder is provided. A video decoder receives data from a bitstream to be decoded as a plurality of video pictures. The video decoder parses the bitstream for a first syntax element in a sequence parameter set (SPS) of a current sequence of video pictures. When the first syntax element indicates that transform skip mode is allowed for the current sequence of video pictures and when transform skip mode is used for a current block in a current picture of the current sequence, the video decoder reconstructs the current block by using quantized residual signals that are not transformed.
Abstract:
Processing methods and apparatuses for video data comprise receiving input data associated with a current transform block in a current video picture, determining if a size of the current transform block is a power of 4, determining a normalized quantization or dequantization scaling factor, generating transform coefficient levels by applying a quantization processing to scaled transform coefficients in the current transform block or generating scaled transform coefficients by applying an inverse quantization processing to transform coefficient levels in the current transform block, and encoding or decoding the current transform block. The quantization processing comprises scaling the scaled transform coefficients by the normalized quantization scaling factor and the inverse quantization processing comprises scaling the transform coefficient levels by the normalized dequantization scaling factor.
Abstract:
Method and apparatus for encoding and decoding prediction residues in a video coding system also disclosed. At the decoder side, a Rice parameter for the target transform coefficient is determined based on a local sum of absolute levels of neighboring transform coefficients of the target transform coefficient. A dependent quantization state is determined and a zero-position variable is determined based on the dependent quantization state and the Rice parameter. One or more coded bits associated with a first syntax element for the target transform coefficient in a transform block are parsed and decoded using one or more codes including a Golomb-Rice code with the Rice parameter, where the first syntax element corresponds to a modified absolute level value of the target transform coefficient. An absolute level value of the target transform coefficient is derived according to the zero-position variable and the first syntax element.
Abstract:
Video processing methods and apparatuses for processing video pictures referring to a high-level syntax set include receiving input data, determining a first syntax element indicating whether reference picture resampling is disabled or constrained, determining a second syntax element indicating whether subpicture partitioning is disabled or constrained, and encoding or decoding the video pictures. The first and second syntax elements are restricted to disable or constrain subpicture partitioning when reference picture resampling is enabled or disable or constrain reference picture resampling when subpicture partitioning is enabled. The first syntax element and the second syntax element are syntax elements signaled in the high level syntax set.