Abstract:
Aspects of the disclosure provide a method for denoising a reconstructed picture in a video coding system. The method can include providing two or more candidate non-local denoising technologies, selecting a target non-local denoising technology from the two or more candidate non-local denoising technologies for a patch group, and denoising the patch group of the reconstructed picture with the target non-local denoising technology. Besides, two parallel processing methods are provided for forming patch groups with predictor-based searching algorithms.
Abstract:
A method and apparatus of video coding using IntraBC (Intra Block Copy) for video data are disclosed. Restricted prediction is determined, in which one or more selected PU (prediction unit) sizes, PU prediction types or both are disabled or disallowed. Video coding is applied to the current block using coding modes including IntraBC prediction in accordance with the restricted prediction. The selected PU sizes, PU prediction types or both may correspond to 8×8 bi-prediction PU. In one example, the selected PU sizes, PU prediction types or both are disabled or disallowed for the restricted prediction except one or more conditions are determined to be satisfied. The conditions may correspond to at least one of two motion vectors of the 8×8 bi-prediction PU having integer values, the conditions may also correspond to at least one of motion vector components of the 8×8 bi-prediction PU having integer valueIntraBCIntraBCIntraBCIntraBCIntraBC.
Abstract:
A method for color index coding of blocks in a picture is disclosed. In one embodiment for the decoder side, a current coding mode is parsed from the input coded bitstream for one or more first currently decoded pixels in a current block starting from a first starting position. If the current coding mode corresponds to a generalized copy-previous-row mode, one or more first reference pixels starting from a first starting reference position located at T-th row above the first currently decoded pixels are identified, where T is an integer equal to or greater than one. The first reference pixels starting from the first starting reference position are then used to decode the first currently decoded pixels starting from a first starting position by copying the first reference pixels starting from the first starting reference position respectively. The system may also include a copy-left mode or escape mode.
Abstract:
In one implementation, a method codes video pictures, in which each of the video pictures is partitioned into LCUs (largest coding units). The method operates by receiving a current LCU, partitioning the current LCU adaptively to result in multiple leaf CUs, determining whether a current leaf CU has at least one nonzero quantized transform coefficient according to both Prediction Mode (PredMode) and Coded Block Flag (CBF), and incorporating quantization parameter information for the current leaf CU in a video bitstream, if the current leaf CU has at least one nonzero quantized transform coefficient. If the current leaf CU has no nonzero quantized transform coefficient, the method excludes the quantization parameter information for the current leaf CU in the video bitstream.
Abstract:
A method for video decoding includes receiving a video frame reconstructed based on data received from a bitstream. The method further includes extracting, from the bitstream, a first syntax element indicating whether a spatial partition for partitioning the video frame is active. The method also includes, responsive to the first syntax element indicating that the spatial partition for partitioning the video frame is active, determining a configuration of the spatial partition for partitioning the video frame, determining a plurality of parameter sets of a neural network, and applying the neural network to the video frame. The video frame is spatially divided based on the determined configuration of the spatial partition for partitioning the video frame into a plurality of portions, and the neural network is applied to the plurality of portions in accordance with the determined plurality of parameter sets.
Abstract:
Methods and apparatus of video coding are disclosed. According to this method, input data comprising pixel data for a current block to be encoded at an encoder side or encoded data of the current block to be decoded at a decoder side is received. When one or more reference blocks or sub-blocks of the current block are coded in an affine mode, the following coding process is applied: one or more derived MVs (Motion Vectors) are determined for the current block according to one or more affine models associated with said one or more reference blocks or sub-blocks; a merge list comprising at least one of said one or more derived MVs as one translational MV candidate is generated; and predictive encoding or decoding is applied to the input data using information comprising the merge list.
Abstract:
A method and apparatus for video coding are disclosed. According to the method for the decoder side, a first syntax, related to whether the current block is coded in a CCLM related mode, is parsed from a bitstream comprising the encoded data for the current block. If the first syntax indicates the current block being coded in the CCLM related mode, a second syntax is parsed from the bitstream, wherein the second syntax is related to whether a multiple model CCLM mode is used or whether one or more model parameters are explicitly signalled or implicitly derived. The model parameters for the second-colour block are determined if the first syntax indicates the current block being coded in a CCLM related mode. The encoded data associated with the second-colour block is then decoded using prediction data comprising the cross-colour predictor for the second-colour block.
Abstract:
A video coding system that uses multiple models to predict chroma samples is provided. The video coding system receives data for a block of pixels to be encoded or decoded as a current block of a current picture of a video. The video coding system derives multiple prediction linear models based on luma and chroma samples neighboring the current block. The video coding system constructs a composite linear model based on the multiple prediction linear models. The video coding system applies the composite linear model to incoming or reconstructed luma samples of the current block to generate a chroma predictor of the current block. The video coding system uses the chroma predictor to reconstruct chroma samples of the current block or to encode the current block.
Abstract:
A method and apparatus for video coding system that utilizes low-latency template-matching motion-vector refinement are disclosed. According to this method, input data associated with a current block of a video unit in a current picture are received. Motion compensation is then applied to the current block according to an initial motion vector (MV) to obtain initial motion-compensated predictors of the current. After applying the motion compensation to the current block, template-matching MV refinement is applied to the current block to obtain a refined MV for the current block. The current block is then encoded or decoded using information including the refined MV. The method may further comprise determining gradient values of the initial motion-compensated predictors. The initial motion-compensated predictors can be adjusted by taking into consideration of the gradient values and/or MV difference between the refined and initial MVs.
Abstract:
Video encoding or decoding methods and apparatuses include receiving input data associated with a current block in a current picture, determining a preload region in a reference picture shared by two or more coding configurations of affine prediction or motion compensation or by two or more affine refinement iterations, loading reference samples in the preload region, generating predictors for the current block, and encoding or decoding the current block according to the predictors. The predictors associated with the affine refinement iterations or coding configurations are generated based on some of the reference samples in the preload region.