Abstract:
Methods and apparatus of video coding for 3D video data are disclosed. According to one method, the gradients of the geometry frame are derived. A reconstructed point cloud is reconstructed using the geometry frame. One or more candidate holes in the reconstructed point cloud are filled based on the gradients of the geometry frame. According to another method of encoding for 3D video data, candidate hole locations in a geometry frame, patch or layer are determined. Source points projected to the candidate hole locations are grouped to generate grouped points. The grouped points are removed from an original patch containing the grouped points.
Abstract:
A method and apparatus for Intra prediction of non-square chroma blocks are disclosed. The system uses a block partitioning structure for partitioning a chroma block into one or more final sub-blocks including at least one non-square final sub-block using block partition recursively. The block partition either splits each given block into intermediate sub-blocks using one or more split modes including one binary split mode or applies no splitting to each given block. The block is treated as an initial given block and each intermediate sub-block is treated as one given block in a next recursion. Any given block that is not further split becomes a final sub-block. After the final sub-blocks are determined, non-square Intra prediction is applied to each non-square final sub-block.
Abstract:
Methods and apparatus of processing cube face images are disclosed. According to embodiments of the present invention, one or more discontinuous boundaries within each assembled cubic frame are determined and used for selective filtering, where the filtering process is skipped at said one or more discontinuous boundaries within each assembled cubic frame when the filtering process is enabled. Furthermore, the filtering process is applied to one or more continuous areas in each assembled cubic frame.
Abstract:
A method and apparatus for applying filter to Intra prediction samples are disclosed. According to an embodiment of the present invention, a filter is applied to one or more prediction samples of the Initial Intra prediction block to form one or more filtered prediction samples. For example, the filter is applied to the prediction sample in the non-boundary locations of the Initial Intra prediction block. Alternatively, the filter is applied to all prediction samples in the Initial Intra prediction block. The filtered Intra prediction block comprising one or more filtered prediction samples is used as a predictor for Intra prediction encoding or decoding of the current block. The filter corresponds to a FIR (finite impulse response) filter or an IIR (infinite impulse response) filter.
Abstract:
Aspects of the disclosure provide a method for video coding. The method includes receiving input data associated with a processing block in a current picture, selecting, from a set of neighboring reconstructed samples for intra-coding pixels in the processing block, a plurality of reference samples for a pixel in the processing block based on a position of the pixel and an intra prediction mode of the processing block, determining a projection phase for the pixel based on the position of the pixel and the intra prediction mode of the processing block, determining coefficients of an interpolation filter based on the projection phase for the pixel, applying the interpolation filter with the determined coefficients on the reference samples to generate a prediction of the pixel, and encoding or decoding the pixel in the processing block using the prediction of the pixel.
Abstract:
A method and apparatus for deriving aligned inter-view prediction for uni-prediction and bi-prediction in three-dimensional and multi-view coding with illumination compensation enabled are disclosed. When the illumination compensation is enabled, the derivation process for a single reference block in reference list_0 or reference list_1 in the case of uni-prediction is aligned with the derivation process for a reference block in reference list_0 and a reference block in reference list_1 in the case of bi-prediction. The reconstruction process generates the reference block based on a temporal reference block when Inter prediction is used and based on an inter-view reference block when inter-view prediction is used. For the uni-prediction and bi-prediction, the same clipping process can be included in the reconstruction process, the illumination compensation process, or both.
Abstract:
The invention provides a motion prediction method. First, a plurality of candidate units corresponding to a current unit of a current frame is determined. A plurality of motion vectors of the candidate units is then obtained. A plurality of scaling factors of the candidate units is then calculated according to a plurality of respective temporal distances depending on a plurality of reference frames of the motion vectors. The motion vectors of the candidate units are then scaled according to the scaling factors to obtain a plurality of scaled motion vectors. The scaled motion vectors are ranked, and a subset of highest ranking motion vectors are identified to be included in a candidate set. Finally, a motion vector predictor for motion prediction of the current unit is then selected from the candidate units.
Abstract:
Methods and apparatus of processing spherical images related to segmented sphere projection (SSP) are disclosed. According to one method, a North Pole region of the spherical image is projected to a first circular image and a South Pole region of the spherical image is projected to a second circular image using a mapping process selected from a mapping group comprising equal-area mapping, non-uniform mapping and cubemap mapping. Methods and apparatus of processing spherical images related to rotated sphere projection (RSP) are also disclosed. According to this method, the spherical image is projected into a first part of rotated sphere projection corresponding to a θ×φ region of the spherical image and a second part of rotated sphere projection corresponding to a remaining part of the spherical image using equal-area mapping.
Abstract:
Methods and apparatus of processing 360-degree virtual reality images are disclosed. According to one method, each 360-degree virtual reality image is projected into one first projection picture using first projection-format conversion. The first projection pictures are encoded and decoded into first reconstructed projection pictures. Each first reconstructed projection picture is then projected into one second reconstructed projection picture or one third reconstructed projection picture corresponding to a selected viewpoint using second projection-format conversion. One or more discontinuous edges in one or more second reconstructed projection pictures or one or more third reconstructed projection pictures corresponding to the selected viewpoint are identified. A post-processing filter is then applied to at least one discontinuous edge in the second reconstructed projection pictures or third reconstructed projection pictures corresponding to the selected viewpoint to generate filtered output.
Abstract:
A method and apparatus or video coding or processing for an image sequence corresponding to virtual reality (VR) video are disclosed. According to embodiments of the present invention, a padded area outside one cubic face frame boundary of one cubic face frame is padded to form a padded cubic face frame using one or more extended cubic faces, where at least one boundary cubic face in said one cubic face frame has one padded area using pixel data derived from one extended cubic face in a same cubic face frame.