Abstract:
Aspects of the disclosure include a method for video coding. The method includes receiving input data associated with a current block in a current image frame of video data, where the current block is coded by intra-prediction or to be coded by intra-prediction. The method also includes determining an intra-prediction mode of the current block, selecting one of a plurality of filters including at least a default filter and an N-tap filter, and generating filtered neighboring samples by filtering neighboring samples adjacent to the current block using the selected filter, where N is a positive integer different from 3. Moreover, the method includes encoding or decoding the current block by predicting the current block based on the filtered neighboring samples and the intra-prediction mode.
Abstract:
A method and apparatus for a three-dimensional or multi-view video encoding or decoding system utilizing unified disparity vector derivation is disclosed. When a three-dimensional coding tool using a derived disparity vector (DV) is selected, embodiments according to the present invention will first obtain the derived DV from one or more neighboring blocks. If the derived DV is available, the selected three-dimensional coding tool is applied to the current block using the derived DV. If the derived DV is not available, the selected three-dimensional coding tool is applied to the current block using a default DV, where the default DV is set to point to an inter-view reference picture in a reference picture list of the current block.
Abstract:
A method and apparatus for determining a derived disparity vector (DV) directly from associated depth block for motion vector prediction in three-dimensional video encoding or decoding are disclosed. Input data associated with current motion information of a current texture block of a current texture picture in a current dependent view and a depth block associated with the current texture block are received. The derived DV for the current texture block based on the depth block is then determined and used for inter-view or temporal motion vector prediction (MVP). If the current motion information corresponds to inter-view prediction, the current DV is encoded or decoded using the derived DV as a MVP. If the current motion information corresponds to temporal prediction, the current MV is encoded or decoded using a derived MV of a corresponding texture block in a reference view as the MVP.
Abstract:
A method and apparatus for signaling one or more coding modes selected from a mode group for multi-view or three-dimensional (3D) video coding are disclosed. The mode group comprises a pulse code modulation (PCM) mode, a segment-wise depth coding (SDC) mode and a depth-based block partitioning (DBBP) mode. According to the present invention, a first mode is determined regarding whether it is enabled, where the first mode corresponds to one of at least two mode members selected from the PCM mode, the SDC mode and the DBBP mode. If the first mode is enabled, a first mode flag corresponding to the first mode is signaled to indicate whether the first mode is asserted. If the first mode is asserted, the current depth or texture block is coded using the first mode and signaling any remaining mode member of at least two mode members is skipped.
Abstract:
A method and apparatus of video coding incorporating Deep Neural Network are disclosed. A target signal is processed using DNN (Deep Neural Network), where the target signal provided to DNN input corresponds to the reconstructed residual, output from the prediction process, the reconstruction process, one or more filtering processes, or a combination of them. The output data from DNN output is provided for the encoding process or the decoding process. The DNN can be used to restore pixel values of the target signal or to predict a sign of one or more residual pixels between the target signal and an original signal. An absolute value of one or more residual pixels can be signalled in the video bitstream and used with the sign to reduce residual error of the target signal.
Abstract:
Method and apparatus of coding a video sequence are disclosed. According to this method, a first syntax is signalled in or parsed from a bitstream, where the first syntax indicates whether a loop filtering process is disabled for one or more virtual boundaries in a corresponding region. A reconstructed filter unit in a current picture is received, wherein the reconstructed filter unit is associated with the loop filter and the reconstructed filter unit comprises reconstructed pixels for applying a loop filtering process associated with the loop filter to a current reconstructed pixel. When the first syntax is true, the loop filter processing is disabled when the reconstructed filter unit is across said one or more virtual boundaries in the corresponding region. When the first syntax is false, the loop filter processing is not disabled when the reconstructed filter unit is across the virtual boundary.
Abstract:
Methods and apparatus of coding a video sequence, wherein pictures from the video sequence contain one or more discontinuous edges are disclosed. The loop filtering process associated with the loop filter is then applied to the current reconstructed pixel to generate a filtered reconstructed pixel, where if the loop filtering process is across a virtual boundary of the current picture, one or more alternative reference pixels are used to replace unexpected reference pixels located in a different side of the virtual boundary from the current reconstructed pixel, and said one or more alternative reference pixels are generated from second reconstructed pixels in a same side of the virtual boundary as the current reconstructed pixel. According to another method, reference pixels are derived from spherical neighbouring reference pixels for the loop filtering process.
Abstract:
Method and apparatus of coding 360-degree virtual reality (VR360) pictures are disclosed. According to the method, when a first MV (motion vector) of a target neighboring block for the current block is not available within the 2D projection picture, or when the target neighboring block is not in a same face as the current block: a true neighboring block picture corresponding to the target neighboring block is identified within the 2D projection; if a second MV of the true neighboring block exists, the second MV of the true neighboring block is transformed into a derived MV; and a current MV of the current block is encoded or decoded using the derived MV or one selected candidate in a MV candidate list including the derived MV as an MV predictor.
Abstract:
A method and apparatus of Intra prediction filtering in an image or video encoder or decoder are disclosed. The method comprises receiving input data associated with a current block (1110); determining a current Intra prediction mode belonging to a set of available Intra prediction modest for the current block (1120); according to the current Intra prediction mode, determining an initial Intra prediction block consisting of initial Intra prediction pixel values based on neighboring reconstructed samples of the current block (1130); applying Intra prediction filter to the initial Intra prediction block according to a current scanning order selected from multiple scanning orders depending on the current Intra prediction mode to generate a filtered Intra prediction block consisting of filtered Intra prediction pixel values, wherein inputs to the Intra prediction filter comprise a current pixel and one or more adjacent pixels, and said multiple scanning orders comprise at least two scanning orders selected from a vertical scanning order, a horizontal scanning order and a diagonal scanning order (1140); applying mode-dependent Intra prediction encoding or depending to the current block using the filtered Intra prediction block as a predictor for the current block (1150).
Abstract:
Methods and apparatus for processing a 360°-VR frame sequence are disclosed. According to one method, input data associated with a 360°-VR frame sequence are received, where each 360°-VR frame comprises one set of faces associated with a polyhedron format. Each set of faces is rearranged into one rectangular whole VR frame consisting of a front sub-frame and a rear sub-frame, where the front sub-frame corresponds to first contents in a first field of view covering front 180°×180° view and the rear sub-frame corresponds to second contents in a second field of view covering rear 180°×180° view. Output data corresponding to a rearranged 360°-VR frame sequence consisting of a sequence of rectangular whole VR frames are provided.