Abstract:
An apparatus and method of deriving a motion vector predictor (MVP) for a current MV of a current block in Inter, Merge or Skip mode are disclosed based on motion vector (MV) attribute search. The system determines first MV attribute search comprising whether a given MV pointing to the target reference picture in the given reference list, or whether the given MV pointing to the target reference picture in other reference list, and determines second MV attribute search comprising whether the given MV pointing to other reference pictures in the given reference list, or whether the given MV pointing to the other reference pictures in the other reference list. The MVP for the current block is then determined from the neighboring blocks according to a search order.
Abstract:
Method and apparatus of coding pictures containing one or more virtual boundaries, such as 360-degree virtual reality (VR360) video are disclosed. According to this method, a reconstructed filtered unit associated with a loop filter for a current reconstructed pixel is received. The loop filtering process associated with the loop filter is applied to the current reconstructed pixel to generate a filtered reconstructed pixel, where if the loop filtering process for the current reconstructed pixel is across a virtual boundary of the picture, the loop filtering process is disabled when fixed-size loop filtering is used or a smaller-size loop filter is selected when adaptive-size loop filtering is used for the current reconstructed pixel, where the filtered reconstructed pixel is the same as the current reconstructed pixel when the loop filtering process is disabled. The filtered reconstructed pixel is the same as the current reconstructed pixel.
Abstract:
A video decoding method includes: decoding a part of a bitstream to generate a decoded frame, and parsing at least one syntax element from the bitstream. The decoded frame is a projection-based frame that includes projection faces packed in a cube-based projection layout. At least a portion of a 360-degree content of a sphere is mapped to the projection faces via cube-based projection. The at least one syntax element is indicative of a guard band configuration of the projection-based frame.
Abstract:
An exemplary video processing method includes: receiving an omnidirectional content corresponding to a sphere; obtaining a plurality of projection faces from the omnidirectional content of the sphere according to a pyramid projection; creating at least one padding region; and generating a projection-based frame by packing the projection faces and the at least one padding region in a pyramid projection layout. The projection faces packed in the pyramid projection layout include a first projection face. The at least one padding region packed in the pyramid projection layout includes a first padding region. The first padding region connects with at least the first projection face, and forms at least a portion of one boundary of the pyramid projection layout.
Abstract:
A video processing method includes: decoding a part of a bitstream to generate a decoded frame, where the decoded frame is a projection-based frame that includes projection faces in a hemisphere cubemap projection layout; and remapping sample locations of the projection-based frame to locations on the sphere, where a sample location within the projection-based frame is converted into a local sample location within a projection face packed in the projection-based frame; in response to adjustment criteria being met, an adjusted local sample location within the projection face is generated by applying adjustment to one coordinate value of the local sample location within the projection face, and the adjusted local sample location within the projection face is remapped to a location on the sphere; and in response to the adjustment criteria not being met, the local sample location within the projection face is remapped to a location on the sphere.
Abstract:
A video decoding method includes decoding a part of a bitstream to generate a decoded frame. The decoded frame is a projection-based frame that comprises at least one projection face and at least one guard band packed in a projection layout. At least a portion of a 360-degree content of a sphere is mapped to the at least one projection face via projection. The decoded frame is in a 4:2:0 chroma format or a 4:2:2 chroma format, and a guard band size of each of the at least one guard band is equal to an even number of luma samples.
Abstract:
Method and apparatus of coding a video sequence are disclosed. According to this method, a first syntax is signalled in or parsed from a bitstream, where the first syntax indicates whether a loop filtering process is disabled for one or more virtual boundaries in a corresponding region. A reconstructed filter unit in a current picture is received, wherein the reconstructed filter unit is associated with the loop filter and the reconstructed filter unit comprises reconstructed pixels for applying a loop filtering process associated with the loop filter to a current reconstructed pixel. When the first syntax is true, the loop filter processing is disabled when the reconstructed filter unit is across said one or more virtual boundaries in the corresponding region. When the first syntax is false, the loop filter processing is not disabled when the reconstructed filter unit is across the virtual boundary.
Abstract:
A video processing method includes: obtaining a plurality of square projection faces from an omnidirectional content of a sphere according to a cube-based projection, scaling the square projection faces to generate a plurality of scaled projection faces, respectively, creating at least one padding region, generating a projection-based frame by packing the scaled projection faces and said at least one padding region in a projection layout of the cube-based projection, and encoding the projection-based frame to generate a part of a bitstream.
Abstract:
A method and apparatus of priority-based MVP (motion vector predictor) derivation for motion compensation in a video encoder or decoder are disclosed. According to this method, one or more final motion vector predictors (MVPs) are derived using priority-based MVP derivation process. The one or more final MVPs are derived by selecting one or more firstly available MVs from a priority-based MVP list for Inter prediction mode, Skip mode or Merge mode based on reference data of one or two target reference pictures that are reconstructed prior to the current block according to a priority order. Therefore, there is no need for transmitting information at the encoder side nor deriving information at the decoder side that is related to one or more MVP indices to identify the one or more final MVPs in the video bitstream.
Abstract:
A method and apparatus for Intra prediction of non-square chroma blocks are disclosed. The system uses a block partitioning structure for partitioning a chroma block into one or more final sub-blocks including at least one non-square final sub-block using block partition recursively. The block partition either splits each given block into intermediate sub-blocks using one or more split modes including one binary split mode or applies no splitting to each given block. The block is treated as an initial given block and each intermediate sub-block is treated as one given block in a next recursion. Any given block that is not further split becomes a final sub-block. After the final sub-blocks are determined, non-square Intra prediction is applied to each non-square final sub-block.