Abstract:
A method and apparatus for deriving a motion vector predictor (MVP) candidate set for a block are disclosed. Embodiments according to the present invention generate a complete full MVP candidate set based on the redundancy-removed MVP candidate set if one or more redundant MVP candidates exist. In one embodiment, the method generates the complete full MVP candidate set by adding replacement MVP candidates to the redundancy-removed MVP candidate set and a value corresponding to a non-redundant MVP is assigned to each replacement MVP candidate. In another embodiment, the method generates the complete full MVP candidate set by adding replacement MVP candidates to the redundancy-removed MVP candidate set and a value is assigned to each replacement MVP candidate according to a rule. The procedure of assigning value, checking redundancy, removing redundant MVP candidate are repeated until the MVP candidate set is complete and full.
Abstract:
In an example, a method of decoding video data includes determining a prediction partitioning structure for predicting pixel values associated with a block of video data. The method also includes determining a transform partitioning structure for applying one or more transforms to the predicted pixel values. Determining the transform split structure includes splitting a parent transform unit, upon determining the transform partitioning structure comprises splitting the parent transform unit into one or more square transforms, determining one or more square transforms such that each of the one or more square transforms correspond to exactly one prediction partition, and upon determining the transform partitioning structure comprises splitting the parent transform unit into one or more non-square transforms, determining whether to split the one or more non-square transforms based at least in part on the one or more non-square transforms being non-square.
Abstract:
While maintaining a high degree of freedom in choosing partition sizes and transformation sizes adapted for local characteristics of videos, the amount of metadata is decreased. A video encoding apparatus (10) divides an input video into blocks of a prescribed size and encodes the video block by block. The video encoding apparatus is provided with: a prediction parameter determining portion (102) that decides the block partition structure; a predictive image producing portion (103) that generates predictive images, partition by partition, as prescribed by the partition structure; a transform coefficient producing portion (107) which applies one of the frequency transformations included in a prescribed transformation preset to prediction residuals, i.e. the differences between predictive images and the input video; a transform restriction deriving portion (104) which generates the list of transform candidate, i.e. lists of frequency transformations that can be applied to each partition, on the basis of partition format information; and a variable-length-encoding portion (108) which, on the basis of the list of transform candidate and the transformation preset, performs variable-length encoding on transformation selection flags.
Abstract:
A video coding method includes at least the following steps: utilizing a visual quality evaluation module for evaluating visual quality based on data involved in a coding loop; and referring to at least the evaluated visual quality for deciding a target bit allocation of a rate-controlled unit in video coding. Besides, a video coding apparatus has a visual quality evaluation module, a rate controller and a coding circuit. The visual quality evaluation module evaluates visual quality based on data involved in a coding loop. The rate controller refers to at least the evaluated visual quality for deciding a target bit allocation of a rate-controlled unit. The coding circuit has the coding loop included therein, and encodes the rate-controlled unit according to the target bit allocation.
Abstract:
Methods for motion estimation with adaptive motion accuracy of the present invention include several techniques for computing motion vectors of high pixel accuracy with a minor increase in computation. One technique uses fast-search strategies in sub-pixel space that smartly searches for the best motion vectors. An alternate technique estimates high-accurate motion vectors using different interpolation filters at different stages in order to reduce computational complexity. Yet another technique uses rate-distortion criteria that adapts according to the different motion accuracies to determine both the best motion vectors and the best motion accuracies. Still another technique uses a VLC table that is interpreted differently at different coding units, according to the associated motion vector accuracy.
Abstract:
During a process to derive an inter-view predicted motion vector candidate (IPMVC) for an Advanced Motion Vector Prediction (AMVP) candidate list, a video coder determines, based on a disparity vector of a current prediction unit (PU), a reference PU for the current PU. Furthermore, when a first reference picture of the reference PU has the same picture order count (POC) value as a target reference picture of the current PU, the video coder determines an IPMVC based on a first motion vector of the reference PU. Otherwise, when a second reference picture of the reference PU has the same POC value as the target reference picture of the current PU, the video coder determines the IPMVC based on a second motion vector of the reference PU.
Abstract:
Embodiments of systems and methods for signaling chroma information for a picture in a compressed video stream are provided. One system embodiment, among others, comprises a memory with logic, and a processor configured with the logic to provide a compressed video stream that includes a picture having chroma samples and luma samples, and provide in the compressed video stream a flag for signaling information corresponding to the location of the chroma samples in relation to the luma samples in the picture, wherein a first defined flag value indicates default locations of the chroma samples in relation to the luma samples in the picture, wherein a second defined flag value indicates a presence in the compressed video stream of auxiliary chroma information corresponding to relative locations of the chroma samples to the luma samples in the picture, and wherein the number of chroma samples in the picture implied by the first defined flag value is equal to the number of chroma samples in the picture implied by the second defined flag value. Other embodiments for signaling chroma information for a picture in a compressed video stream are included herein.
Abstract:
Techniques are disclosed for detecting sudden illumination changes using radiance consistency within a spatial neighborhood. A background/foreground (BG/FG) component of a behavior recognition system may be configured to generate a background image depicting a scene background. Further, the (BG/FG) component may periodically evaluate a current video frame to determine whether a sudden illumination change has occurred. A sudden illumination change occurs when scene lighting changes dramatically from one frame to the next (or over a small number of frames).
Abstract:
A motion vector derivation unit includes a comparison unit for comparing a parameter TR1 for a reference vector with a predetermined value to determine whether it exceeds the predetermined value or not; a switching unit for switching selection between the maximum value of a pre-stored parameter TR and the parameter TR1 according to the comparison result by the comparison unit; a multiplier parameter table (for multipliers); and a multiplier parameter table (for divisors) for associating the parameter TR1 with a value approximate to the inverse value (1/TR1) of this parameter TR1.