Abstract:
Systems and methods are disclosed for compressing a target video. A computer-implemented method may use a computer system that include one or more physical computer processors and non-transient electronic storage. The computer-implemented method may include: obtaining the target video, extracting one or more frames from the target video, and generating an estimated optical flow based on a displacement of pixels between the one or more frames. The one or more frames may include one or more of a key frame and a target frame.
Abstract:
A method of representing displacement information between the frames of a video and/or multiview sequence, comprising the steps of assigning a plurality of the frames to a Group of Pictures (GOPs), providing a base displacement model for each GOP, the base displacement model describing a displacement field that carries each location in a designated base frame of the GOP to a corresponding location in each other the frame of the GOP, and inferring other displacement relationships between the frames of the GOP from the base displacement model. Embodiments include a piecewise smooth displacement field, deformable mesh, reverse displacement field, double mapping, synthesize optical blur and foreground-background discrimination process.
Abstract:
Systems and methods in accordance with embodiments of the invention enable feature based high resolution motion estimation from low resolution images captured using an array camera. One embodiment includes performing feature detection with respect to a sequence of low resolution images to identify initial locations for a plurality of detected features in the sequence of low resolution images, where the at least one sequence of low resolution images is part of a set of sequences of low resolution images captured from different perspectives. The method also includes synthesizing high resolution image portions, where the synthesized high resolution image portions contain the identified plurality of detected features from the sequence of low resolution images. The method further including performing feature detection within the high resolution image portions to identify high precision locations for the detected features, and estimating camera motion using the high precision locations for said plurality of detected features.
Abstract:
It is proposed a Method and System for Global Motion Estimation and Compensation. The system operates on two depth maps which contain a reference depth map and a current depth map as well as their associated texture images if available. The system executes four major steps which are feature detection (step 1), global motion compensation (GMC) (step 2), major moving objects (MMO) detection and the estimation of their motion parameters denoted as major motion parameters (MMPs) (step 3), and local motion compensation (LMC) of macroblocks or other coding/prediction units (step 4). The output of the system is the global motion parameters (GMPs), major moving objects (MMOs) and MMPs, the local motion vectors (LMVs), and the coding modes.
Abstract:
3D video coding/decoding supporting inter-component prediction is made operatable with respect to a broader range of settings with respect to accompanying individual views with depth and deactivation/enablement of inter-component prediction, by using one or more flags in the data stream to control the enablement/disablement of the availability of inter-component prediction for coding units of a current picture.
Abstract:
3D video coding/decoding supporting inter-component prediction is made operatable with respect to a broader range of settings with respect to accompanying individual views with depth and deactivation/enablement of inter-component prediction, by using one or more flags in the data stream to control the enablement/disablement of the availability of inter-component prediction for coding units of a current picture.
Abstract:
A method is provided for coding a current image. The method includes determining, in a first image, different from the current image, a group of k′ pixels corresponding to a current group of k pixels (k′≦k) to be coded of the current image; calculating a motion vector between each of the k′ pixels of the first image and a corresponding pixel of a second image different from the current image, on completion of which a field of k′ motion vectors is obtained; and predicting the pixels or of the motion of the current group of k pixels of the current image on the basis of the field of k′ motion vectors which is obtained.
Abstract:
An input segment of an input video is encoded by first extracting and storing, for each segment of previously encoded videos, a set of reference features. The set of input features are matched with each set of the reference features to produce a set of scores. The reference segments having largest scores are selected to produce a first reduced set of reference segments. A rate-distortion cost for each reference segment in the first reduced set of reference segments is estimated. The reference segments in the first reduced set of reference segments is selected to produce a second reduced set of reference segments. Then, the input segment are encoded based on second reduced set of reference segments.
Abstract:
A system method and computer program product for creating a composited video frame sequence for an application. A current scene graph state for the application is compared to a previous scene graph state wherein each scene graph state includes a plurality of hierarchical nodes that represent one or more objects at each node. A video construction engine determines if one or more objects have moved based upon a comparison of the scene graph states. If one or more objects have moved based upon the scene graph comparison, motion information about the objects is determined and the motion information is forwarded to a stitcher module. The motion information is used to define motion vectors for use by the stitcher module in construction of the composited video frame sequence.
Abstract:
Embodiments of the present application use motion information of an imaging device in encoding video or image frames captured by the imaging device. In one embodiment, a video encoding system that processes image data captured by an image sensor comprises at least one encoder that receives motion information describing motion of the imaging sensor during capturing of the image data and that encodes the image data with assistance of the motion information.