Video object segmentation by reference-guided mask propagation

    公开(公告)号:US10671855B2

    公开(公告)日:2020-06-02

    申请号:US15949935

    申请日:2018-04-10

    Applicant: Adobe Inc.

    Abstract: Various embodiments describe video object segmentation using a neural network and the training of the neural network. The neural network both detects a target object in the current frame based on a reference frame and a reference mask that define the target object and propagates the segmentation mask of the target object for a previous frame to the current frame to generate a segmentation mask for the current frame. In some embodiments, the neural network is pre-trained using synthetically generated static training images and is then fine-tuned using training videos.

    Segmenting Objects In Video Sequences
    3.
    发明申请

    公开(公告)号:US20200143171A1

    公开(公告)日:2020-05-07

    申请号:US16183560

    申请日:2018-11-07

    Applicant: Adobe Inc.

    Abstract: In implementations of segmenting objects in video sequences, user annotations designate an object in any image frame of a video sequence, without requiring user annotations for all image frames. An interaction network generates a mask for an object in an image frame annotated by a user, and is coupled both internally and externally to a propagation network that propagates the mask to other image frames of the video sequence. Feature maps are aggregated for each round of user annotations and couple the interaction network and the propagation network internally. The interaction network and the propagation network are trained jointly using synthetic annotations in a multi-round training scenario, in which weights of the interaction network and the propagation network are adjusted after multiple synthetic annotations are processed, resulting in a trained object segmentation system that can reliably generate realistic object masks.

    Joint Trimap Estimation and Alpha Matte Prediction for Video Matting

    公开(公告)号:US20230360177A1

    公开(公告)日:2023-11-09

    申请号:US17736397

    申请日:2022-05-04

    Applicant: Adobe Inc.

    Abstract: In implementations of systems for joint trimap estimation and alpha matte prediction, a computing device implements a matting system to estimate a trimap for a frame of a digital video using a first stage of a machine learning model. An alpha matte is predicted for the frame based on the trimap and the frame using a second stage of the machine learning model. The matting system generates a refined trimap and a refined alpha matte for the frame based on the alpha matte, the trimap, and the frame using a third stage of the machine learning model. An additional trimap is estimated for an additional frame of the digital video based on the refined trimap and the refined alpha matte using the first stage of the machine learning model.

    VIDEO OBJECT SEGMENTATION BY REFERENCE-GUIDED MASK PROPAGATION

    公开(公告)号:US20200250436A1

    公开(公告)日:2020-08-06

    申请号:US16856292

    申请日:2020-04-23

    Applicant: Adobe Inc.

    Abstract: Various embodiments describe video object segmentation using a neural network and the training of the neural network. The neural network both detects a target object in the current frame based on a reference frame and a reference mask that define the target object and propagates the segmentation mask of the target object for a previous frame to the current frame to generate a segmentation mask for the current frame. In some embodiments, the neural network is pre-trained using synthetically generated static training images and is then fine-tuned using training videos.

    VIDEO OBJECT SEGMENTATION BY REFERENCE-GUIDED MASK PROPAGATION

    公开(公告)号:US20190311202A1

    公开(公告)日:2019-10-10

    申请号:US15949935

    申请日:2018-04-10

    Applicant: Adobe Inc.

    Abstract: Various embodiments describe video object segmentation using a neural network and the training of the neural network. The neural network both detects a target object in the current frame based on a reference frame and a reference mask that define the target object and propagates the segmentation mask of the target object for a previous frame to the current frame to generate a segmentation mask for the current frame. In some embodiments, the neural network is pre-trained using synthetically generated static training images and is then fine-tuned using training videos.

    Space-time memory network for locating target object in video content

    公开(公告)号:US11200424B2

    公开(公告)日:2021-12-14

    申请号:US16293126

    申请日:2019-03-05

    Applicant: Adobe Inc.

    Abstract: Certain aspects involve using a space-time memory network to locate one or more target objects in video content for segmentation or other object classification. In one example, a video editor generates a query key map and a query value map by applying a space-time memory network to features of a query frame from video content. The video editor retrieves a memory key map and a memory value map that are computed, with the space-time memory network, from a set of memory frames from the video content. The video editor computes memory weights by applying a similarity function to the memory key map and the query key map. The video editor classifies content in the query frame as depicting the target feature using a weighted summation that includes the memory weights applied to memory locations in the memory value map.

    Video object segmentation by reference-guided mask propagation

    公开(公告)号:US11176381B2

    公开(公告)日:2021-11-16

    申请号:US16856292

    申请日:2020-04-23

    Applicant: Adobe Inc.

    Abstract: Various embodiments describe video object segmentation using a neural network and the training of the neural network. The neural network both detects a target object in the current frame based on a reference frame and a reference mask that define the target object and propagates the segmentation mask of the target object for a previous frame to the current frame to generate a segmentation mask for the current frame. In some embodiments, the neural network is pre-trained using synthetically generated static training images and is then fine-tuned using training videos.

Patent Agency Ranking