SUPERVISED LEARNING TECHNIQUES FOR ENCODER TRAINING

    公开(公告)号:US20220121932A1

    公开(公告)日:2022-04-21

    申请号:US17384378

    申请日:2021-07-23

    申请人: Adobe Inc.

    IPC分类号: G06N3/08 G06N3/04

    摘要: Systems and methods train an encoder neural network for fast and accurate projection into the latent space of a Generative Adversarial Network (GAN). The encoder is trained by providing an input training image to the encoder and producing, by the encoder, a latent space representation of the input training image. The latent space representation is provided as input to the GAN to generate a generated training image. A latent code is sampled from a latent space associated with the GAN and the sampled latent code is provided as input to the GAN. The GAN generates a synthetic training image based on the sampled latent code. The sampled latent code is provided as input to the encoder to produce a synthetic training code. The encoder is updated by minimizing a loss between the generated training image and the input training image, and the synthetic training code and the sampled latent code.

    DETERMINING CAMERA PARAMETERS FROM A SINGLE DIGITAL IMAGE

    公开(公告)号:US20210358170A1

    公开(公告)日:2021-11-18

    申请号:US17387207

    申请日:2021-07-28

    申请人: Adobe Inc.

    IPC分类号: G06T7/80 G06T7/12 G06T7/13

    摘要: The present disclosure relates to systems, non-transitory computer-readable media, and methods for utilizing a critical edge detection neural network and a geometric model to determine camera parameters from a single digital image. In particular, in one or more embodiments, the disclosed systems can train and utilize a critical edge detection neural network to generate a vanishing edge map indicating vanishing lines from the digital image. The system can then utilize the vanishing edge map to more accurately and efficiently determine camera parameters by applying a geometric model to the vanishing edge map. Further, the system can generate ground truth vanishing line data from a set of training digital images for training the critical edge detection neural network.

    VIDEO INPAINTING VIA MACHINE-LEARNING MODELS WITH MOTION CONSTRAINTS

    公开(公告)号:US20210287007A1

    公开(公告)日:2021-09-16

    申请号:US16817100

    申请日:2020-03-12

    申请人: Adobe Inc.

    IPC分类号: G06K9/00 G06N20/00 G06K9/46

    摘要: Certain aspects involve video inpainting in which content is propagated from a user-provided reference frame to other video frames depicting a scene. For example, a computing system accesses a set of video frames with annotations identifying a target region to be modified. The computing system determines a motion of the target region's boundary across the set of video frames, and also interpolates pixel motion within the target region across the set of video frames. The computing system also inserts, responsive to user input, a reference frame into the set of video frames. The reference frame can include reference color data from a user-specified modification to the target region. The computing system can use the reference color data and the interpolated motion to update color data in the target region across set of video frames.

    Skin tone assisted digital image color matching

    公开(公告)号:US10936853B1

    公开(公告)日:2021-03-02

    申请号:US16593872

    申请日:2019-10-04

    申请人: Adobe Inc.

    IPC分类号: G06K9/00 G06T7/90

    摘要: In implementations of skin tone assisted digital image color matching, a device implements a color editing system, which includes a facial detection module to detect faces in an input image and in a reference image, and includes a skin tone model to determine a skin tone value reflective of a skin tone of each of the faces. A color matching module can be implemented to group the faces into one or more face groups based on the skin tone value of each of the faces, match a face group pair as an input image face group paired with a reference image face group, and generate a modified image from the input image based on color features of the reference image, the color features including face skin tones of the respective faces in the face group pair as part of the color features applied to modify the input image.

    TEXTURE INTERPOLATION USING NEURAL NETWORKS
    26.
    发明申请

    公开(公告)号:US20200342634A1

    公开(公告)日:2020-10-29

    申请号:US16392968

    申请日:2019-04-24

    申请人: Adobe Inc.

    IPC分类号: G06T11/00 G06T7/40

    摘要: Techniques are disclosed for neural network based interpolation of image textures. A methodology implementing the techniques according to an embodiment includes training a global encoder network to generate global latent vectors based on training texture images, and training a local encoder network to generate local latent tensors based on the training texture images. The method further includes interpolating between the global latent vectors associated with each set of training images, and interpolating between the local latent tensors associated with each set of training images. The method further includes training a decoder network to generate reconstructions of the training texture images and to generate an interpolated texture based on the interpolated global latent vectors and the interpolated local latent tensors. The training of the encoder and decoder networks is based on a minimization of a loss function of the reconstructions and a minimization of a loss function of the interpolated texture.

    3D object reconstruction using photometric mesh representation

    公开(公告)号:US10769848B1

    公开(公告)日:2020-09-08

    申请号:US16421729

    申请日:2019-05-24

    申请人: Adobe, Inc.

    摘要: Techniques are disclosed for 3D object reconstruction using photometric mesh representations. A decoder is pretrained to transform points sampled from 2D patches of representative objects into 3D polygonal meshes. An image frame of the object is fed into an encoder to get an initial latent code vector. For each frame and camera pair from the sequence, a polygonal mesh is rendered at the given viewpoints. The mesh is optimized by creating a virtual viewpoint, rasterized to obtain a depth map. The 3D mesh projections are aligned by projecting the coordinates corresponding to the polygonal face vertices of the rasterized mesh to both selected viewpoints. The photometric error is determined from RGB pixel intensities sampled from both frames. Gradients from the photometric error are backpropagated into the vertices of the assigned polygonal indices by relating the barycentric coordinates of each image to update the latent code vector.

    Patch validity test
    29.
    发明授权

    公开(公告)号:US10586311B2

    公开(公告)日:2020-03-10

    申请号:US15921457

    申请日:2018-03-14

    申请人: ADOBE INC.

    摘要: Embodiments of the present invention provide systems, methods, and computer storage media for improved patch validity testing for patch-based synthesis applications using similarity transforms. The improved patch validity tests are used to validate (or invalidate) candidate patches as valid patches falling within a sampling region of a source image. The improved patch validity tests include a hole dilation test for patch validity, a no-dilation test for patch invalidity, and a comprehensive pixel test for patch invalidity. A fringe test for range invalidity can be used to identify pixels with an invalid range and invalidate corresponding candidate patches. The fringe test for range invalidity can be performed as a precursor to any or all of the improved patch validity tests. In this manner, validated candidate patches are used to automatically reconstruct a target image.

    Image lighting transfer via multi-dimensional histogram matching

    公开(公告)号:US10521892B2

    公开(公告)日:2019-12-31

    申请号:US15253655

    申请日:2016-08-31

    申请人: ADOBE INC.

    摘要: Embodiments of the present invention provide systems, methods, and computer storage media directed at relighting a target image based on a lighting effect from a reference image. In one embodiment, a target image and a reference image are received, the reference image includes a lighting effect desired to be applied to the target image. A lighting transfer is performed using color data and geometrical data associated with the reference image and color data and geometrical data associated with the target image. The lighting transfer causes generation of a relit image that corresponds with the target image having a lighting effect of the reference image. The relit image is provided for display to a user via one or more output devices. Other embodiments may be described and/or claimed.