-
公开(公告)号:US20220122222A1
公开(公告)日:2022-04-21
申请号:US17384283
申请日:2021-07-23
申请人: Adobe Inc.
发明人: Cameron Smith , Ratheesh Kalarot , Wei-An Lin , Richard Zhang , Niloy Mitra , Elya Shechtman , Shabnam Ghadar , Zhixin Shu , Yannick Hold-Geoffrey , Nathan Carr , Jingwan Lu , Oliver Wang , Jun-Yan Zhu
摘要: An improved system architecture uses a Generative Adversarial Network (GAN) including a specialized generator neural network to generate multiple resolution output images. The system produces a latent space representation of an input image. The system generates a first output image at a first resolution by providing the latent space representation of the input image as input to a generator neural network comprising an input layer, an output layer, and a plurality of intermediate layers and taking the first output image from an intermediate layer, of the plurality of intermediate layers of the generator neural network. The system generates a second output image at a second resolution different from the first resolution by providing the latent space representation of the input image as input to the generator neural network and taking the second output image from the output layer of the generator neural network.
-
公开(公告)号:US20220121932A1
公开(公告)日:2022-04-21
申请号:US17384378
申请日:2021-07-23
申请人: Adobe Inc.
发明人: Ratheesh Kalarot , Wei-An Lin , Cameron Smith , Zhixin Shu , Baldo Faieta , Shabnam Ghadar , Jingwan Lu , Aliakbar Darabi , Jun-Yan Zhu , Niloy Mitra , Richard Zhang , Elya Shechtman
摘要: Systems and methods train an encoder neural network for fast and accurate projection into the latent space of a Generative Adversarial Network (GAN). The encoder is trained by providing an input training image to the encoder and producing, by the encoder, a latent space representation of the input training image. The latent space representation is provided as input to the GAN to generate a generated training image. A latent code is sampled from a latent space associated with the GAN and the sampled latent code is provided as input to the GAN. The GAN generates a synthetic training image based on the sampled latent code. The sampled latent code is provided as input to the encoder to produce a synthetic training code. The encoder is updated by minimizing a loss between the generated training image and the input training image, and the synthetic training code and the sampled latent code.
-
公开(公告)号:US20210358170A1
公开(公告)日:2021-11-18
申请号:US17387207
申请日:2021-07-28
申请人: Adobe Inc.
摘要: The present disclosure relates to systems, non-transitory computer-readable media, and methods for utilizing a critical edge detection neural network and a geometric model to determine camera parameters from a single digital image. In particular, in one or more embodiments, the disclosed systems can train and utilize a critical edge detection neural network to generate a vanishing edge map indicating vanishing lines from the digital image. The system can then utilize the vanishing edge map to more accurately and efficiently determine camera parameters by applying a geometric model to the vanishing edge map. Further, the system can generate ground truth vanishing line data from a set of training digital images for training the critical edge detection neural network.
-
公开(公告)号:US20210287007A1
公开(公告)日:2021-09-16
申请号:US16817100
申请日:2020-03-12
申请人: Adobe Inc.
发明人: Oliver Wang , Matthew Fisher , John Nelson , Geoffrey Oxholm , Elya Shechtman , Wenqi Xian
摘要: Certain aspects involve video inpainting in which content is propagated from a user-provided reference frame to other video frames depicting a scene. For example, a computing system accesses a set of video frames with annotations identifying a target region to be modified. The computing system determines a motion of the target region's boundary across the set of video frames, and also interpolates pixel motion within the target region across the set of video frames. The computing system also inserts, responsive to user input, a reference frame into the set of video frames. The reference frame can include reference color data from a user-specified modification to the target region. The computing system can use the reference color data and the interpolated motion to update color data in the target region across set of video frames.
-
公开(公告)号:US10936853B1
公开(公告)日:2021-03-02
申请号:US16593872
申请日:2019-10-04
申请人: Adobe Inc.
发明人: Kartik Sethi , Oliver Wang , Tharun Mohandoss , Elya Shechtman , Chetan Nanda
摘要: In implementations of skin tone assisted digital image color matching, a device implements a color editing system, which includes a facial detection module to detect faces in an input image and in a reference image, and includes a skin tone model to determine a skin tone value reflective of a skin tone of each of the faces. A color matching module can be implemented to group the faces into one or more face groups based on the skin tone value of each of the faces, match a face group pair as an input image face group paired with a reference image face group, and generate a modified image from the input image based on color features of the reference image, the color features including face skin tones of the respective faces in the face group pair as part of the color features applied to modify the input image.
-
公开(公告)号:US20200342634A1
公开(公告)日:2020-10-29
申请号:US16392968
申请日:2019-04-24
申请人: Adobe Inc.
发明人: Connelly Barnes , Sohrab Amirghodsi , Michal Lukac , Elya Shechtman , Ning Yu
摘要: Techniques are disclosed for neural network based interpolation of image textures. A methodology implementing the techniques according to an embodiment includes training a global encoder network to generate global latent vectors based on training texture images, and training a local encoder network to generate local latent tensors based on the training texture images. The method further includes interpolating between the global latent vectors associated with each set of training images, and interpolating between the local latent tensors associated with each set of training images. The method further includes training a decoder network to generate reconstructions of the training texture images and to generate an interpolated texture based on the interpolated global latent vectors and the interpolated local latent tensors. The training of the encoder and decoder networks is based on a minimization of a loss function of the reconstructions and a minimization of a loss function of the interpolated texture.
-
公开(公告)号:US10769848B1
公开(公告)日:2020-09-08
申请号:US16421729
申请日:2019-05-24
申请人: Adobe, Inc.
发明人: Oliver Wang , Vladimir Kim , Matthew Fisher , Elya Shechtman , Chen-Hsuan Lin , Bryan Russell
摘要: Techniques are disclosed for 3D object reconstruction using photometric mesh representations. A decoder is pretrained to transform points sampled from 2D patches of representative objects into 3D polygonal meshes. An image frame of the object is fed into an encoder to get an initial latent code vector. For each frame and camera pair from the sequence, a polygonal mesh is rendered at the given viewpoints. The mesh is optimized by creating a virtual viewpoint, rasterized to obtain a depth map. The 3D mesh projections are aligned by projecting the coordinates corresponding to the polygonal face vertices of the rasterized mesh to both selected viewpoints. The photometric error is determined from RGB pixel intensities sampled from both frames. Gradients from the photometric error are backpropagated into the vertices of the assigned polygonal indices by relating the barycentric coordinates of each image to update the latent code vector.
-
公开(公告)号:US20200082591A1
公开(公告)日:2020-03-12
申请号:US16681218
申请日:2019-11-12
申请人: Adobe Inc.
发明人: David Simons , Michal Lukac , Daniel Sykora , Elya Shechtman , Paul Asente , Jingwan Lu , Jakub Fiser , Ondrej Jamriska
IPC分类号: G06T13/40 , G06T13/80 , G06T11/00 , G06T5/50 , G06T5/00 , G06T3/00 , G06T7/90 , G06T7/73 , G06T7/11 , G06T7/194 , G06K9/00 , G06T15/04
摘要: Certain embodiments involve generating one or more of appearance guide and a positional guide and using one or more of the guides to synthesize a stylized image or animation. For example, a system obtains data indicating a target image and a style exemplar image. The system generates an appearance guide, a positional guide, or both from the target image and the style exemplar image. The system uses one or more of the guides to transfer a texture or style from the style exemplar image to the target image.
-
公开(公告)号:US10586311B2
公开(公告)日:2020-03-10
申请号:US15921457
申请日:2018-03-14
申请人: ADOBE INC.
摘要: Embodiments of the present invention provide systems, methods, and computer storage media for improved patch validity testing for patch-based synthesis applications using similarity transforms. The improved patch validity tests are used to validate (or invalidate) candidate patches as valid patches falling within a sampling region of a source image. The improved patch validity tests include a hole dilation test for patch validity, a no-dilation test for patch invalidity, and a comprehensive pixel test for patch invalidity. A fringe test for range invalidity can be used to identify pixels with an invalid range and invalidate corresponding candidate patches. The fringe test for range invalidity can be performed as a precursor to any or all of the improved patch validity tests. In this manner, validated candidate patches are used to automatically reconstruct a target image.
-
公开(公告)号:US10521892B2
公开(公告)日:2019-12-31
申请号:US15253655
申请日:2016-08-31
申请人: ADOBE INC.
发明人: Kalyan K. Sunkavalli , Sunil Hadap , Elya Shechtman , Zhixin Shu
摘要: Embodiments of the present invention provide systems, methods, and computer storage media directed at relighting a target image based on a lighting effect from a reference image. In one embodiment, a target image and a reference image are received, the reference image includes a lighting effect desired to be applied to the target image. A lighting transfer is performed using color data and geometrical data associated with the reference image and color data and geometrical data associated with the target image. The lighting transfer causes generation of a relit image that corresponds with the target image having a lighting effect of the reference image. The relit image is provided for display to a user via one or more output devices. Other embodiments may be described and/or claimed.
-
-
-
-
-
-
-
-
-