-
1.
公开(公告)号:US20200372284A1
公开(公告)日:2020-11-26
申请号:US16616235
申请日:2019-10-16
Applicant: Google LLC
Inventor: Christoph Rhemann , Abhimitra Meka , Matthew Whalen , Jessica Lynn Busch , Sofien Bouaziz , Geoffrey Douglas Harvey , Andrea Tagliasacchi , Jonathan Taylor , Paul Debevec , Peter Joseph Denny , Sean Ryan Francesco Fanello , Graham Fyffe , Jason Angelo Dourgarian , Xueming Yu , Adarsh Prakash Murthy Kowdle , Julien Pascal Christophe Valentin , Peter Christopher Lincoln , Rohit Kumar Pandey , Christian Häne , Shahram Izadi
Abstract: Methods, systems, and media for relighting images using predicted deep reflectance fields are provided. In some embodiments, the method comprises: identifying a group of training samples, wherein each training sample includes (i) a group of one-light-at-a-time (OLAT) images that have each been captured when one light of a plurality of lights arranged on a lighting structure has been activated, (ii) a group of spherical color gradient images that have each been captured when the plurality of lights arranged on the lighting structure have been activated to each emit a particular color, and (iii) a lighting direction, wherein each image in the group of OLAT images and each of the spherical color gradient images are an image of a subject, and wherein the lighting direction indicates a relative orientation of a light to the subject; training a convolutional neural network using the group of training samples, wherein training the convolutional neural network comprises: for each training iteration in a series of training iterations and for each training sample in the group of training samples: generating an output predicted image, wherein the output predicted image is a representation of the subject associated with the training sample with lighting from the lighting direction associated with the training sample; identifying a ground-truth OLAT image included in the group of OLAT images for the training sample that corresponds to the lighting direction for the training sample; calculating a loss that indicates a perceptual difference between the output predicted image and the identified ground-truth OLAT image; and updating parameters of the convolutional neural network based on the calculated loss; identifying a test sample that includes a second group of spherical color gradient images and a second lighting direction; and generating a relit image of the subject included in each of the second group of spherical color gradient images with lighting from the second lighting direction using the trained convolutional neural network.
-
2.
公开(公告)号:US10997457B2
公开(公告)日:2021-05-04
申请号:US16616235
申请日:2019-10-16
Applicant: Google LLC
Inventor: Christoph Rhemann , Abhimitra Meka , Matthew Whalen , Jessica Lynn Busch , Sofien Bouaziz , Geoffrey Douglas Harvey , Andrea Tagliasacchi , Jonathan Taylor , Paul Debevec , Peter Joseph Denny , Sean Ryan Francesco Fanello , Graham Fyffe , Jason Angelo Dourgarian , Xueming Yu , Adarsh Prakash Murthy Kowdle , Julien Pascal Christophe Valentin , Peter Christopher Lincoln , Rohit Kumar Pandey , Christian Häne , Shahram Izadi
Abstract: Methods, systems, and media for relighting images using predicted deep reflectance fields are provided. In some embodiments, the method comprises: identifying a group of training samples, wherein each training sample includes (i) a group of one-light-at-a-time (OLAT) images that have each been captured when one light of a plurality of lights arranged on a lighting structure has been activated, (ii) a group of spherical color gradient images that have each been captured when the plurality of lights arranged on the lighting structure have been activated to each emit a particular color, and (iii) a lighting direction, wherein each image in the group of OLAT images and each of the spherical color gradient images are an image of a subject, and wherein the lighting direction indicates a relative orientation of a light to the subject; training a convolutional neural network using the group of training samples, wherein training the convolutional neural network comprises: for each training iteration in a series of training iterations and for each training sample in the group of training samples: generating an output predicted image, wherein the output predicted image is a representation of the subject associated with the training sample with lighting from the lighting direction associated with the training sample; identifying a ground-truth OLAT image included in the group of OLAT images for the training sample that corresponds to the lighting direction for the training sample; calculating a loss that indicates a perceptual difference between the output predicted image and the identified ground-truth OLAT image; and updating parameters of the convolutional neural network based on the calculated loss; identifying a test sample that includes a second group of spherical color gradient images and a second lighting direction; and generating a relit image of the subject included in each of the second group of spherical color gradient images with lighting from the second lighting direction using the trained convolutional neural network.
-
公开(公告)号:US20240020915A1
公开(公告)日:2024-01-18
申请号:US18353213
申请日:2023-07-17
Applicant: Google LLC
Inventor: Yinda Zhang , Feitong Tan , Sean Ryan Francesco Fanello , Abhimitra Meka , Sergio Orts Escolano , Danhang Tang , Rohit Kumar Pandey , Jonathan James Taylor
Abstract: Techniques include introducing a neural generator configured to produce novel faces that can be rendered at free camera viewpoints (e.g., at any angle with respect to the camera) and relit under an arbitrary high dynamic range (HDR) light map. A neural implicit intrinsic field takes a randomly sampled latent vector as input and produces as output per-point albedo, volume density, and reflectance properties for any queried 3D location. These outputs are aggregated via a volumetric rendering to produce low resolution albedo, diffuse shading, specular shading, and neural feature maps. The low resolution maps are then upsampled to produce high resolution maps and input into a neural renderer to produce relit images.
-
公开(公告)号:US20240290025A1
公开(公告)日:2024-08-29
申请号:US18588948
申请日:2024-02-27
Applicant: GOOGLE LLC
Inventor: Yinda Zhang , Sean Ryan Francesco Fanello , Ziqian Bai , Feitong Tan , Zeng Huang , Kripasindhu Sarkar , Danhang Tang , Di Qiu , Abhimitra Meka , Ruofei Du , Mingsong Dou , Sergio Orts Escolano , Rohit Kumar Pandey , Thabo Beeler
CPC classification number: G06T13/40 , G06T7/90 , G06T17/20 , G06V10/44 , G06T2207/10024 , G06T2207/20084
Abstract: A method comprises receiving a first sequence of images of a portion of a user, the first sequence of images being monocular images; generating an avatar based on the first sequence of images, the avatar being based on a model including a feature vector associated with a vertex; receiving a second sequence of images of the portion of the user; and based on the second sequence of images, modifying the avatar with a displacement of the vertex to represent a gesture of the avatar.
-
5.
公开(公告)号:US20240029333A1
公开(公告)日:2024-01-25
申请号:US18355154
申请日:2023-07-19
Applicant: GOOGLE LLC
Inventor: Abhimitra Meka , Thabo Beeler , Franziska Müller , Gengyan Li , Marcel Bühler , Otmar Hilliges
CPC classification number: G06T13/40 , G06T17/20 , G06T15/50 , G06T2210/62 , G06T2210/44
Abstract: A method including selecting a first point from a 3D model representing an avatar, the first point being associated with an eye, selecting a second point from the 3D model, the second point being associated with a periocular region associated with the eye, generating an albedo and spherical harmonics (SH) coefficients based on the first point and the second point, and generating an image point based on the albedo, and the SH coefficients.
-
公开(公告)号:US20250111477A1
公开(公告)日:2025-04-03
申请号:US18477219
申请日:2023-09-28
Applicant: GOOGLE LLC
Inventor: Sergio Orts Escolano , Zhiwen Fan , Di Qiu , Yinda Zhang , Daoye Wang , Erroll Wood , Abhimitra Meka , Hossam Isack , Paulo Fabiano Urnau Gotardo , Kripasindhu Sarkar , Thabo Beeler , Zhengyang Shen , Alexander Sahba Koumis
Abstract: A method including capturing a first plurality of images that include a foreground object and a background, capturing a second plurality of images that include the background, generating an alpha matte based on the first plurality of images and the second plurality of images using a trained machine learned model trained using a loss function configured to cause the trained machine learned model to learn high-frequency details of the foreground object, generating a foreground object image based on the first plurality of images and the second plurality of images using the trained machine learned model, and synthesizing an image including the foreground object image and a second background scene using the alpha matte.
-
公开(公告)号:US20250078397A1
公开(公告)日:2025-03-06
申请号:US18823613
申请日:2024-09-03
Applicant: Google LLC
Inventor: Abhimitra Meka , Marcel Bühler , Kripasindhu Sarkar , Tanmay Shah , Gengyan Li , Daoye Wang , Leonhard Helminger , Sergio Orts Escolano , Dmitry Lagun , Thabo Beeler
Abstract: A method including determining a viewpoint, generating a first image using an image generator, the first image including an object in a first orientation based on the viewpoint, modifying the image generator based on a second orientation of the object, and generating a second image based on the first image using the modified image generator.
-
-
-
-
-
-