COMPRESSING IMAGE-TO-IMAGE MODELS WITH AVERAGE SMOOTHING

    公开(公告)号:US20250054199A1

    公开(公告)日:2025-02-13

    申请号:US18923108

    申请日:2024-10-22

    Applicant: Snap Inc.

    Abstract: System and methods for compressing image-to-image models. Generative Adversarial Networks (GANs) have achieved success in generating high-fidelity images. An image compression system and method adds a novel variant to class-dependent parameters (CLADE), referred to as CLADE-Avg, which recovers the image quality without introducing extra computational cost. An extra layer of average smoothing is performed between the parameter and normalization layers. Compared to CLADE, this image compression system and method smooths abrupt boundaries, and introduces more possible values for the scaling and shift. In addition, the kernel size for the average smoothing can be selected as a hyperparameter, such as a 3×3 kernel size. This method does not introduce extra multiplications but only addition, and thus does not introduce much computational overhead, as the division can be absorbed into the parameters after training.

    Motion representations for articulated animation

    公开(公告)号:US11798213B2

    公开(公告)日:2023-10-24

    申请号:US17364218

    申请日:2021-06-30

    Applicant: Snap Inc.

    Abstract: Systems and methods herein describe novel motion representations for animating articulated objects consisting of distinct parts. The described systems and method access source image data, identify driving image data to modify image feature data in the source image sequence data, generate, using an image transformation neural network, modified source image data comprising a plurality of modified source images depicting modified versions of the image feature data, the image transformation neural network being trained to identify, for each image in the source image data, a driving image from the driving image data, the identified driving image being implemented by the image transformation neural network to modify a corresponding source image in the source image data using motion estimation differences between the identified driving image and the corresponding source image, and stores the modified source image data.

    Messaging system with neural hair rendering

    公开(公告)号:US12272015B2

    公开(公告)日:2025-04-08

    申请号:US18653609

    申请日:2024-05-02

    Applicant: Snap Inc.

    Abstract: A messaging system performs neural network hair rendering for images provided by users of the messaging system. A method of neural network hair rendering includes processing a three-dimensional (3D) model of fake hair and a first real hair image depicting a first person to generate a fake hair structure, and encoding, using a fake hair encoder neural subnetwork, the fake hair structure to generate a coded fake hair structure. The method further includes processing, using a cross-domain structure embedding neural subnetwork, the coded fake hair structure to generate a fake and real hair structure, and encoding, using an appearance encoder neural subnetwork, a second real hair image depicting a second person having a second head to generate an appearance map. The method further includes processing, using a real appearance renderer neural subnetwork, the appearance map and the fake and real hair structure to generate a synthesized real image.

    AUTOMATIC IMAGE QUALITY EVALUATION
    7.
    发明公开

    公开(公告)号:US20240296535A1

    公开(公告)日:2024-09-05

    申请号:US18176843

    申请日:2023-03-01

    Applicant: Snap Inc.

    Abstract: Examples disclosed herein describe techniques for automatic image quality evaluation. A first set of images generated by a first automated image generator and a second set of images generated by a second automated image generator are accessed. A first machine learning model generates a first quality indicator for each image in the first set of images and the second set of images. A second machine learning model generates a second quality indicator for each image in the first set of images and the second set of images. Based on the generated indicators, a first image from the first set of images and a second image from the second set of images are automatically selected and compared. A first ranking of the first automated image generator and the second automated image generator is generated based on the comparison, and ranking data is caused to be presented on a device.

    3D MODELING BASED ON NEURAL LIGHT FIELD
    8.
    发明公开

    公开(公告)号:US20240273809A1

    公开(公告)日:2024-08-15

    申请号:US18644653

    申请日:2024-04-24

    Applicant: Snap Inc.

    CPC classification number: G06T15/06 G06T7/97 G06T2207/20081 G06T2207/20084

    Abstract: Methods and systems are disclosed for performing operations for generating a 3D model of a scene. The operations include: receiving a set of two-dimensional (2D) images representing a first view of a real-world environment; applying a machine learning model comprising a neural light field network to the set of 2D images to predict pixel values of a target image representing a second view of the real-world environment, the machine learning model being trained to map a ray origin and direction directly to a given pixel value; and generating a three-dimensional (3D) model of the real-world environment based on the set of 2D images and the predicted target image.

    3D modeling based on neural light field

    公开(公告)号:US12002146B2

    公开(公告)日:2024-06-04

    申请号:US17656778

    申请日:2022-03-28

    Applicant: Snap Inc.

    CPC classification number: G06T15/06 G06T7/97 G06T2207/20081 G06T2207/20084

    Abstract: Methods and systems are disclosed for performing operations for generating a 3D model of a scene. The operations include: receiving a set of two-dimensional (2D) images representing a first view of a real-world environment; applying a machine learning model comprising a neural light field network to the set of 2D images to predict pixel values of a target image representing a second view of the real-world environment, the machine learning model being trained to map a ray origin and direction directly to a given pixel value; and generating a three-dimensional (3D) model of the real-world environment based on the set of 2D images and the predicted target image.

    CROSS-MODAL SHAPE AND COLOR MANIPULATION
    10.
    发明公开

    公开(公告)号:US20230386158A1

    公开(公告)日:2023-11-30

    申请号:US17814391

    申请日:2022-07-22

    Applicant: Snap Inc.

    CPC classification number: G06T19/20 G06T17/00 G06T2219/2012 G06T2219/2021

    Abstract: Systems, computer readable media, and methods herein describe an editing system where a three-dimensional (3D) object can be edited by editing a 2D sketch or 2D RGB views of the 3D object. The editing system uses multi-modal (MM) variational auto-decoders (VADs)(MM-VADs) that are trained with a shared latent space that enables editing 3D objects by editing 2D sketches of the 3D objects. The system determines a latent code that corresponds to an edited or sketched 2D sketch. The latent code is then used to generate a 3D object using the MM-VADs with the latent code as input. The latent space is divided into a latent space for shapes and a latent space for colors. The MM-VADs are trained with variational auto-encoders (VAE) and a ground truth.

Patent Agency Ranking