Image processing method, storage medium, and computer device

    公开(公告)号:US11276207B2

    公开(公告)日:2022-03-15

    申请号:US16880883

    申请日:2020-05-21

    IPC分类号: G06T11/00 G06T3/40 G06T9/00

    摘要: An image processing method for a computer device. The method includes obtaining a to-be-processed image belonging to a first image category; inputting the to-be-processed image into a first stage image conversion model, to obtain a first intermediate image; and converting the first intermediate image into a second intermediate image through a second stage image conversion model. The method also includes determining a first weight matrix corresponding to the first intermediate image; determining a second weight matrix corresponding to the second intermediate image; and fusing the first intermediate image and the second intermediate image according to the corresponding first weight matrix and second weight matrix, to obtain a target image corresponding to the to-be-processed image and belonging to a second image category. A sum of the first weight matrix and the second weight matrix being a preset matrix.

    Sound-picture matching method of video, related apparatus, and storage medium

    公开(公告)号:US11972778B2

    公开(公告)日:2024-04-30

    申请号:US17712060

    申请日:2022-04-01

    摘要: A video sound-picture matching includes: acquiring a voice sequence; acquiring a voice segment from the voice sequence; acquiring an initial position of a start-stop mark and a moving direction of the start-stop mark from an image sequence; determining an active segment according to the initial position of the start-stop mark, the moving direction of the start-stop mark, and the voice segment; and synthesizing the voice segment and the active segment to obtain a video segment. In a video synthesizing process, the present disclosure uses start-stop marks to locate positions of active segments in an image sequence, so as to match the active segments having actions with voice segments, so that the synthesized video segments are more in line with natural laws of a character during speaking, and have better authenticity.

    Image fusion method, model training method, and related apparatuses

    公开(公告)号:US11776097B2

    公开(公告)日:2023-10-03

    申请号:US17336561

    申请日:2021-06-02

    IPC分类号: G06T5/50 G06N3/04 G06N3/08

    摘要: Methods, devices, and storage medium for fusing at least one image are disclosed. The method includes obtaining a first to-be-fused image and a second to-be-fused image, the first to-be-fused image comprising first regions, and the second to-be-fused image comprising second regions; obtaining a first feature set according to the first to-be-fused image and obtaining a second feature set according to the second to-be-fused image; performing first fusion processing on the first to-be-fused image and the second to-be-fused image by using a shape fusion network model to obtain a third to-be-fused image, the third to-be-fused image comprising at least one first encoding feature and at least one second encoding feature; and performing second fusion processing on the third to-be-fused image and the first to-be-fused image by using a condition fusion network model to obtain a target fused image. Model training methods, apparatus, and storage medium are also disclosed.

    Video synthesis method, model training method, device, and storage medium

    公开(公告)号:US11356619B2

    公开(公告)日:2022-06-07

    申请号:US17239438

    申请日:2021-04-23

    摘要: Embodiments of this application disclose methods, systems, and devices for video synthesis. In one aspect, a method comprises obtaining a plurality of frames corresponding to source image information of a first to-be-synthesized video, each frame of the source image information. The method also comprises obtaining a plurality of frames corresponding to target image information of a second to-be-synthesized video. For each frame of the plurality of frames corresponding to the target image information of the second to-be-synthesized video, the method comprises fusing a respective source image from the first to-be-synthesized video, a corresponding source motion key point, and a respective target motion key point corresponding to the frame using a pre-trained video synthesis model, and generating a respective output image in accordance with the fusing. The method further comprises repeating the fusing and the generating steps for the second to-be-synthesized video to produce a synthesized video.