HIGH-LEVEL SYNTAX FOR SIGNALING NEURAL NETWORKS WITHIN A MEDIA BITSTREAM

    公开(公告)号:US20220256227A1

    公开(公告)日:2022-08-11

    申请号:US17649915

    申请日:2022-02-03

    Abstract: An example method is provided to include receiving a media bitstream comprising one or more media units and a first enhancement information message, wherein the first enhancement information message comprises at least two independently parsable structures, a first independently parsable structure comprising information about at least one purpose of one or more neural networks (NNs) to be applied to the one or more media units, and a second independently parsable structure comprising or identifying one or more neural networks; decoding the one or more media units; and using the one or more neural networks to enhance or filter one or more frames of the decoded the one or more media units, based on the at least one purpose. An example method includes. Corresponding apparatuses and computer program products are also provided.

    METHOD, APPARATUS AND COMPUTER PROGRAM PRODUCT FOR PROVIDING AN ATTENTION BLOCK FOR NEURAL NETWORK-BASED IMAGE AND VIDEO COMPRESSION

    公开(公告)号:US20240289590A1

    公开(公告)日:2024-08-29

    申请号:US18572100

    申请日:2022-06-16

    CPC classification number: G06N3/045

    Abstract: Various embodiments provide a method, an apparatus, and computer program product. The method comprising: defining an attention block comprising: a set of initial neural network layers, wherein each layer is caused to process an output of a previous layer, and wherein a first layer processes an input of a dense split attention block; core attention blocks process one or more outputs of the set of initial neural network layers; a concatenation block for concatenating one or more outputs of the core attention blocks and at least one intermediate output of the set of initial neural network layers; one or more final neural network layers process at least the output of the concatenation block; and a summation block caused to sum an output of the final neural network layers and an input to the attention block; and providing an output of the summation block as a final output of the attention block.

Patent Agency Ranking