Dense feature scale detection for image matching

    公开(公告)号:US12198357B2

    公开(公告)日:2025-01-14

    申请号:US18367034

    申请日:2023-09-12

    Applicant: Snap Inc.

    Abstract: Dense feature scale detection can be implemented using multiple convolutional neural networks trained on scale data to more accurately and efficiently match pixels between images. An input image can be used to generate multiple scaled images. The multiple scaled images are input into a feature net, which outputs feature data for the multiple scaled images. An attention net is used to generate an attention map from the input image. The attention map assigns emphasis as a soft distribution to different scales based on texture analysis. The feature data and the attention data can be combined through a multiplication process and then summed to generate dense features for comparison.

    Acoustic neural network scene detection

    公开(公告)号:US11545170B2

    公开(公告)日:2023-01-03

    申请号:US17247137

    申请日:2020-12-01

    Applicant: Snap Inc.

    Abstract: An acoustic environment identification system is disclosed that can use neural networks to accurately identify environments. The acoustic environment identification system can use one or more convolutional neural networks to generate audio feature data. A recursive neural network can process the audio feature data to generate characterization data. The characterization data can be modified using a weighting system that weights signature data items. Classification neural networks can be used to generate a classification of an environment.

    IMAGE AND POINT CLOUD BASED TRACKING AND IN AUGMENTED REALITY SYSTEMS

    公开(公告)号:US20220406008A1

    公开(公告)日:2022-12-22

    申请号:US17856720

    申请日:2022-07-01

    Applicant: Snap Inc.

    Abstract: Systems and methods for image based location estimation are described. In one example embodiment, a first positioning system is used to generate a first position estimate. Point cloud data describing an environment is then accessed. A two-dimensional surface of an image of an environment is captured, and a portion of the image is matched to a portion of key points in the point cloud data. An augmented reality object is then aligned within one or more images of the environment based on the match of the point cloud with the image. In some embodiments, building façade data may additionally be used to determine a device location and place the augmented reality object within an image.

    Dense feature scale detection for image matching

    公开(公告)号:US11367205B1

    公开(公告)日:2022-06-21

    申请号:US16721483

    申请日:2019-12-19

    Applicant: Snap Inc.

    Abstract: Dense feature scale detection can be implemented using multiple convolutional neural networks trained on scale data to more accurately and efficiently match pixels between images. An input image can be used to generate multiple scaled images. The multiple scaled images are input into a feature net, which outputs feature data for the multiple scaled images. An attention net is used to generate an attention map from the input image. The attention map assigns emphasis as a soft distribution to different scales based on texture analysis. The feature data and the attention data can be combined through a multiplication process and then summed to generate dense features for comparison.

    Dense captioning with joint interference and visual context

    公开(公告)号:US11361489B2

    公开(公告)日:2022-06-14

    申请号:US16946346

    申请日:2020-06-17

    Applicant: Snap Inc.

    Abstract: A dense captioning system and method is provided for analyzing an image to generate proposed bounding regions for a plurality of visual concepts within the image, generating a region feature for each proposed bounding region to generate a plurality of region features of the image, and determining a context feature for the image using a proposed bounding region that is a largest in size of the proposed bounding regions. For each region feature of the plurality of region features of the image, the dense captioning system and method further provides for analyzing the region feature to determine for the region feature a detection score that indicates a likelihood that the region feature comprises an actual object, and generating a caption for a visual concept in the image using the region feature and the context feature when a detection score is above a specified threshold value.

Patent Agency Ranking