AUDIO-BASED MEDIA EDIT POINT SELECTION
    11.
    发明公开

    公开(公告)号:US20240112668A1

    公开(公告)日:2024-04-04

    申请号:US18529170

    申请日:2023-12-05

    Applicant: Adobe Inc.

    CPC classification number: G10L15/04 G06F40/253 G06F40/30

    Abstract: A media edit point selection process can include a media editing software application programmatically converting speech to text and storing a timestamp-to-text map. The map correlates text corresponding to speech extracted from an audio track for the media clip to timestamps for the media clip. The timestamps correspond to words and some gaps in the speech from the audio track. The probability of identified gaps corresponding to a grammatical pause by the speaker is determined using the timestamp-to-text map and a semantic model. Potential edit points corresponding to grammatical pauses in the speech are stored for display or for additional use by the media editing software application. Text can optionally be displayed to a user during media editing.

    ADAPTIVE SEARCH RESULTS FOR MULTIMEDIA SEARCH QUERIES

    公开(公告)号:US20220414149A1

    公开(公告)日:2022-12-29

    申请号:US17902457

    申请日:2022-09-02

    Applicant: Adobe Inc.

    Abstract: A system identifies a video comprising frames associated with content tags. The system detects features for each frame of the video. The system identifies, based on the detected features, scenes of the video. The system determines, for each frame for each scene, a frame score that indicates a number of content tags that match the other frames within the scene. The system selects, for each scene, a set of key frames that represent the scene based on the determined frame scores. The system receives a search query comprising a keyword. The system generates, for display, search results responsive to the search query including a dynamic preview of the video. The dynamic preview comprises an arrangement of frames of the video corresponding to each scene of the video. Each of the arrangement of frames is selected from the selected set of key frames representing the respective scene of the video.

    IMAGE PROCESSING TECHNIQUES TO QUICKLY FIND A DESIRED OBJECT AMONG OTHER OBJECTS FROM A CAPTURED VIDEO SCENE

    公开(公告)号:US20220343647A1

    公开(公告)日:2022-10-27

    申请号:US17811897

    申请日:2022-07-12

    Applicant: Adobe Inc.

    Abstract: Techniques are provided for identifying objects (such as products within a physical store) within a captured video scene and indicating which of object in the captured scene matches a desired object requested by a user. The matching object is then displayed in an accentuated manner to the user in real-time (via augmented reality). Object identification is carried out via a multimodal methodology. Objects within the captured video scene are identified using a neural network trained to identify different types of objects. The identified objects can then be compared against a database of pre-stored images of the desired product to determine if a close match is found. Additionally, text on the identified objects is analyzed and compared to the text of the desired object. Based on either or both identification methods, the desired object is indicated to the user on their display, via an augmented reality graphic.

    Providing context aware video searching

    公开(公告)号:US10998007B2

    公开(公告)日:2021-05-04

    申请号:US16588662

    申请日:2019-09-30

    Applicant: Adobe Inc.

    Abstract: This disclosure relates to methods, non-transitory computer readable media, and systems that can generate a context-aware-video-progress bar including a video-scene-proportionate timeline with time-interval sections sized according to relative scene proportions within time intervals of a video. In some implementations, for instance, the disclosed systems determine relative proportions of scenes within a video across time intervals of the video and generate a video-scene-proportionate timeline comprising time-interval sections sized proportionate to the relative proportions of scenes across the time intervals. By integrating the video-scene-proportionate timeline within a video-progress bar, the disclosed systems generate a context-aware-video-progress bar for a video. Such a context-aware-video-progress bar can facilitate more precise and intelligent scrubbing through a video, a dynamic graphical user interface for navigating within and identifying frames of the video, and a flexible user-friendly tool for quickly identifying scenes.

    GENERATING AND PROVIDING COMPOSITION EFFECT TUTORIALS FOR CREATING AND EDITING DIGITAL CONTENT

    公开(公告)号:US20200082586A1

    公开(公告)日:2020-03-12

    申请号:US16128904

    申请日:2018-09-12

    Applicant: Adobe Inc.

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for generating and providing composition effect tutorials for creating and editing digital content based on a metadata composite structure. For example, the disclosed systems can generate and/or access a metadata composite structure that includes nodes corresponding to composition effects applied to a digital content item, where a given node can include location information indicating where a composition effect is applied relative to a digital content item. The disclosed systems can further generate a tutorial to guide a user to implement a selected composition effect by identifying composition effects of nodes that correspond to a location selected within a composition interface and presenting instructions for a particular composition effect.

    Adaptive search results for multimedia search queries

    公开(公告)号:US11941049B2

    公开(公告)日:2024-03-26

    申请号:US17902457

    申请日:2022-09-02

    Applicant: Adobe Inc.

    Abstract: A system identifies a video comprising frames associated with content tags. The system detects features for each frame of the video. The system identifies, based on the detected features, scenes of the video. The system determines, for each frame for each scene, a frame score that indicates a number of content tags that match the other frames within the scene. The system selects, for each scene, a set of key frames that represent the scene based on the determined frame scores. The system receives a search query comprising a keyword. The system generates, for display, search results responsive to the search query including a dynamic preview of the video. The dynamic preview comprises an arrangement of frames of the video corresponding to each scene of the video. Each of the arrangement of frames is selected from the selected set of key frames representing the respective scene of the video.

    Editing digital images using multi-panel graphical user interfaces

    公开(公告)号:US11294556B1

    公开(公告)日:2022-04-05

    申请号:US17141745

    申请日:2021-01-05

    Applicant: Adobe Inc.

    Inventor: Amol Jindal

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media that utilize a multi-panel graphical user interface for modifying digital images. For example, in one or more embodiments, the disclosed systems divide the graphical user interface of a client device into first panel and a second panel. Further, the disclosed systems provide different portions of a digital image for display within the first and second panels. In some implementations, the disclosed systems receive a user interaction with the portion of the digital image displayed within the first panel. Based on the received user interaction, the disclosed systems modify the second portion of the digital image displayed within the second panel.

    AUDIO-BASED SELECTION OF EDIT POINTS FOR VIDEO CLIPS

    公开(公告)号:US20220068258A1

    公开(公告)日:2022-03-03

    申请号:US17008427

    申请日:2020-08-31

    Applicant: Adobe Inc.

    Abstract: A media edit point selection process can include a media editing software application programmatically converting speech to text and storing a timestamp-to-text map. The map correlates text corresponding to speech extracted from an audio track for the media clip to timestamps for the media clip. The timestamps correspond to words and some gaps in the speech from the audio track. The probability of identified gaps corresponding to a grammatical pause by the speaker is determined using the timestamp-to-text map and a semantic model. Potential edit points corresponding to grammatical pauses in the speech are stored for display or for additional use by the media editing software application. Text can optionally be displayed to a user during media editing.

Patent Agency Ranking