Avatar Creation From Natural Language Description

    公开(公告)号:US20250022187A1

    公开(公告)日:2025-01-16

    申请号:US18590707

    申请日:2024-02-28

    Inventor: Kun Jin Siva Penke

    Abstract: In one embodiment, a method includes accessing, by a computing device, a natural-language input comprising a description of an avatar and generating, from the natural-language input and by a trained avatar-creation model, a set of avatar-attribute feature vectors. The method further includes determining, by a trained avatar-attribute classifier, one or more avatar attributes from the set of avatar-attribute feature vectors; and generating the avatar based on the determined one or more avatar attributes for presentation on a display of a computing device.

    MACHINE LEARNING-BASED APPROACH FOR AUDIO-DRIVEN AVATAR ANIMATION OR OTHER FUNCTIONS

    公开(公告)号:US20240203014A1

    公开(公告)日:2024-06-20

    申请号:US18299248

    申请日:2023-04-12

    CPC classification number: G06T13/205 G06T13/40 G10L17/02 G10L17/04 G10L17/18

    Abstract: A method includes obtaining, using at least one processing device of an electronic device, an audio input associated with a speaker. The method also includes extracting, using a feature extractor of a trained machine learning model, audio features from the audio input. The method further includes generating (i) one or more content parameter predictions using content embeddings extracted by a content encoder and decoded by a content decoder of the trained machine learning model and (ii) one or more style parameter predictions using style embeddings extracted by a style encoder and decoded by a style decoder of the trained machine learning model. The content embeddings and the style embeddings are based on the audio features of the audio input. The trained machine learning model is trained to generate the one or more content parameter predictions and the one or more style parameter predictions using disentangled content and style embeddings.

    Real-Time Avatar Animation
    6.
    发明申请

    公开(公告)号:US20250104318A1

    公开(公告)日:2025-03-27

    申请号:US18601097

    申请日:2024-03-11

    Abstract: In one embodiment, a method includes accessing an audio input that includes a mixture of vocal sounds and non-vocal sounds and separating, by a trained audio source separation model, the audio input into a first audio output representing the vocal sounds and a second audio output representing the non-vocal sounds. The method further includes determining, by one or more trained avatar animation models and by separately encoding the first audio output representing the vocal sounds and the second audio output representing the non-vocal sounds, an avatar animation temporally corresponding to the audio input; and rendering, in real time and temporally coincident with the audio input, the determined avatar animation.

    METHOD FOR TRANSMITTING PHONETIC DATA
    8.
    发明申请
    METHOD FOR TRANSMITTING PHONETIC DATA 有权
    传输电话数据的方法

    公开(公告)号:US20150235638A1

    公开(公告)日:2015-08-20

    申请号:US14185198

    申请日:2014-02-20

    Inventor: Siva Penke

    CPC classification number: G10L19/0018 G10L2015/025

    Abstract: A method for transmitting and receiving a communication is provided. The method of receiving a communication includes detecting a selection of a phonetic function which requests or causes inclusion of at least one phonetic data file in the communication, and receiving, in the communication, the one or more phonetic data file.

    Abstract translation: 提供了一种用于发送和接收通信的方法。 接收通信的方法包括检测在通信中请求或导致包含至少一个语音数据文件的语音功能的选择,以及在通信中接收一个或多个语音数据文件。

    AUTOMATIC 3D AVATAR GENERATION BY MESH DEFORMATION BASED ON OPTIMAL TRANSPORT FUNCTION

    公开(公告)号:US20240355052A1

    公开(公告)日:2024-10-24

    申请号:US18610772

    申请日:2024-03-20

    Inventor: Kun Jin Siva Penke

    CPC classification number: G06T17/20

    Abstract: A method includes obtaining, using at least one processing device of an electronic device, an identification of multiple three-dimensional (3D) objects of interest. The method also includes generating, using the at least one processing device, multiple intermediate 3D objects based on the 3D objects of interest using optimal transport, where the intermediate 3D objects are generated using interpolation or extrapolation based on shapes of the 3D objects of interest. The method further includes presenting, using the at least one processing device, one or more of the intermediate 3D objects to a user.

Patent Agency Ranking