-
公开(公告)号:US20250148682A1
公开(公告)日:2025-05-08
申请号:US18938478
申请日:2024-11-06
Applicant: Korea Electronics Technology Institute
Inventor: Choong Il KIM , Sa Im SHIN , Jin Yea JANG , San KIM , Byung Gill JOE
Abstract: Provided are a method and a system for generating a human animation from SMPL series parameters. An animation generation method according to an embodiment includes: extracting, by a motion extraction module, motion information from an image content; converting, by a motion conversion module, the extracted motion information into parameters required by an animation generation module; and generating, by the animation generation module, a motion image of a human object from the converted motion information. Accordingly, an animation accurately showing motions of an object appearing in an image content by a more realistic 3D model may be generated and digital human contents of high reality may be created and used.
-
公开(公告)号:US20220230640A1
公开(公告)日:2022-07-21
申请号:US17716445
申请日:2022-04-08
Applicant: KOREA ELECTRONICS TECHNOLOGY INSTITUTE
Inventor: Jin Yea JANG , Min Young JUNG , San KIM , Sa Im SHIN
Abstract: This application relates to an apparatus for adaptive conversation. In one aspect, the apparatus includes a server communication circuit that forms a communication channel with a user terminal and a server processor functionally connected to the server communication circuit. The server processor may receive, from the user terminal, a user utterance and surrounding external information acquired at a point in time at which the user utterance is collected. The server processor may also generate one word input by combining the surrounding external information with input information generated by performing natural language processing on the user utterance. The server processor may further generate a response sentence by applying the word input to a neural network model, and transmit the response sentence to the user terminal.
-
3.
公开(公告)号:US20250139153A1
公开(公告)日:2025-05-01
申请号:US18927089
申请日:2024-10-25
Applicant: Korea Electronics Technology Institute
Inventor: San KIM , Sa Im SHIN , Jin Yea JANG , Byung Gill JOE
IPC: G06F16/432 , G06F40/35
Abstract: There is provided a multi-modal knowledge-based conversation data generation and additional information labeling system. A conversation data generation method according to an embodiment includes: receiving input of a user utterance; searching pieces of text knowledge related to the user utterance; searching images related to the user utterance; receiving input of an answer to the user utterance referring to the searched text knowledge and images; and collecting the user utterances, the pieces of text knowledge, the images, and the answer to the user utterances as conversation data. Accordingly, open domain knowledge-based conversation data may be established by using both images and texts, and an image may be used as utterance and information acquired from the image may be used as utterance, so that AI conversation data similar to actual conversations can be established.
-
公开(公告)号:US20230069655A1
公开(公告)日:2023-03-02
申请号:US17894061
申请日:2022-08-23
Applicant: Korea Electronics Technology Institute
Inventor: Jin Yea JANG , Sa Im SHIN , Min Young JUNG , San KIM
Abstract: There are provided a method and a system for automatically generating a back-channel in an interactive agent system. According to an embodiment of the disclosure, an automatic back-channel generation method includes: predicting a back-channel by analyzing an utterance of a user inputted in a back-channel prediction model; and generating the predicted back-channel, and the back-channel prediction model is an AI model that is trained to predict a back-channel to express from the utterance of the user. Accordingly, a back-channel is automatically generated by utilizing a back-channel prediction module which is based on a language model, so that a natural dialogue interaction with a user may be implemented in an interactive agent system, and quality of a dialogue service provided to a user may be enhanced.
-
公开(公告)号:US20220415009A1
公开(公告)日:2022-12-29
申请号:US17563205
申请日:2021-12-28
Applicant: Korea Electronics Technology Institute
Inventor: Han Mu PARK , Jin Yea JANG , Yoon Young JEONG , Sa Im SHIN
IPC: G06V10/42 , G09B21/00 , G06V10/774 , G06V10/84
Abstract: There are provided a method for segmenting a sign language video by gloss to recognize a sign language sentence, and a method for training. According to an embodiment, a sign language video segmentation method receives an input of a sign language sentence video, and segments the inputted sign language sentence video by gloss. Accordingly, there is suggested a method for segmenting a sign language sentence video by gloss, analyzing various gloss sequences from the linguistic perspective, understanding meanings robustly in spite of various changes in sentences, and translating sign language into appropriate Korean sentences.
-
公开(公告)号:US20220245970A1
公开(公告)日:2022-08-04
申请号:US17726325
申请日:2022-04-21
Applicant: KOREA ELECTRONICS TECHNOLOGY INSTITUTE
Inventor: Min Young JUNG , Sa Im SHIN , Jin Yea JANG , San KIM
Abstract: This application relates to an adaptive inference system and an operation method therefor. In one aspect, the system includes a user terminal for collecting multi-modal information including at least visual information, voice information and text information. The system may also include an inference support device for receiving the multi-modal information from the user terminal, and inferring the intention of a user on the basis of pre-stored history information related to the user terminal, individualized information and the multi-modal information.
-
公开(公告)号:US20240420691A1
公开(公告)日:2024-12-19
申请号:US18704593
申请日:2023-08-23
Applicant: Korea Electronics Technology Institute
Inventor: Min Young JUNG , Sa Im SHIN , Jin Yea JANG , San KIM
IPC: G10L15/183 , G10L13/02
Abstract: There is provided a conversation method and system for operating a conversation model according to the presence or absence of related knowledge. A conversation method according to an embodiment may receive an input of a user utterance, may determine a conversation model based on an inputted user utterance content, may create a conversation content by using the determined conversation model, and may convert the created conversation content into a voice, and may output the voice. Accordingly, different conversation models may be operated according to the presence or absence of related knowledge, so that an empathic conversation may be generated instead of generating an inappropriate response in a knowledge-based conversation, and user's satisfaction on a conversation may be enhanced.
-
公开(公告)号:US20220415093A1
公开(公告)日:2022-12-29
申请号:US17851639
申请日:2022-06-28
Applicant: Korea Electronics Technology Institute
Inventor: Han Mu PARK , Jin Yea JANG , Sa Im SHIN
IPC: G06V40/20 , G06V10/774 , G06V20/40 , G06V40/10
Abstract: There are provided a method and a system for recognizing a finger language video in units of syllables based on AI. The finger language video recognition system includes: an extraction unit configured to extract posture information of a speaker from a finger language video; and a recognition unit configured to recognize a finger language of the speaker from the extracted posture information of the speaker in units of syllables, and to output a text. Accordingly, a language text in units of syllables may be generated from a finger language video, by using an AI-based syllable unit finger language recognition model.
-
9.
公开(公告)号:US20240249558A1
公开(公告)日:2024-07-25
申请号:US18289958
申请日:2022-06-20
Applicant: Korea Electronics Technology Institute
Inventor: Han Mu PARK , Jin Yea JANG , Sa Im SHIN
IPC: G06V40/20 , G06V10/774 , G06V20/40 , G06V40/16
CPC classification number: G06V40/28 , G06V10/774 , G06V20/46 , G06V20/49 , G06V40/175
Abstract: Provided are a method for segmenting sign language into morphemes, a method for predicting morpheme positions, and a method for augmenting data. A system for analyzing sign language according to an embodiment of the present invention comprises: a recognition unit which recognizes key points of a speaker from a sign language video; and a prediction unit which inputs the recognized key points into an artificial intelligence model, segments the sign language into morphemes, and predicts position information of the segmented morphemes. Accordingly, by recognizing the morphemes of the sign language video frame by frame on the basis of a skeletal model and thereby segmenting the sign language into morphemes and predicting morpheme positions, it is possible to lay the foundations for accurate sign language translation.
-
10.
公开(公告)号:US20240144955A1
公开(公告)日:2024-05-02
申请号:US18384105
申请日:2023-10-26
Applicant: Korea Electronics Technology Institute
Inventor: Min Young JUNG , Jin Yea JANG , San KIM , Sa Im SHIN
Abstract: There is a method for monitoring emotion and behavior occurring during a conversation with a user who is in need of protection. According to an embodiment, a user monitoring method receives an input of a user utterance and generates a conversation content by using a conversation model based on a user utterance content, converts the generated conversation content into a voice and outputs the voice, and recognizes emotion of the user by using an emotion analysis model based on the user utterance content and monitors the emotion. Accordingly, the method may monitor behavior/emotion of a user through a natural conversation while allowing a specialist/guardian to monitor an abnormal sign of a user who is in need of protection.
-
-
-
-
-
-
-
-
-