-
公开(公告)号:US10108606B2
公开(公告)日:2018-10-23
申请号:US15214215
申请日:2016-07-19
Inventor: Seung Yun , Ki Hyun Kim , Sang Hun Kim , Yun Young Kim , Jeong Se Kim , Min Kyu Lee , Soo Jong Lee , Young Jik Lee , Mu Yeol Choi
IPC: G06F17/28 , G10L13/033 , G10L13/06 , G10L25/24 , G10L25/75
Abstract: Provided are an automatic interpretation system and method for generating a synthetic sound having characteristics similar to those of an original speaker's voice. The automatic interpretation system for generating a synthetic sound having characteristics similar to those of an original speaker's voice includes a speech recognition module configured to generate text data by performing speech recognition for an original speech signal of an original speaker and extract at least one piece of characteristic information among pitch information, vocal intensity information, speech speed information, and vocal tract characteristic information of the original speech, an automatic translation module configured to generate a synthesis-target translation by translating the text data, and a speech synthesis module configured to generate a synthetic sound of the synthesis-target translation.
-
公开(公告)号:US10796106B2
公开(公告)日:2020-10-06
申请号:US16114388
申请日:2018-08-28
Inventor: Dong Hyun Kim , Young Jik Lee , Sang Hun Kim
IPC: G06F40/58 , G10L13/033 , G02B27/01 , G10L13/04
Abstract: Provided are an apparatus and method for selecting a speaker by using smart glasses. The apparatus includes a camera configured to capture a front angle video of a user and track guest interpretation interlocutors in the captured video, smart glasses configured to display a virtual space map image including the guest interpretation interlocutors tracked through the camera, a gaze-tracking camera configured to select a target person for interpretation by tracking a gaze of the user so that a guest interpretation interlocutor displayed in the video may be selected, and an interpretation target processor configured to provide an interpretation service in connection with the target person selected through the gaze-tracking camera.
-
公开(公告)号:US10249294B2
公开(公告)日:2019-04-02
申请号:US15646302
申请日:2017-07-11
Inventor: Dong Hyun Kim , Young Jik Lee , Sang Hun Kim , Seung Hi Kim , Min Kyu Lee , Mu Yeol Choi
IPC: G10L15/14 , G10L17/04 , G10L15/065 , G10L15/06 , G10L15/08
Abstract: A speech recognition method capable of automatic generation of phones according to the present invention includes: unsupervisedly learning a feature vector of speech data; generating a phone set by clustering acoustic features selected based on an unsupervised learning result; allocating a sequence of phones to the speech data on the basis of the generated phone set; and generating an acoustic model on the basis of the sequence of phones and the speech data to which the sequence of phones is allocated.
-
公开(公告)号:US20190188265A1
公开(公告)日:2019-06-20
申请号:US16114388
申请日:2018-08-28
Inventor: Dong Hyun Kim , Young Jik Lee , Sang Hun Kim
IPC: G06F17/28 , G10L13/04 , G10L13/033
Abstract: Provided are an apparatus and method for selecting a speaker by using smart glasses. The apparatus includes a camera configured to capture a front angle video of a user and track guest interpretation interlocutors in the captured video, smart glasses configured to display a virtual space map image including the guest interpretation interlocutors tracked through the camera, a gaze-tracking camera configured to select a target person for interpretation by tracking a gaze of the user so that a guest interpretation interlocutor displayed in the video may be selected, and an interpretation target processor configured to provide an interpretation service in connection with the target person selected through the gaze-tracking camera.
-
公开(公告)号:US20180075844A1
公开(公告)日:2018-03-15
申请号:US15646302
申请日:2017-07-11
Inventor: Dong Hyun KIM , Young Jik Lee , Sang Hun Kim , Seung Hi Kim , Min Kyu Lee , Mu Yeol Choi
IPC: G10L15/14 , G10L17/04 , G10L15/065
CPC classification number: G10L15/144 , G10L15/063 , G10L15/065 , G10L15/08 , G10L15/142 , G10L17/04 , G10L2015/0631
Abstract: A speech recognition method capable of automatic generation of phones according to the present invention includes: unsupervisedly learning a feature vector of speech data; generating a phone set by clustering acoustic features selected based on an unsupervised learning result; allocating a sequence of phones to the speech data on the basis of the generated phone set; and generating an acoustic model on the basis of the sequence of phones and the speech data to which the sequence of phones is allocated.
-
-
-
-