Abstract:
An apparatus for improving context-based automatic interpretation performance includes: an uttered voice input unit configured to receive a voice signal from a user; a previous sentence input unit configured to determine whether there is a user’s previous utterance when the voice signal is input by the uttered voice input unit; a voice encoding processing unit configured to decode only the voice signal through the uttered voice input unit when it is determined that there is no user’s previous utterance and extract a vector of the voice signal when it is determined that there is the user’s previous utterance; a context encoding processing unit configured to extract a context vector from a previous utterance when there is the previous utterance and transmit the extracted context vector of the previous utterance; and an interpretation decoding processing unit configured to output an interpretation result text.
Abstract:
Provided is a method for interpretation and translation accomplished by an interpretation and translation apparatus of a user through interfacing with an interpretation and translation apparatus of the other party. The method includes: automatically setting a translation target language which enables to communicate with the other party based on a message from the interpretation and translation apparatus of the other party by using a communication connection in a network; receiving input information of a use language of the user; calling a translator corresponding to the translation target language to transmit a result obtained by translating the input information into the translation target language to the interpretation and translation apparatus of the other party; and outputting received data from the interpretation and translation apparatus of the other party or outputting the result obtained by translating the received data into the use language of the user by using the translator.
Abstract:
The present disclosure relates to a method and device for improving the performance of an AI model that uses voice recognition results as text input. A method of training an AI model according to an embodiment of the present disclosure may include: generating first time information on a plurality of words included in a voice and transcription, using a first learning sample including the voice and the transcription; generating second time information by adding a pre-configured delay time to the first time information; generating a modified transcription based on an end time of a last word among the plurality of words and the second time information; and performing training of the AI model based on a second training sample including the voice and the modified transcription.
Abstract:
A method for generating multiple phoneme for foreign proper nouns according to the present invention comprises: converting a second language proper noun uttered in a first language to a second language word using an automatic translator; generating second language phoneme strings corresponding to the second language word using a second language G2P; converting the second language phoneme strings to first language phoneme strings; generating first language phoneme strings corresponding to the second language proper noun uttered in the first language using a first language G2P; and generating a plurality of phoneme strings by using the first language phoneme strings obtained through the step of converting to the first language phoneme strings and the first language phoneme strings obtained through the step of generating the first language phoneme strings.
Abstract:
The present invention relates to an automatic interpretation method and system for converting only voice of a speaker into a target language. The present invention may significantly improve performance of automatic interpretation with a foreigner to be communicated with, even in a high-noise environment in which multiple speakers utter at the same time by utilizing voice and image information input to a smart device in a complex manner. In addition, the present invention may determine a situation based on text information and image information existing around a user, and reflect the situation information together with multimodal information to an interpretation engine in real time. In addition, the present invention may significantly improve user convenience of an automatic interpretation system by directly augmenting and displaying an interpreted sentence directly next to a speaker image or generating a synthesized sound by distinguishing the interpreted sentence from other speeches.
Abstract:
Provided a method performed by an automatic interpretation server based on a zero user interface (UI), which communicates with a plurality of terminal devices having a microphone function, a speaker function, a communication function, and a wearable function. The method includes connecting terminal devices disposed within a designated automatic interpretation zone, receiving a voice signal of a first user from a first terminal device among the terminal devices within the automatic interpretation zone, matching a plurality of users placed within a speech-receivable distance of the first terminal device, and performing automatic interpretation on the voice signal and transmitting results of the automatic interpretation to a second terminal device of at least one second user corresponding to a result of the matching.
Abstract:
Provided are an apparatus and method for providing a personal assistant service based on automatic translation. The apparatus for providing a personal assistant service based on automatic translation includes an input section configured to receive a command of a user, a memory in which a program for providing a personal assistant service according to the command of the user is stored, and a processor configured to execute the program. The processor updates at least one of a speech recognition model, an automatic interpretation model, and an automatic translation model on the basis of an intention of the command of the user using a recognition result of the command of the user and provides the personal assistant service on the basis of an automatic translation call.
Abstract:
An automatic translation device includes a communications module transmitting and receiving data to and from an ear-set device including a speaker, a first microphone, and a second microphone, a memory storing a program generating a result of translation using a dual-channel audio signal, and a processor executing the program stored in the memory. When the program is executed, the processor compares a first audio signal including a voice signal of a user, received using the first microphone, with a second audio signal including a noise signal and the voice signal of the user, received using the second microphone, and entirely or selectively extracting the voice signal of the user from the first and second audio signals, based on a result of the comparison, to perform automatic translation.
Abstract:
A method of correcting errors in a speech recognition system includes a process of searching a speech recognition error-answer pair DB based on a sound model for a first candidate answer group for a speech recognition error, a process of searching a word relationship information DB for a second candidate answer group for the speech recognition error, a process of searching a user error correction information DB for a third candidate answer group for the speech recognition error, a process of searching a domain articulation pattern DB and a proper noun DB for a fourth candidate answer group for the speech recognition error, and a process of aligning candidate answers within each of the retrieved candidate answer groups and displaying the aligned candidate answers.
Abstract:
Provided is a method of performing automatic interpretation based on speaker separation by a user terminal, the method including: receiving a first speech signal including at least one of a user speech of a user and a user surrounding speech around the user from an automatic interpretation service providing terminal, separating the first speech signal into speaker-specific speech signals, performing interpretation on the speaker-specific speech signals in a language selected by the user on the basis of an interpretation mode, and providing a second speech signal generated as a result of the interpretation to at least one of a counterpart terminal and the automatic interpretation service providing terminal according to the interpretation mode.