Invention Grant
US06366885B1 Speech driven lip synthesis using viseme based hidden markov models 有权
使用基于Viseme的隐马尔可夫模型的语音驱动唇形合成

Speech driven lip synthesis using viseme based hidden markov models
Abstract:
A method of speech driven lip synthesis which applies viseme based training models to units of visual speech. The audio data is grouped into a smaller number of visually distinct visemes rather than the larger number of phonemes. These visemes then form the basis for a Hidden Markov Model (HMM) state sequence or the output nodes of a neural network. During the training phase, audio and visual features are extracted from input speech, which is then aligned according to the apparent viseme sequence with the corresponding audio features being used to calculate the HMM state output probabilities or the output of the neutral network. During the synthesis phase, the acoustic input is aligned with the most likely viseme HMM sequence (in the case of an HMM based model) or with the nodes of the network (in the case of a neural network based system), which is then used for animation.
Information query
Patent Agency Ranking
0/0