-
公开(公告)号:US20190392625A1
公开(公告)日:2019-12-26
申请号:US16564546
申请日:2019-09-09
Inventor: Jianxiang Wang , Fuqiang Lyu , Xiao Liu , Jianchao Ji
Abstract: Embodiments of the present disclosure provide a method and apparatus for generating an animation. A method may include: extracting an audio feature from target speech segment by segment, to aggregate the audio feature into an audio feature sequence composed of an audio feature of each speech segment; inputting the audio feature sequence into a pre-trained mouth-shape information prediction model, to obtain a mouth-shape information sequence corresponding to the audio feature sequence; generating, for mouth-shape information in the mouth-shape information sequence, a face image including a mouth-shape object indicated by the mouth-shape information; and using the generated face image as a key frame of a facial animation, to generate the facial animation.
-
公开(公告)号:US20200234478A1
公开(公告)日:2020-07-23
申请号:US16668963
申请日:2019-10-30
Inventor: Xiao Liu , Fuqiang Lyu , Jianxiang Wang , Jianchao Ji
Abstract: Embodiments of the present disclosure provide a method and apparatus for processing information. A method may include: generating voice response information based on voice information sent by a user; generating a phoneme sequence based on the voice response information; generating mouth movement information based on the phoneme sequence, the mouth movement information being used for controlling a mouth movement of a displayed three-dimensional human image when playing the voice response information; and playing the voice response information, and controlling the mouth movement of the three-dimensional human image based on the mouth movement information.
-
公开(公告)号:US11158102B2
公开(公告)日:2021-10-26
申请号:US16668963
申请日:2019-10-30
Inventor: Xiao Liu , Fuqiang Lyu , Jianxiang Wang , Jianchao Ji
Abstract: Embodiments of the present disclosure provide a method and apparatus for processing information. A method may include: generating voice response information based on voice information sent by a user; generating a phoneme sequence based on the voice response information; generating mouth movement information based on the phoneme sequence, the mouth movement information being used for controlling a mouth movement of a displayed three-dimensional human image when playing the voice response information; and playing the voice response information, and controlling the mouth movement of the three-dimensional human image based on the mouth movement information.
-
公开(公告)号:US10997764B2
公开(公告)日:2021-05-04
申请号:US16564546
申请日:2019-09-09
Inventor: Jianxiang Wang , Fuqiang Lyu , Xiao Liu , Jianchao Ji
Abstract: Embodiments of the present disclosure provide a method and apparatus for generating an animation. A method may include: extracting an audio feature from target speech segment by segment, to aggregate the audio feature into an audio feature sequence composed of an audio feature of each speech segment; inputting the audio feature sequence into a pre-trained mouth-shape information prediction model, to obtain a mouth-shape information sequence corresponding to the audio feature sequence; generating, for mouth-shape information in the mouth-shape information sequence, a face image including a mouth-shape object indicated by the mouth-shape information; and using the generated face image as a key frame of a facial animation, to generate the facial animation.
-
公开(公告)号:US20200320303A1
公开(公告)日:2020-10-08
申请号:US16905482
申请日:2020-06-18
Inventor: Dongliang He , Xiang Zhao , Jizhou Huang , Fu Li , Xiao Liu , Shilei Wen
IPC: G06K9/00 , G06F16/783 , G06N3/08
Abstract: A method and an apparatus for grounding a target video clip in a video are provided. The method includes: determining a current video clip in the video based on a current position; acquiring descriptive information indicative of a pre-generated target video clip descriptive feature, and executing a target video clip determining step which includes: determining current state information of the current video clip, wherein the current state information includes information indicative of a feature of the current video clip; generating a current action policy based on the descriptive information and the current state information, the current action policy being indicative of a position change of the current video clip in the video; the method further comprises: in response to reaching a preset condition, using a video clip resulting from executing the current action policy on the current video clip as the target video clip.
-
公开(公告)号:US11410422B2
公开(公告)日:2022-08-09
申请号:US16905482
申请日:2020-06-18
Inventor: Dongliang He , Xiang Zhao , Jizhou Huang , Fu Li , Xiao Liu , Shilei Wen
Abstract: A method and an apparatus for grounding a target video clip in a video are provided. The method includes: determining a current video clip in the video based on a current position; acquiring descriptive information indicative of a pre-generated target video clip descriptive feature, and executing a target video clip determining step which includes: determining current state information of the current video clip, wherein the current state information includes information indicative of a feature of the current video clip; generating a current action policy based on the descriptive information and the current state information, the current action policy being indicative of a position change of the current video clip in the video; the method further comprises: in response to reaching a preset condition, using a video clip resulting from executing the current action policy on the current video clip as the target video clip.
-
7.
公开(公告)号:US10861133B1
公开(公告)日:2020-12-08
申请号:US16810986
申请日:2020-03-06
Inventor: Chao Li , Dongliang He , Xiao Liu , Yukang Ding , Shilei Wen , Errui Ding , Henan Zhang , Hao Sun
IPC: G06T3/40
Abstract: A super-resolution video reconstruction method, device, apparatus and a computer-readable storage medium are provided. The method includes: extracting a hypergraph from consecutive frames of an original video; inputting a hypergraph vector of the hypergraph into a residual convolutional neural network to obtain an output result of the residual convolutional neural network; and inputting the output result of the residual convolutional neural network into a spatial upsampling network to obtain a super-resolution frame, wherein a super-resolution video of the original video is formed by multiple super-resolution frames.
-
-
-
-
-
-