-
公开(公告)号:US12300012B2
公开(公告)日:2025-05-13
申请号:US17984034
申请日:2022-11-09
Inventor: Shangwen Lyu , Hongyu Li , Jing Liu , Hua Wu , Haifeng Wang
IPC: G06V30/19 , G06F40/205 , G06V30/194 , G06V30/412
Abstract: A method for training a document reading comprehension model includes: acquiring a question sample and a rich-text document sample, in which the rich-text document sample includes a real answer of the question sample; acquiring text information and layout information of the rich-text document sample by performing OCR processing on image information of the rich-text document sample; acquiring a predicted answer of the question sample by inputting the text information, the layout information and the image information of the rich-text document sample into a preset reading comprehension model; and training the reading comprehension model based on the real answer and the predicted answer. The method may enhance comprehension ability of the reading comprehension model to the long rich-text document, and save labor cost.
-
公开(公告)号:US12282746B2
公开(公告)日:2025-04-22
申请号:US17820765
申请日:2022-08-18
Inventor: Haifeng Wang , Zhanyi Liu , Zhongjun He , Hua Wu , Zhi Li , Xing Wan , Jingxuan Zhao , Ruiqing Zhang , Chuanqiang Zhang , Fengtao Huang , Hanbing Song , Wei Di , Shuangshuang Cui , Yongzheng Xin
IPC: G06F40/58
Abstract: A display method, an electronic device, and a storage medium, which relate to a field of natural language processing and a field of display. The display method includes: acquiring a content to be displayed; extracting a target term from the content using a term extraction rule; acquiring an annotation information for at least one target term, responsive to an extraction of the at least one target term; and displaying the annotation information for the at least one target term and the content.
-
公开(公告)号:US20230015313A1
公开(公告)日:2023-01-19
申请号:US17656160
申请日:2022-03-23
Inventor: Chuanqiang Zhang , Ruiqing Zhang , Zhongjun He , Zhi Li , Hua Wu
IPC: G06F40/58 , G06F40/279
Abstract: Disclosed are a translation method, a classification model training method, a device and a storage medium, which relate to the field of computer technologies, particularly to the field of artificial intelligence such as natural language processing and deep learning. The translation method includes: obtaining a current processing unit of a source language text based on a segmented word in the source language text; determining a classification result of the current processing unit with a classification model; and in response to determining that the classification result is the current processing unit being translatable separately, translating the current processing unit to obtain translation result in a target language corresponding to the current processing unit.
-
4.
公开(公告)号:US20220327809A1
公开(公告)日:2022-10-13
申请号:US17809133
申请日:2022-06-27
Inventor: Wei Li , Can Gao , Guocheng Niu , Xinyan Xiao , Hao Liu , Jiachen Liu , Hua Wu , Haifeng Wang
IPC: G06V10/778 , G06V10/774 , G06V10/26 , G06F40/284
Abstract: A method for training a model based on multi-modal data joint learning, includes: obtaining multi-modal data; in which the multi-modal data include at least one type of single-modal data and at least one type of Pair multi-modal data; inputting the single-modal data and the Pair multi-modal data into a decoupling attention Transformer network model to generate respectively Token semantic representation features and cross-modal semantic representation features; and training the decoupling attention Transformer network model based on the Token semantic representation features and the cross-modal semantic representation features.
-
公开(公告)号:US12277398B2
公开(公告)日:2025-04-15
申请号:US17694034
申请日:2022-03-14
Inventor: Jian Gong , Yu Sun , Hao Tian , Hua Wu , Haifeng Wang , Qiaoqiao She
IPC: G06F40/40 , G06F40/205 , G06F40/284
Abstract: A model training method, a model training platform, an electronic device and a storage medium are provided, which can be used in the field of artificial intelligence, particularly the fields of natural language processing and deep learning. The model training method includes: receiving an input; determining, based on the input, a user-oriented prefabricated function; determining, based on the input, a model training function; determining, based on the input, a pre-trained model; determining, based on the input, a network structure associated with the pre-trained model so as to support use of the pre-trained model; training, based on the input, the model by using the prefabricated function, the model training function, and the pre-trained model; and providing an output associated with a trained model.
-
公开(公告)号:US12277387B2
公开(公告)日:2025-04-15
申请号:US18056197
申请日:2022-11-16
Inventor: Ruiqing Zhang , Zhongjun He , Zhi Li , Hua Wu
IPC: G06F40/232 , G06F40/279 , G06F40/53
Abstract: A text processing method is provided. The method includes: a first probability value of each candidate character of a plurality of candidate characters corresponding to a target position is determined based on character feature information corresponding to the target position in a text fragment to be processed, wherein the character feature information is determined based on a context at the target position in the text fragment to be processed; a second probability value of each candidate character of the plurality of candidate characters is determined based on a character string including the candidate character and at least one character in at least one position in the text fragment to be processed adjacent to the target position; and a correction character at the target position is determined based on the first probability value and the second probability value of each candidate character of the plurality of candidate characters.
-
公开(公告)号:US20250094806A1
公开(公告)日:2025-03-20
申请号:US18967167
申请日:2024-12-03
Inventor: Junyuan Shang , Yilong Chen , Zhenyu Zhang , Shuohuan Wang , Yu Sun , Hua Wu
IPC: G06N3/082 , G06N3/0475
Abstract: Provided is a large language model training method, an electronic device and a storage medium, relating to the field of artificial intelligence technologies, and in particular, to the fields of deep learning, natural language processing and large model. The method includes: performing dimension reduction parameter fusion on a two-dimensional parameter matrix on each channel in each network layer in a first large language model, respectively, to obtain a second large language model; performing layer reduction parameter fusion on network layers in the second large language model based on a three-dimensional parameter matrix of each network layer in the second large language model to obtain a third large language model; and training the third large language model to obtain a target large language model under the condition that the target loss function determined based on the first and third large language models meets a preset first function condition.
-
公开(公告)号:US12118319B2
公开(公告)日:2024-10-15
申请号:US17655772
申请日:2022-03-21
Inventor: Jun Xu , Zeming Liu , Zeyang Lei , Zhengyu Niu , Hua Wu , Haifeng Wang
Abstract: The present disclosure provides a dialog method and system, an electronic device and a storage medium, and relates to the field of artificial intelligence (AI) technologies such as deep learning and natural language processing. A specific implementation scheme involves: rewriting a corresponding dialog state based on received dialog information of a user; determining to-be-used dialog action information based on the dialog information of the user and the dialog state; and generating a reply statement based on the dialog information of the user and the dialog action information. According to the present disclosure, the to-be-used dialog action information can be determined based on the dialog information of the user and the dialog state; and then the reply statement is generated based on the dialog action information, thereby providing an efficient dialog scheme.
-
公开(公告)号:US12086555B2
公开(公告)日:2024-09-10
申请号:US17643053
申请日:2021-12-07
Inventor: Jianglu Hu , Hehan Li , Huifeng Sun , Shuqi Sun , Yue Chang , Tingting Li , Hua Wu , Haifeng Wang
IPC: G06F40/35 , G06F16/332
CPC classification number: G06F40/35 , G06F16/3329
Abstract: The disclosure provides a method for generating a dialogue. The method includes: obtaining an input sentence; determining a type of a task-based response sentence that is to be generated, by updating a current dialogue state based on the input sentence; generating the task-based response sentence by inputting the input sentence into a task-based dialogue response generator; and determining the task-based response sentence as a target response sentence in response to the type of the task-based response sentence being a designated type.
-
公开(公告)号:US20230029687A1
公开(公告)日:2023-02-02
申请号:US17655772
申请日:2022-03-21
Inventor: Jun Xu , Zeming Liu , Zeyang Lei , Zhengyu Niu , Hua Wu , Haifeng Wang
Abstract: The present disclosure provides a dialog method and system, an electronic device and a storage medium, and relates to the field of artificial intelligence (AI) technologies such as deep learning and natural language processing. A specific implementation scheme involves: rewriting a corresponding dialog state based on received dialog information of a user; determining to-be-used dialog action information based on the dialog information of the user and the dialog state; and generating a reply statement based on the dialog information of the user and the dialog action information. According to the present disclosure, the to-be-used dialog action information can be determined based on the dialog information of the user and the dialog state; and then the reply statement is generated based on the dialog action information, thereby providing an efficient dialog scheme.
-
-
-
-
-
-
-
-
-