MACHINE TRANSLATION MODEL TRAINING METHOD, APPARATUS, ELECTRONIC DEVICE AND STORAGE MEDIUM

    公开(公告)号:US20210200963A1

    公开(公告)日:2021-07-01

    申请号:US17200588

    申请日:2021-03-12

    Abstract: The present disclosure provides a machine translation model training method, apparatus, electronic device and storage medium, which relates to the technical field of natural language processing. A specific implementation solution is as follows: selecting, from parallel corpuses, a set of samples whose translation quality satisfies a preset requirement and which have universal-field features and/or target-field features, to constitute a first training sample set; selecting, from the parallel corpuses, a set of samples whose translation quality satisfies a preset requirement and which do not have universal-field features and target-field features, to constitute a second training sample set; training an encoder in the machine translation model in the target field, a discriminator configured in encoding layers of the encoder, and the encoder and a decoder in the machine translation model in the target field in turn with the first training sample set and second training sample set, respectively. The training method according to the present disclosure is time-saving and effort-saving, and may effectively improve the training efficiency of the machine translation model in the target field.

    METHOD AND APPARATUS FOR TRAINING MODELS IN MACHINE TRANSLATION, ELECTRONIC DEVICE AND STORAGE MEDIUM

    公开(公告)号:US20210390266A1

    公开(公告)日:2021-12-16

    申请号:US17200551

    申请日:2021-03-12

    Abstract: A method and apparatus for training models in machine translation, an electronic device and a storage medium are disclosed, which relates to the field of natural language processing technologies and the field of deep learning technologies. An implementation includes mining similar target sentences of a group of samples based on a parallel corpus using a machine translation model and a semantic similarity model, and creating a first training sample set; training the machine translation model with the first training sample set; mining a negative sample of each sample in the group of samples based on the parallel corpus using the machine translation model and the semantic similarity model, and creating a second training sample set; and training the semantic similarity model with the second sample training set. With the above-mentioned technical solution of the present application, by training the two models jointly, while the semantic similarity model is trained, the machine translation model may be optimized and nurtures the semantic similarity model, thus further improving the accuracy of the semantic similarity model.

Patent Agency Ranking