-
公开(公告)号:US20210287044A1
公开(公告)日:2021-09-16
申请号:US17104165
申请日:2020-11-25
Inventor: Long LI , Haifeng WANG , Weibao GONG
Abstract: A method for updating a parameter of a model, a distributed training system, and an electric device are related to a field of deep learning technologies. The method includes: obtaining a batch training period of batch training data to be trained for a model; increasing priorities of tasks ranked at the bottom in a sequence of gradient communication tasks for parameters of the model when the batch training period is greater than or equal to a preset period threshold; and performing a communication of gradients of the parameters and updating the parameters based on priorities of the gradient communication tasks for the parameters in the model.
-
公开(公告)号:US20210406767A1
公开(公告)日:2021-12-30
申请号:US17142822
申请日:2021-01-06
Inventor: Daxiang DONG , Weibao GONG , Yi LIU , Dianhai YU , Yanjun MA , Haifeng WANG
IPC: G06N20/00 , G06F16/182 , G06N5/04
Abstract: The present application discloses a distributed training method and system, a device and a storage medium, and relates to technical fields of deep learning and cloud computing. The method includes: sending, by a task information server, a first training request and information of an available first computing server to at least a first data server; sending, by the first data server, a first batch of training data to the first computing server, according to the first training request; performing, by the first computing server, model training according to the first batch of training data, sending model parameters to the first data server so as to be stored after the training is completed, and sending identification information of the first batch of training data to the task information server so as to be recorded; wherein the model parameters are not stored at any one of the computing servers.
-