Abstract:
A parallel data processing method based on multiple graphic processing units (GPUs) is provided, including: creating, in a central processing unit (CPU), a plurality of worker threads for controlling a plurality of worker groups respectively, the worker groups including one or more GPUs; binding each worker thread to a corresponding GPU; loading a plurality of batches of training data from a nonvolatile memory to GPU video memories in the plurality of worker groups; and controlling the plurality of GPUs to perform data processing in parallel through the worker threads. The method can enhance efficiency of multi-GPU parallel data processing. In addition, a parallel data processing apparatus is further provided.
Abstract:
A parallel data processing method based on multiple graphic processing units (GPUs) is provided, including: creating, in a central processing unit (CPU), a plurality of worker threads for controlling a plurality of worker groups respectively, the worker groups including a plurality of GPUs; binding each worker thread to a corresponding GPU; loading one batch of training data from a nonvolatile memory to a GPU video memory corresponding to one worker group; transmitting, between a plurality of GPUs corresponding to one worker group, data required by data processing performed by the GPUs through peer to peer; and controlling the plurality of GPUs to perform data processing in parallel through the worker threads.
Abstract:
A data processing method is described. A processor provides different training data sets to a plurality of graphics processing units (GPUs), respectively. The processor controls the plurality of GPUs to generate respective sets of modification parameters by performing respective training processes in a parallel manner using respectively managed versions of a predictive model according to the corresponding training data sets. The processor controls the plurality of GPUs to exchange, before completion of the respective training processes, at least a portion of the sets of modification parameters that have been generated by the plurality of GPUs. The processor also causes the plurality of GPUs to modify the respectively managed versions of the predictive model according to at least the portion of the sets of modification parameters exchanged among the plurality of GPUs.
Abstract:
A parallel data processing method based on multiple graphic processing units (GPUs) is provided, including: creating, in a central processing unit (CPU), a plurality of worker threads for controlling a plurality of worker groups respectively, the worker groups including a plurality of GPUs; binding each worker thread to a corresponding GPU; loading one batch of training data from a nonvolatile memory to a GPU video memory corresponding to one worker group; transmitting, between a plurality of GPUs corresponding to one worker group, data required by data processing performed by the GPUs through peer to peer; and controlling the plurality of GPUs to perform data processing in parallel through the worker threads.
Abstract:
A data processing method in a data processing device is provided. First to-be-processed data input into a neural network are obtained. Iterative training is performed on the neural network for a first preset number of times by using first target data in the first to-be-processed data, to obtain a seed model of the neural network. First newly added data generated after an elapse of time corresponding to the first time window is obtained, and the first newly added data and the first to-be-processed data are combined into second to-be-processed data. Iterative training is performed on the seed model for a second preset number of times by using second target data in the second to-be-processed data, to obtain a first incremental model of the neural network. A first preset area overlaps between the second time window and the first time window. The first incremental model online is published.
Abstract:
A parallel data processing method based on multiple graphic processing units (GPUs) is provided, including: creating, in a central processing unit (CPU), a plurality of worker threads for controlling a plurality of worker groups respectively, the worker groups including one or more GPUs; binding each worker thread to a corresponding GPU; loading a plurality of batches of training data from a nonvolatile memory to GPU video memories in the plurality of worker groups; and controlling the plurality of GPUs to perform data processing in parallel through the worker threads. The method can enhance efficiency of multi-GPU parallel data processing. In addition, a parallel data processing apparatus is further provided.
Abstract:
Embodiments of this disclosure disclose a virtual item control method and apparatus, a terminal, and a non-transitory storage medium, and belong to the field of computer technologies. The method includes: displaying a plurality of item icons after a press operation on a throw button is detected; controlling, in response to the press operation moving to the target item icon and a duration reaches a target duration, a virtual object to hold a target virtual item corresponding to the target item icon; controlling, in response to a movement of the press operation, the virtual object to aim at a target position; and controlling, in response to a release of the press operation, the virtual object to throw the target virtual item to the target position.