Hybrid aggregation for deep learning neural networks

    公开(公告)号:US10783437B2

    公开(公告)日:2020-09-22

    申请号:US15450010

    申请日:2017-03-05

    摘要: A processing unit topology of a neural network including a plurality of processing units is determined. The neural network includes at least one machine in which each machine includes a plurality of nodes, and wherein each node includes at least one of the plurality of processing units. One or more of the processing units are grouped into a first group according to a first affinity. The first group is configured, using a processor and a memory, to use a first aggregation procedure for exchanging model parameters of a model of the neural network between the processing units of the first group. One or more of the processing units are grouped into a second group according to a second affinity. The second group is configured to use a second aggregation procedure for exchanging the model parameters between the processing units of the second group.

    Method for visualizing neural network models

    公开(公告)号:US10936938B2

    公开(公告)日:2021-03-02

    申请号:US15857587

    申请日:2017-12-28

    摘要: A method for providing a graphical visualization of a neural network to a user is provided. The method includes generating the graphical visualization of the neural network at least in part by: representing layers of the neural network as respective three-dimensional blocks, wherein at least a first dimension of a given block is proportional to a computational complexity of a layer of the neural network represented by the given block; and representing data flows between the layers of the neural network as respective three-dimensional structures connecting blocks representing the layers of the neural network, wherein a first dimension of a given structure is proportional to each of a first dimension and a second dimension of a data flow represented by the given structure. The method also includes displaying the graphical visualization of the neural network to the user.

    PARAMETER DATA SHARING FOR MULTI-LEARNER TRAINING OF MACHINE LEARNING APPLICATIONS

    公开(公告)号:US20240020582A1

    公开(公告)日:2024-01-18

    申请号:US18355058

    申请日:2023-07-19

    IPC分类号: G06N20/20 G06N20/00

    CPC分类号: G06N20/20 G06N20/00 H04L67/10

    摘要: A machine receives a first set of global parameters from a global parameter server. Multiple learner processors in the machine execute an algorithm that models an entity type using the first set of global parameters and a mini-batch of data known to describe the entity type. The machine generates a consolidated set of gradients that describes a direction for the first set of global parameters in order to improve an accuracy of the algorithm in modeling the entity type when using the first set of global parameters and the mini-batch of data. The machine transmits the consolidated set of gradients to the global parameter server. The machine then receives a second set of global parameters from the global parameter server, where the second set of global parameters is a modification of the first set of global parameters based on the consolidated set of gradients.

    STORAGE CONTROLLER ACCELARATION FOR NEURAL NETWORK TRAINING AND INFERENCE

    公开(公告)号:US20180322383A1

    公开(公告)日:2018-11-08

    申请号:US15584136

    申请日:2017-05-02

    IPC分类号: G06N3/08 G06N3/04 G06N3/063

    CPC分类号: G06N3/08 G06N3/04 G06N3/063

    摘要: A storage controller of a machine receives training data associated with a neural network model. The neural network model includes a plurality of layers, and the machine further including at least one graphics processing unit. The storage controller trains at least one layer of the plurality of layers of the neural network model using the training data to generate processed training data. A size of the processed data is less than a size of the training data. Training of the at least one layer includes adjusting one or more weights of the at least one layer using the training data. The storage controller sends the processed training data to at least one graphics processing unit of the machine. The at least one graphics processing unit is configured to store the processed training data and train one or more remaining layers of the plurality of layers using the processed training data.

    HYBRID AGGREGATION FOR DEEP LEARNING NEURAL NETWORKS

    公开(公告)号:US20180253646A1

    公开(公告)日:2018-09-06

    申请号:US15450010

    申请日:2017-03-05

    IPC分类号: G06N3/08

    CPC分类号: G06N3/084 G06N3/0454

    摘要: A processing unit topology of a neural network including a plurality of processing units is determined. The neural network includes at least one machine in which each machine includes a plurality of nodes, and wherein each node includes at least one of the plurality of processing units. One or more of the processing units are grouped into a first group according to a first affinity. The first group is configured, using a processor and a memory, to use a first aggregation procedure for exchanging model parameters of a model of the neural network between the processing units of the first group. One or more of the processing units are grouped into a second group according to a second affinity. The second group is configured to use a second aggregation procedure for exchanging the model parameters between the processing units of the second group.

    Parameter data sharing for multi-learner training of machine learning applications

    公开(公告)号:US11748666B2

    公开(公告)日:2023-09-05

    申请号:US15347875

    申请日:2016-11-10

    IPC分类号: G06N20/20 G06N20/00 H04L67/10

    CPC分类号: G06N20/20 G06N20/00 H04L67/10

    摘要: A machine receives a first set of global parameters from a global parameter server. The first set of global parameters includes data that weights one or more operands used in an algorithm that models an entity type. Multiple learner processors in the machine execute the algorithm using the first set of global parameters and a mini-batch of data known to describe the entity type. The machine generates a consolidated set of gradients that describes a direction for the first set of global parameters in order to improve an accuracy of the algorithm in modeling the entity type when using the first set of global parameters and the mini-batch of data. The machine transmits the consolidated set of gradients to the global parameter server. The machine then receives a second set of global parameters from the global parameter server, where the second set of global parameters is a modification of the first set of global parameters based on the consolidated set of gradients.

    Storage controller acceleration for neural network training and inference

    公开(公告)号:US11138494B2

    公开(公告)日:2021-10-05

    申请号:US15584136

    申请日:2017-05-02

    IPC分类号: G06N3/063 G06N3/04 G06N3/08

    摘要: A storage controller of a machine receives training data associated with a neural network model. The neural network model includes a plurality of layers, and the machine further including at least one graphics processing unit. The storage controller trains at least one layer of the plurality of layers of the neural network model using the training data to generate processed training data. A size of the processed data is less than a size of the training data. Training of the at least one layer includes adjusting one or more weights of the at least one layer using the training data. The storage controller sends the processed training data to at least one graphics processing unit of the machine. The at least one graphics processing unit is configured to store the processed training data and train one or more remaining layers of the plurality of layers using the processed training data.

    METHOD FOR VISUALIZING NEURAL NETWORK MODELS

    公开(公告)号:US20190205728A1

    公开(公告)日:2019-07-04

    申请号:US15857587

    申请日:2017-12-28

    IPC分类号: G06N3/04 G06F17/30

    CPC分类号: G06N3/04 G06F16/904

    摘要: A method for providing a graphical visualization of a neural network to a user is provided. The method includes generating the graphical visualization of the neural network at least in part by: representing layers of the neural network as respective three-dimensional blocks, wherein at least a first dimension of a given block is proportional to a computational complexity of a layer of the neural network represented by the given block; and representing data flows between the layers of the neural network as respective three-dimensional structures connecting blocks representing the layers of the neural network, wherein a first dimension of a given structure is proportional to each of a first dimension and a second dimension of a data flow represented by the given structure. The method also includes displaying the graphical visualization of the neural network to the user.