-
1.
公开(公告)号:US20230401446A1
公开(公告)日:2023-12-14
申请号:US18238016
申请日:2023-08-25
Applicant: HUAWEI TECHNOLOGIES CO., LTD.
Inventor: Yehui TANG , Yixing XU , Yunhe WANG , Chunjing XU
IPC: G06N3/082 , G06N3/0464
CPC classification number: G06N3/082 , G06N3/0464 , G06V10/82
Abstract: Embodiments of this application disclose a convolutional neural network pruning processing method, a data processing method, and a device, which may be applied to the field of artificial intelligence. The convolutional neural network pruning processing method includes: performing sparse training on a convolutional neural network by using a constructed objective loss function, where the objective loss function may include three sub-loss functions.
-
公开(公告)号:US20220180199A1
公开(公告)日:2022-06-09
申请号:US17680630
申请日:2022-02-25
Applicant: HUAWEI TECHNOLOGIES CO., LTD.
Inventor: Yixing XU , Hanting CHEN , Kai HAN , Yunhe WANG , Chunjing XU
Abstract: This application provides a neural network model compression method in the field of artificial intelligence. The method includes: obtaining, by a server, a first neural network model and training data of the first neural network that are uploaded by user equipment; obtaining a PU classifier based on the training data of the first neural network and unlabeled data stored in the server; selecting, by using the PU classifier, extended data from the unlabeled data stored in the server, where the extended data has a property and distribution similar to a property and distribution of the training data of the first neural network model; and training a second neural network model by using a knowledge distillation (KD) method based on the extended data, where the first neural network model is used as a teacher network model and the second neural network model is used as a student network model.
-
公开(公告)号:US20230153615A1
公开(公告)日:2023-05-18
申请号:US18147297
申请日:2022-12-28
Applicant: HUAWEI TECHNOLOGIES CO., LTD.
Inventor: Yixing XU , Xinghao CHEN , Yunhe WANG , Chunjing XU
Abstract: The technology of this application relates to a neural network distillation method, applied to the field of artificial intelligence, and includes processing to-be-processed data by using a first neural network and a second neural network to obtain a first target output and a second target output, where the first target output is obtained by performing kernel function-based transformation on an output of the first neural network layer, and the second target output is obtained by performing kernel function-based transformation on an output of the second neural network layer. The method further includes performing knowledge distillation on the first neural network based on a target loss constructed by using the first target output and the second target output.
-
公开(公告)号:US20210312261A1
公开(公告)日:2021-10-07
申请号:US17220158
申请日:2021-04-01
Applicant: HUAWEI TECHNOLOGIES CO., LTD.
Inventor: Yixing XU , Kai HAN , Yunhe WANG , Chunjing XU , Qi TIAN
Abstract: The present application discloses a neural network search method in the field of artificial intelligence, and the neural network search method includes: obtaining a feature tensor of each of a plurality of neural networks, where the feature tensor of each neural network is used to represent a computing capability of the neural network; inputting the feature tensor of each of the plurality of neural networks into an accuracy prediction model for calculation, to obtain accuracy of each neural network, where the accuracy prediction model is obtained through training based on a ranking-based loss function; and determining a neural network corresponding to the maximum accuracy as a target neural network. Embodiments of the present invention help improve accuracy of a network structure found through search.
-
-
-