METHODS AND APPARATUS TO PERFORM PARALLEL DOUBLE-BATCHED SELF-DISTILLATION IN RESOURCE-CONSTRAINED IMAGE RECOGNITION APPLICATIONS

    公开(公告)号:US20240331371A1

    公开(公告)日:2024-10-03

    申请号:US18573973

    申请日:2021-11-30

    CPC classification number: G06V10/82

    Abstract: Methods and apparatus to perform parallel double-batched self-distillation in resource-constrained image recognition environments are disclosed herein. Example apparatus disclosed herein are to identify a source data batch and an augmented data batch, the augmented data generated based on at least one data augmentation technique. Disclosed example apparatus is also to share one or more parameters between a student neural network corresponding to the source data batch and a teacher neural network corresponding to the augmented data batch, the one or more parameters including one or more convolution layers to be shared between the teacher neural network and the student neural network. Disclosed example apparatus is further to align knowledge corresponding to the teacher neural network and the student neural network, the knowledge corresponding to the one or more parameters shared between the student neural network and the teacher neural network.

    DYNAMIC TEMPORAL NORMALIZATION FOR DEEP LEARNING IN VIDEO UNDERSTANDING APPLICATIONS

    公开(公告)号:US20240273873A1

    公开(公告)日:2024-08-15

    申请号:US18563305

    申请日:2021-09-01

    CPC classification number: G06V10/7715 G06V10/32 G06V10/82

    Abstract: Techniques related to application of deep neural networks to video for video recognition and understanding are discussed. A feature map of a deep neural network for a current time stamp of input video is standardized to a standardized feature map and pooled to a feature vector. The feature vector and transform parameters for a prior time stamp are used to generate transform parameters for the current time stamp based on application of a meta temporal relay. The resultant current time stamp transform parameters, such as a hidden state and a cell state of the meta temporal relay, are used to transform the standardized feature map to a normalized feature map for use by a subsequent layer of the deep neural network.

    Methods and systems for boosting deep neural networks for deep learning

    公开(公告)号:US11790223B2

    公开(公告)日:2023-10-17

    申请号:US16475076

    申请日:2017-04-07

    Abstract: Methods and systems are disclosed for boosting deep neural networks for deep learning. In one example, in a deep neural network including a first shallow network and a second shallow network, a first training sample is processed by the first shallow network using equal weights. A loss for the first shallow network is determined based on the processed training sample using equal weights. Weights for the second shallow network are adjusted based on the determined loss for the first shallow network. A second training sample is processed by the second shallow network using the adjusted weights. In another example, in a deep neural network including a first weak network and a second weak network, a first subset of training samples is processed by the first weak network using initialized weights. A classification error for the first weak network on the first subset of training samples is determined. The second weak network is boosted using the determined classification error of the first weak network with adjusted weights. A second subset of training samples is processed by the second weak network using the adjusted weights.

    METHODS AND APPARATUS TO DYNAMICALLY NORMALIZE DATA IN NEURAL NETWORKS

    公开(公告)号:US20230274132A1

    公开(公告)日:2023-08-31

    申请号:US18005804

    申请日:2020-08-26

    CPC classification number: G06N3/08

    Abstract: Methods, apparatus, systems, and articles of manufacture to dynamically normalize data in neural networks are disclosed. An apparatus for use with a machine learning model includes at least one normalization calculator to generate a plurality of alternate normalized outputs associated with input data for the machine learning model. Different ones of the alternate normalized outputs based on different normalization techniques. The apparatus further includes a soft weighting engine to generate a plurality of soft weights based on the input data. The apparatus also includes a normalized output generator to generate a final normalized output based on the plurality of alternate normalized outputs and the plurality of soft weights.

    METHOD AND APPARATUS FOR DYNAMIC NORMALIZATION AND RELAY IN A NEURAL NETWORK

    公开(公告)号:US20220207359A1

    公开(公告)日:2022-06-30

    申请号:US17485406

    申请日:2021-09-25

    Abstract: Embodiments are generally directed to methods and apparatuses for dynamic normalization and relay in a neural network. An embodiment of an apparatus for dynamic normalization and relay in a neural network including a hyper normalization layer comprises: a compute engine to: generate a hidden state and a cell state for the hyper normalization layer based on an input feature map for the hyper normalization layer as well as a previous hidden state and a previous cell state; and normalize the input feature map in the hyper normalization layer with the hidden state and the cell state for the hyper normalization layer.

    FEATURE FUSION FOR MULTI-MODAL MACHINE LEARNING ANALYSIS

    公开(公告)号:US20200279156A1

    公开(公告)日:2020-09-03

    申请号:US16645425

    申请日:2017-10-09

    Abstract: A system to perform multi-modal analysis has at least three distinct characteristics: an early abstraction layer for each data modality integrating homogeneous feature cues coming from different deep learning architectures for that data modality, a late abstraction layer for further integrating heterogeneous features extracted from different models or data modalities and output from the early abstraction layer, and a propagation-down strategy for joint network training in an end-to-end manner. The system is thus able to consider correlations among homogeneous features and correlations among heterogenous features at different levels of abstraction. The system further extracts and fuses discriminative information contained in these models and modalities for high performance emotion recognition.

    METHODS AND SYSTEMS USING IMPROVED TRAINING AND LEARNING FOR DEEP NEURAL NETWORKS

    公开(公告)号:US20200026988A1

    公开(公告)日:2020-01-23

    申请号:US16475075

    申请日:2017-04-07

    Abstract: Methods and systems are disclosed using improved training and learning for deep neural networks. In one example, a deep neural network includes a plurality of layers, and each layer has a plurality of nodes. For each L layer in the plurality of layers, the nodes of each L layer are randomly connected to nodes in a L+1 layer. For each L+1 layer in the plurality of layers, the nodes of each L+1 layer are connected to nodes in a subsequent L layer in a one-to-one manner. Parameters related to the nodes of each L layer are fixed. Parameters related to the nodes of each L+1 layers are updated, and L is an integer starting with 1. In another example, a deep neural network includes an input layer, output layer, and a plurality of hidden layers. Inputs for the input layer and labels for the output layer are determined related to a first sample. Similarity between different pairs of inputs and labels between a second sample with the first sample is estimated using Gaussian regression process.

Patent Agency Ranking