Abstract:
A method for generating a personalized artificial neural network (ANN) model receives an input at a first artificial neural network. The input is processed to extract a set of intermediate features. The method determines if the input is out-of-distribution relative to a dataset used to train the first artificial neural network. The intermediate features corresponding to the input are provided to a second artificial neural network bases on the out-of-distribution determination. Additionally, the system resources for performing the training and inference tasks of the first artificial neural network and the second, personalized artificial neural network are allocated according to a computational complexity of the training and inference tasks and a power consumption of the resources.
Abstract:
Various embodiments include methods and devices for generating source code of one or more trained machine learning models for use with an existing toolchain of an edge processing device. Embodiments may include parsing a trained machine learning model, generating weight data from the parsed trained machine learning model, generating layer code from the parsed trained machine learning model, and generating a network construct source code of the trained machine learning model from the weight data and the layer code in which the network construct source code is compileable for and executable by the edge processing device.
Abstract:
Aspects of the disclosure are related to a method, apparatus, and system for using display content from a rich operating system (OS) environment as a background image in a trusted user interface (UI), comprising: capturing a display buffer of the rich OS environment; passing the captured display buffer to a Trusted Application; and displaying, with the Trusted Application, the captured display buffer as the background image in the trusted UI, wherein the Trusted Application is executed in a Trusted Execution Environment (TEE).
Abstract:
A method for compressing of a deep neural network model includes determining an architecture of a teacher model. An initial layer of the teacher model is preserved and included in a student model. Repeated layers of the teacher model having a same type are identified. The second of such layers is removed such that the student model includes fewer layers than the teacher model. Knowledge distillation is applied to train the student model. The identifying and removal of layers of the same type is repeated to compress the student model.
Abstract:
Various embodiments include methods and devices for weight layout transformation of a weight tensor. Embodiments may include, accessing a first memory to retrieve weights of the weight tensor in a transformed order that is different than an order for retrieving the weights for a calculation at a network layer of a trained machine learning model, and loading the weights to a second memory in the transformed order. Embodiments may further include retrieving the weights from the second memory in the transformed order, and reordering the weights to the order for implementing the calculation at the network layer of the trained machine learning model.
Abstract:
The disclosure is related to searching for a second device to provide a service that a first device is attempting to establish. The first device sends a search profile and a capabilities profile to the second device using near field communication (NFC), the search profile including criteria describing the service the first device is attempting to establish, the capabilities profile including connection capabilities of the first device, receives a score from the second device, the score indicating a closeness of a match between the search profile and the capabilities profile and one or more services and capabilities of the second device, and determines whether to connect with the second device to establish the service based on the received score.
Abstract:
A method of anomaly detection and energy-efficient inference determination includes receiving an input. A set of features of the input are extracted using an artificial neural network (ANN) to generate a latent representation of the input. A reconstruction of the input is generated using the ANN, based on the latent representation. A reconstruction error is computed based on the generated reconstruction and the input. The reconstruction error is compared to a predefined threshold to determine whether the in-distribution data or out-of-distribution data. An anomaly is detected in response to an out-of-distribution determination. A decision model is provided with the latent representation in response to the input being determined to be in-distribution data. In turn, the decision model computes an inference based on the latent representation.
Abstract:
A method receives a first program code including one or more nested loops. A loop order is determined for the nested loop(s). The determined loop order aligns an input data layout and an output data layout. The nested loop(s) are transformed based on the loop order. A second program code is generated based on the transformed nested loop(s).
Abstract:
A method for energy-efficient classification receiving, via a first circuit, an input data stream from one or more sensors. The first circuit detects, while a second circuit is in a dormant state, if a state change has occurred between a first input of the input data stream and a second input of the input data stream. The second input is a next succeeding input of the input data stream. The first circuit triggers the second circuit to perform a classification of the input data stream in response to detecting the state change.
Abstract:
Certain aspects of the present disclosure are generally directed to apparatus and techniques for event state detection. One example method generally includes receiving a plurality of sensor signals at a computing device, determining, at the computing device, probabilities of sub-event states based on the plurality of sensor signals using an artificial neural network for each of a plurality of time intervals, and detecting, at the computing device, the event state based on the probabilities of the sub-event states via a state sequence model.