DISTRIBUTED MACHINE LEARNING COMPILER OPTIMIZATION

    公开(公告)号:US20240177048A1

    公开(公告)日:2024-05-30

    申请号:US18056379

    申请日:2022-11-17

    CPC classification number: G06N20/00

    Abstract: A method for optimizing the compilation of a machine learning model to be executed on target edge devices is provided. Compute nodes of a plurality of compute nodes are allocated to a compiler optimization process for a compiler of said machine learning model. The machine learning model has a compute graph representation having nodes that are kernel operators necessary to execute the machine learning model and edges that connect said kernel operators to define precedence constraints. A round of optimization is scheduled for the process amongst the allocated compute nodes. At each allocated compute node a sequencing and scheduling solution is applied per round to obtain a performance metric for the machine learning model. From each compute node the performance metric is received and a solution that has the best performance metric is identified and implemented for execution of the machine learning model on the target edge devices.

    NEURAL DIRECTED ACYCLIC GRAPH (DAG) SCHEDULING VIA ONE-SHOT PRIORITY SAMPLING

    公开(公告)号:US20240119301A1

    公开(公告)日:2024-04-11

    申请号:US18464996

    申请日:2023-09-11

    CPC classification number: G06N3/092

    Abstract: A processor-implemented method includes sampling, according to a priority sampling policy, a set of node priorities from a computation graph. Each node priority of the set of node priorities may be associated with a respective node on the computation graph. Additionally, each node may represent an operation of a task performed by an artificial neural network. The method also includes converting, via a list scheduling function, the node priorities to a schedule that associates each node of the computation graph with a processor of a group of processors of a device associated with the artificial neural network, the schedule associated with a makespan. The method further includes performing the task in accordance with the schedule.

    GAIN SCALING OF INPUT TO NEURAL NETWORK FOR END-TO-END LEARNING IN WIRELESS COMMUNICATION SYSTEM

    公开(公告)号:US20230114870A1

    公开(公告)日:2023-04-13

    申请号:US17498651

    申请日:2021-10-11

    Abstract: A method of wireless communication by a user equipment (UE) includes receiving different sets of parameters from different sources as input to a receiver neural network. The method also includes receiving, from a base station, a set of target long-term energy values associated with the receiver neural network. The method further includes calculating a scaling factor for each of the different sets of parameters based on the set of target long-term energy values, and separately scaling each of the different sets of parameters based on the scaling factor calculated for that set in order to generate multiple sets of scaled parameters. The method still further includes transmitting the multiple sets of scaled parameters to the receiver neural network.

    QUALIFYING MACHINE LEARNING-BASED CSI PREDICTION

    公开(公告)号:US20210376895A1

    公开(公告)日:2021-12-02

    申请号:US16888593

    申请日:2020-05-29

    Abstract: Certain aspects of the present disclosure provide techniques for qualifying machine learning model-based channel state information (CSI) predictions. An example method generally includes receiving, from a network entity, a channel state information (CSI) prediction model for quantized CSI, calculating CSI based on downlink reference signal measurements, generating a quantized CSI difference value based a quantization of a difference between the calculated CSI and CSI predicted based on a CSI prediction model, and reporting, to the network entity, the calculated CSI and the quantized CSI difference value.

Patent Agency Ranking