WEIGHT SKIPPING DEEP LEARNING ACCELERATOR
    3.
    发明申请

    公开(公告)号:US20190303757A1

    公开(公告)日:2019-10-03

    申请号:US16221295

    申请日:2018-12-14

    Applicant: MediaTek Inc.

    Abstract: A deep learning accelerator (DLA) includes processing elements (PEs) grouped into PE groups to perform convolutional neural network (CNN) computations, by applying multi-dimensional weights on an input activation to produce an output activation. The DLA also includes a dispatcher which dispatches input data in the input activation and non-zero weights in the multi-dimensional weights to the processing elements according to a control mask. The DLA also includes a buffer memory which stores the control mask which specifies positions of zero weights in the multi-dimensional weights. The PE groups generate output data of respective output channels in the output activation, and share a same control mask specifying same positions of the zero weights.

    NEURAL NETWORK ENGINE WITH TILE-BASED EXECUTION

    公开(公告)号:US20190220742A1

    公开(公告)日:2019-07-18

    申请号:US16246884

    申请日:2019-01-14

    Applicant: MediaTek Inc.

    CPC classification number: G06N3/08

    Abstract: An accelerator for neural network computing includes hardware engines and a buffer memory. The hardware engines include a convolution engine and at least a second engine. Each hardware engine includes circuitry to perform neural network operations. The buffer memory stores a first input tile and a second input tile of an input feature map. The second input tile overlaps with the first input tile in the buffer memory. The convolution engine is operative to retrieve the first input tile from the buffer memory, perform convolution operations on the first input tile to generate an intermediate tile of an intermediate feature map, and pass the intermediate tile to the second engine via the buffer memory.

Patent Agency Ranking