COMPILING CODE FOR A MACHINE LEARNING MODEL FOR EXECUTION ON A SPECIALIZED PROCESSOR

    公开(公告)号:WO2020242686A1

    公开(公告)日:2020-12-03

    申请号:PCT/US2020/029937

    申请日:2020-04-24

    Applicant: APPLE INC.

    Abstract: The subject technology receives a neural network model in a model format, the model format including information for a set of layers of the neural network model, each layer of the set of layers including a set of respective operations. The subject technology generates neural network (NN) code from the neural network model, the NN code being in a programming language distinct from the model format, and the NN code comprising a respective memory allocation for each respective layer of the set of layers of the neural network model, where the generating comprises determining the respective memory allocation for each respective layer based at least in part on a resource constraint of a target device. The subject technology compiles the NN code into a binary format. The subject technology generates a package for deploying the compiled NN code on the target device.

    DYNAMIC TASK ALLOCATION FOR NEURAL NETWORKS
    3.
    发明申请

    公开(公告)号:WO2018222299A1

    公开(公告)日:2018-12-06

    申请号:PCT/US2018/029201

    申请日:2018-04-24

    Applicant: APPLE INC.

    Abstract: The subject technology provides for dynamic task allocation for neural network models. The subject technology determines an operation performed at a node of a neural network model. The subject technology assigns an annotation to indicate whether the operation is better performed on a CPU or a GPU based at least in part on hardware capabilities of a target platform. The subject technology determines whether the neural network model includes a second layer. The subject technology, in response to determining that the neural network model includes a second layer, for each node of the second layer of the neural network model, determines a second operation performed at the node. Further the subject technology assigns a second annotation to indicate whether the second operation is better performed on the CPU or the GPU based at least in part on the hardware capabilities of the target platform.

    DYNAMIC TASK ALLOCATION FOR NEURAL NETWORKS
    4.
    发明公开

    公开(公告)号:EP4283526A2

    公开(公告)日:2023-11-29

    申请号:EP23202256.6

    申请日:2018-04-24

    Applicant: Apple Inc.

    Abstract: The subject technology provides for dynamic task allocation for neural network models. The subject technology determines an operation performed at a node of a neural network model. The subject technology assigns an annotation to indicate whether the operation is better performed on a CPU or a GPU based at least in part on hardware capabilities of a target platform. The subject technology determines whether the neural network model includes a second layer. The subject technology, in response to determining that the neural network model includes a second layer, for each node of the second layer of the neural network model, determines a second operation performed at the node. Further the subject technology assigns a second annotation to indicate whether the second operation is better performed on the CPU or the GPU based at least in part on the hardware capabilities of the target platform.

    DYNAMIC TASK ALLOCATION FOR NEURAL NETWORKS
    5.
    发明公开

    公开(公告)号:EP4283526A3

    公开(公告)日:2024-01-17

    申请号:EP23202256.6

    申请日:2018-04-24

    Applicant: Apple Inc.

    Abstract: The subject technology provides for dynamic task allocation for neural network models. The subject technology determines an operation performed at a node of a neural network model. The subject technology assigns an annotation to indicate whether the operation is better performed on a CPU or a GPU based at least in part on hardware capabilities of a target platform. The subject technology determines whether the neural network model includes a second layer. The subject technology, in response to determining that the neural network model includes a second layer, for each node of the second layer of the neural network model, determines a second operation performed at the node. Further the subject technology assigns a second annotation to indicate whether the second operation is better performed on the CPU or the GPU based at least in part on the hardware capabilities of the target platform.

Patent Agency Ranking