METHODS AND APPARATUS TO OFFLOAD EXECUTION OF A PORTION OF A MACHINE LEARNING MODEL

    公开(公告)号:US20210397999A1

    公开(公告)日:2021-12-23

    申请号:US17359395

    申请日:2021-06-25

    Abstract: Methods, apparatus, systems and articles of manufacture to offload execution of a portion of a machine learning model are disclosed. An example apparatus includes processor circuitry to instantiate offload controller circuitry to select a first portion of layers of the machine learning model for execution at a first node and a second portion of the layers for remote execution for execution at a second node, model executor circuitry to execute the first portion of the layers, serialization circuitry to serialize the output of the execution of the first portion of the layers, and a network interface to transmit a request for execution of the machine learning model to the second node, the request including the serialized output of the execution of the first portion of the layers of the machine learning model and a layer identifier identifying the second portion of the layers of the machine learning model.

    Technologies for accelerating edge device workloads

    公开(公告)号:US11159454B2

    公开(公告)日:2021-10-26

    申请号:US16748232

    申请日:2020-01-21

    Abstract: Technologies for accelerating edge device workloads at a device edge network include a network computing device which includes a processor platform that includes at least one processor which supports a plurality of non-accelerated function-as-a-service (FaaS) operations and an accelerated platform that includes at least one accelerator which supports a plurality of accelerated FaaS (AFaaS) operation. The network computing device is configured to receive a request to perform a FaaS operation, determine whether the received request indicates that an AFaaS operation is to be performed on the received request, and identify compute requirements for the AFaaS operation to be performed. The network computing device is further configured to select an accelerator platform to perform the identified AFaaS operation and forward the received request to the selected accelerator platform to perform the identified AFaaS operation. Other embodiments are described and claimed.

    Technologies for quality of service based throttling in fabric architectures

    公开(公告)号:US10951516B2

    公开(公告)日:2021-03-16

    申请号:US16291541

    申请日:2019-03-04

    Abstract: Technologies for quality of service based throttling in a fabric architecture include a network node of a plurality of network nodes interconnected across the fabric architecture via an interconnect fabric. The network node includes a host fabric interface (HFI) configured to facilitate the transmission of data to/from the network node, monitor quality of service levels of resources of the network node used to process and transmit the data, and detect a throttling condition based on a result of the monitored quality of service levels. The HFI is further configured to generate and transmit a throttling message to one or more of the interconnected network nodes in response to having detected a throttling condition. The HFI is additionally configured to receive a throttling message from another of the network nodes and perform a throttling action on one or more of the resources based on the received throttling message. Other embodiments are described herein.

    Scalable edge computing
    129.
    发明授权

    公开(公告)号:US10944689B2

    公开(公告)日:2021-03-09

    申请号:US16024465

    申请日:2018-06-29

    Abstract: There is disclosed in one example a communication apparatus, including: a telemetry interface; a management interface; and an edge gateway configured to: identify diverted traffic, wherein the diverted traffic includes traffic to be serviced by an edge microcloud configured to provide a plurality of services; receive telemetry via the telemetry interface; use the telemetry to anticipate a future per-service demand within the edge microcloud; compute a scale for a resource to meet the future per-service demand; and operate the management interface to instruct the edge microcloud to perform the scale before the future per-service demand occurs.

Patent Agency Ranking