-
公开(公告)号:US10686688B2
公开(公告)日:2020-06-16
申请号:US15655846
申请日:2017-07-20
Applicant: INTEL CORPORATION
Inventor: Francesc Guim Bernat , Susanne M. Balle , Daniel Rivas Barragan , John Chun Kwok Leung , Suraj Prabhakaran , Murugasamy K. Nachimuthu , Slawomir Putyrski
IPC: H04L12/26 , G06F16/22 , G06F16/23 , H04L12/24 , H04L12/927 , H04Q9/00 , H04L29/08 , H04L12/925
Abstract: Techniques for reducing fragmentation in software-defined infrastructures are described. A compute node, including one or more processor circuits, may be configured to access one or more remote resources via a fabric, the compute node may be configured to receive a dynamic tolerated fragmentation for the one or more remote resources. The compute node may be configured to monitor the performance of the one or more remote resources. For example, the compute node may be configured to monitor if one or more of the monitored resources were to exceed a threshold bandwidth or latency range as defined by the dynamic tolerated fragmentation. The compute node may be configured to determine that the monitored performance of the one or more remote resources is outside a threshold defined by the dynamic tolerated fragmentation. If one or more of the remote resources is outside the threshold, for a predetermined period of time, or otherwise, the compute node may be configured to determine so and take appropriate measures, such as generating a message indicating that performance of the one or more remote resources is outside a threshold defined by the dynamic tolerated fragmentation. Other embodiments are described and claimed.
-
公开(公告)号:US20190044886A1
公开(公告)日:2019-02-07
申请号:US15941943
申请日:2018-03-30
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Anil Rao , Suraj Prabhakaran , Mohan Kumar , Karthik Kumar
IPC: H04L12/947 , H04L12/931 , H04L12/801 , H04L12/66
Abstract: Technologies for accelerating edge device workloads at a device edge network include a network computing device which includes a processor platform that includes at least one processor which supports a plurality of non-accelerated function-as-a-service (FaaS) operations and an accelerated platform that includes at least one accelerator which supports a plurality of accelerated FaaS (AFaaS) operation. The network computing device is configured to receive a request to perform a FaaS operation, determine whether the received request indicates that an AFaaS operation is to be performed on the received request, and identify compute requirements for the AFaaS operation to be performed. The network computing device is further configured to select an accelerator platform to perform the identified AFaaS operation and forward the received request to the selected accelerator platform to perform the identified AFaaS operation. Other embodiments are described and claimed.
-
公开(公告)号:US20190034340A1
公开(公告)日:2019-01-31
申请号:US15855104
申请日:2017-12-27
Applicant: Intel Corporation
Inventor: Kshitij A. Doshi , Francesc Guim Bernat , Daniel Rivas Barragan , Suraj Prabhakaran
IPC: G06F12/0831 , G06F17/30 , G06F13/16
Abstract: An embodiment of a semiconductor package apparatus may include technology to create a tracking structure for a memory controller to track a range of memory addresses of a persistent memory, identify a write request at the memory controller for a memory location within the range of tracked memory addresses, and set a flag in the tracking structure to indicate that the memory location had the identified write request. Other embodiments are disclosed and claimed.
-
公开(公告)号:US20250097306A1
公开(公告)日:2025-03-20
申请号:US18894452
申请日:2024-09-24
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Patrick Bohan , Kshitij Arun Doshi , Brinda Ganesh , Andrew J. Herdrich , Monica Kenguva , Karthik Kumar , Patrick G. Kutch , Felipe Pastor Beneyto , Rashmin Patel , Suraj Prabhakaran , Ned M. Smith , Petar Torre , Alexander Vul
IPC: H04L67/148 , G06F9/48 , H04L41/5003 , H04L41/5019 , H04L43/0811 , H04L47/70 , H04L67/00 , H04L67/10 , H04W4/40 , H04W4/70
Abstract: An architecture to perform resource management among multiple network nodes and associated resources is disclosed. Example resource management techniques include those relating to: proactive reservation of edge computing resources; deadline-driven resource allocation; speculative edge QOS pre-allocation; and automatic QoS migration across edge computing nodes. In a specific example, a technique for service migration includes: identifying a service operated with computing resources in an edge computing system, involving computing capabilities for a connected edge device with an identified service level; identifying a mobility condition for the service, based on a change in network connectivity with the connected edge device; and performing a migration of the service to another edge computing system based on the identified mobility condition, to enable the service to be continued at the second edge computing apparatus to provide computing capabilities for the connected edge device with the identified service level.
-
公开(公告)号:US12132805B2
公开(公告)日:2024-10-29
申请号:US17542175
申请日:2021-12-03
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Karthik Kumar , Thomas Willhalm , Petar Torre , Ned Smith , Brinda Ganesh , Evan Custodio , Suraj Prabhakaran
IPC: H04L67/60 , H04L12/66 , H04L47/70 , H04L67/2885 , H04L67/5681 , H04L67/62
CPC classification number: H04L67/60 , H04L12/66 , H04L47/70 , H04L67/2885 , H04L67/5681 , H04L67/62
Abstract: Technologies for fulfilling service requests in an edge architecture include an edge gateway device to receive a request from an edge device or an intermediate tier device of an edge network to perform a function of a service by an entity hosting the service. The edge gateway device is to identify one or more input data to fulfill the request by the service and request the one or more input data from an edge resource identified to provide the input data. The edge gateway device is to provide the input data to the entity associated with the request.
-
公开(公告)号:US12038861B2
公开(公告)日:2024-07-16
申请号:US17973268
申请日:2022-10-25
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Da-Ming Chiang , Kshitij A. Doshi , Suraj Prabhakaran , Mark A. Schmisseur
IPC: G06F13/40 , G06F9/455 , G06F9/50 , G06F9/54 , G06F13/362 , G06F13/42 , G06N3/02 , G06N3/04 , G06N3/045 , G06N3/08
CPC classification number: G06F13/4068 , G06F9/45533 , G06F9/5027 , G06F9/54 , G06F13/362 , G06F13/4265 , G06F13/4282 , G06N3/02 , G06N3/04 , G06N3/045 , G06N3/08 , G06F2213/0026
Abstract: There is disclosed an example of an artificial intelligence (AI) system, including: a first hardware platform; a fabric interface configured to communicatively couple the first hardware platform to a second hardware platform; a processor hosted on the first hardware platform and programmed to operate on an AI problem; and a first training accelerator, including: an accelerator hardware; a platform inter-chip link (ICL) configured to communicatively couple the first training accelerator to a second training accelerator on the first hardware platform without aid of the processor; a fabric ICL to communicatively couple the first training accelerator to a third training accelerator on a second hardware platform without aid of the processor; and a system decoder configured to operate the fabric ICL and platform ICL to share data of the accelerator hardware between the first training accelerator and second and third training accelerators without aid of the processor.
-
公开(公告)号:US20230222363A1
公开(公告)日:2023-07-13
申请号:US18091874
申请日:2022-12-30
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Suraj Prabhakaran , Kshitij Arun Doshi , Da-Ming Chiang , Joe Cahill
IPC: G06N5/04
CPC classification number: G06N5/04
Abstract: Various systems and methods of initiating and performing contextualized AI inferencing, are described herein. In an example, operations performed with a gateway computing device to invoke an inferencing model include receiving and processing a request for an inferencing operation, selecting an implementation of the inferencing model on a remote service based on a model specification and contextual data from the edge device, and executing the selected implementation of the inferencing model, such that results from the inferencing model are provided back to the edge device. Also in an example, operations performed with an edge computing device to request an inferencing model include collecting contextual data, generating an inferencing request, transmitting the inference request to a gateway device, and receiving and processing the results of execution. Further techniques for implementing a registration of the inference model, and invoking particular variants of an inference model, are also described.
-
公开(公告)号:US11586575B2
公开(公告)日:2023-02-21
申请号:US17584092
申请日:2022-01-25
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Da-Ming Chiang , Kshitij A. Doshi , Suraj Prabhakaran , Mark A. Schmisseur
IPC: G06F13/40 , G06F13/362 , G06N3/04 , G06F13/42 , G06F9/455 , G06N3/08 , G06F9/50 , G06F9/54 , G06N3/02
Abstract: There is disclosed an example of an artificial intelligence (AI) system, including: a first hardware platform; a fabric interface configured to communicatively couple the first hardware platform to a second hardware platform; a processor hosted on the first hardware platform and programmed to operate on an AI problem; and a first training accelerator, including: an accelerator hardware; a platform inter-chip link (ICL) configured to communicatively couple the first training accelerator to a second training accelerator on the first hardware platform without aid of the processor; a fabric ICL to communicatively couple the first training accelerator to a third training accelerator on a second hardware platform without aid of the processor; and a system decoder configured to operate the fabric ICL and platform ICL to share data of the accelerator hardware between the first training accelerator and second and third training accelerators without aid of the processor.
-
公开(公告)号:US11540355B2
公开(公告)日:2022-12-27
申请号:US16995011
申请日:2020-08-17
Applicant: Intel Corporation
Inventor: Dario Sabella , Ned M. Smith , Neal Oliver , Kshitij Arun Doshi , Suraj Prabhakaran , Francesc Guim Bernat , Miltiadis Filippou
IPC: H04W88/18 , H04L67/10 , H04W48/08 , H04L9/40 , H04W4/44 , H04W88/16 , H04L41/5019 , H04L67/00 , G06F9/455 , H04L67/12 , H04L41/083 , G06F8/70 , H04W12/00 , H04L67/55 , H04W84/12
Abstract: Various systems and methods for enhancing a distributed computing environment with multiple edge hosts and user devices, including in multi-access edge computing (MEC) network platforms and settings, are described herein. A device of a lifecycle management (LCM) proxy apparatus obtains a request, from a device application, for an application multiple context of an application. The application multiple context for the application is determined. The request from the device application for the application multiple context for the application is authorized. A device application identifier based on the request is added to the application multiple context. A created response for the device application based on the authorization of the request is transmitted to the device application. The response includes an identifier of the application multiple context.
-
公开(公告)号:US20220407784A1
公开(公告)日:2022-12-22
申请号:US17832903
申请日:2022-06-06
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Kshitij Arun Doshi , Suraj Prabhakaran , Raghu Kondapalli , Alexander Bachmutsky
IPC: H04L41/5019 , H04L67/12 , H04L41/0806 , H04L41/5041 , H04L67/61 , H04L67/63
Abstract: Various systems and methods for implementing a service-level agreement (SLA) apparatus receive a request from a requester via a network interface of the gateway, the request comprising an inference model identifier that identifies a handler of the request, and a response time indicator. The response time indicator relates to a time within which the request is to be handled indicates an undefined time within which the request is to be handled. The apparatus determines a network location of a handler that is a platform or an inference model to handle the request consistent with the response time indicator, and routes the request to the handler at the network location.
-
-
-
-
-
-
-
-
-