INTERCONNECTING GLOBAL VIRTUAL PLANES

    公开(公告)号:US20250123875A1

    公开(公告)日:2025-04-17

    申请号:US18912318

    申请日:2024-10-10

    Abstract: A network environment comprises a plurality of host machines that are coupled to each other via a network fabric comprising a plurality of switches, that in turn include a plurality of ports. Each host machine comprises one or more GPUs. A first subset of ports from is associated with a first virtual plane, wherein the first virtual plane identifies a first collection of resources to be used for communicating packets from/to host machines associated with the first virtual plane. A second subset of ports is associated with a second virtual plane that is different from the first virtual plane. A first host machine and a second host machine are associated with the first virtual plane and the second virtual plane, respectively. A packet is communicated from the first host machine to the second host machine using ports from the first subset of ports and the second subset of ports.

    IN-MEMORY WORKFLOW MANAGEMENT IN EDGE DEVICES

    公开(公告)号:US20220326848A1

    公开(公告)日:2022-10-13

    申请号:US17531566

    申请日:2021-11-19

    Abstract: Techniques discussed herein relate to providing in-memory workflow management at an edge device (e.g., a computing device distinct from and operating remotely with respect to a data center). The edge device can operate as a computing node in a computing cluster of edge devices and implement a hosting environment (e.g., a distributed data plane). A work request can be obtained by an in-memory workflow manager of the edge device. The work request may include an intended state of a data plane resource (e.g., a computing cluster, a virtual machine, etc.). The in-memory workflow manager can determine the work request has not commenced and initialize an in-memory execution thread to execute orchestration tasks to configure a data plane of the computing cluster according to the intended state. Current state data corresponding to the configured data plane may be provided to the user device and eventually displayed.

    CLOUD SCALE MULTI-TENANCY FOR RDMA OVER CONVERGED ETHERNET (RoCE)

    公开(公告)号:US20220209993A1

    公开(公告)日:2022-06-30

    申请号:US17165877

    申请日:2021-02-02

    Abstract: Techniques and apparatus for data networking are described. In one example, a method includes receiving a first Layer-2 Remote Direct Memory Access (RDMA) packet which includes a virtual local area network (VLAN) tag and a quality-of-service (QoS) data field; converting the first Layer-2 RDMA packet to a first Layer-3 encapsulated packet; and forwarding the first Layer-3 encapsulated packet to a switch fabric. In this method, the converting includes adding at least one header to the first Layer-2 RDMA packet, where the at least one header includes: a virtual network identifier that is based on information from the VLAN tag, and a QoS value that is based on information from the QoS data field.

    Composable edge device platforms
    6.
    发明授权

    公开(公告)号:US11349710B1

    公开(公告)日:2022-05-31

    申请号:US17531632

    申请日:2021-11-19

    Abstract: Techniques discussed herein relate to providing composable edge devices. In some embodiments, a user request specifying a set of services to be executed at a cloud-computing edge device may be received by a computing device operated by a cloud computing provider. A manifest may be generated in accordance with the user request. The manifest may specify a configuration for the cloud-computing edge device. Another request can be received specifying the same or a different set of services to be executed at another edge device. Another manifest which specifies the configuration for that edge device may be generated and subsequently used to provision the request set of services on that device. In this manner, manifests can be used to compose the platform to be utilized at any given edge device.

    INTERCONNECTION OF GLOBAL VIRTUAL PLANES

    公开(公告)号:US20250126080A1

    公开(公告)日:2025-04-17

    申请号:US18912251

    申请日:2024-10-10

    Abstract: A network environment comprises a plurality of host machines that are coupled to each other via a network fabric comprising a plurality of switches, that in turn include a plurality of ports. Each host machine comprises one or more GPUs. A first subset of ports from is associated with a first virtual plane, wherein the first virtual plane identifies a first collection of resources to be used for communicating packets from/to host machines associated with the first virtual plane. A second subset of ports is associated with a second virtual plane that is different from the first virtual plane. A first host machine and a second host machine are associated with the first virtual plane and the second virtual plane, respectively. A packet is communicated from the first host machine to the second host machine using ports from the first subset of ports and the second subset of ports.

    NETWORK LOCALITY IN A GPU SUPER-CLUSTER
    8.
    发明公开

    公开(公告)号:US20240160495A1

    公开(公告)日:2024-05-16

    申请号:US18500474

    申请日:2023-11-02

    CPC classification number: G06F9/5083

    Abstract: Each host machine of a plurality of host machines stores hierarchical locality information for the host machine that identifies at least a rack comprising the host machine, and a block of a plurality of blocks hosting the rack. The host machine is associated with one or more graphical processing units (GPUs), and wherein GPUs included in a first block operate at a first speed and GPUs included in a second block operate at a second speed that is different than the first speed. Responsive to receiving a request requesting execution of a workload, one or more host machines are identified as being available for executing the workload, and the hierarchical locality information and linkage information of the one or more host machines is provided in response to the request.

    SUPERCLUSTER NETWORK OF GRAPHICAL PROCESSING UNITS (GPUS)

    公开(公告)号:US20240152396A1

    公开(公告)日:2024-05-09

    申请号:US18500463

    申请日:2023-11-02

    CPC classification number: G06F9/5027 G06F13/4022 G06F2209/505

    Abstract: Described herein is a network fabric including a plurality of graphical processing unit (GPU) clusters. The plurality of GPU clusters includes at least a first GPU cluster operating at a first speed and a second GPU cluster operating at a second speed that is different than the first speed. The network fabric includes a plurality of blocks, wherein each block includes: (a) one or more racks that host a GPU cluster, and (b) a plurality of switches arranged in a hierarchical structure that communicatively couple the block to other blocks included in the network fabric. Responsive to receiving a request to execute a workload, allocating one or more GPUs from the plurality of GPU clusters to execute the workload.

Patent Agency Ranking