Process synchronization between engines using data in a memory location

    公开(公告)号:US10078879B2

    公开(公告)日:2018-09-18

    申请号:US14692984

    申请日:2015-04-22

    CPC classification number: G06T1/20 G06F9/48 G06F9/52 G06F15/167 G09G5/001

    Abstract: Memory-based semaphores are described that are useful for synchronizing processes between different processing engines. In one example, operations include executing a first process at a first processing engine, the executing including updating a memory register, sending a signal from the first processing engine to a second processing engine that the memory register has been updated, the signal including a memory register address to identify the updated memory register inline data and a dataword, fetching data from the memory register by the second processing engine, comparing the fetched data to the received dataword, and conditionally executing a next command of a second process at the second processing engine based on the comparison.

    Hardware assist for privilege access violation checks

    公开(公告)号:US09633230B2

    公开(公告)日:2017-04-25

    申请号:US13649798

    申请日:2012-10-11

    CPC classification number: G06F21/74 G06F21/84 G06T15/005

    Abstract: Techniques are disclosed for processing rendering engine workload of a graphics system in a secure fashion, wherein at least some security processing of the workload is offloaded from software-based security parsing to hardware-based security parsing. In some embodiments, commands from a given application are received by a user-mode driver (UMD), which is configured to generate a command buffer delineated into privileged and/or non-privileged command sections. The delineated command buffer can then be passed by the UMD to a kernel-mode driver (KMD), which is configured to parse and validate only privileged buffer sections, but to issue all other batch buffers with a privilege indicator set to non-privileged. A graphics processing unit (GPU) can receive the privilege-designated batch buffers from the KMD, and is configured to disallow execution of any privileged command from a non-privileged batch buffer, while any privileged commands from privileged batch buffers are unrestricted by the GPU.

    Memory address re-mapping of graphics data
    55.
    发明授权
    Memory address re-mapping of graphics data 有权
    图形数据的内存地址重映射

    公开(公告)号:US09223603B2

    公开(公告)日:2015-12-29

    申请号:US13932963

    申请日:2013-07-01

    Abstract: A method and apparatus for creating, updating, and using guest physical address (GPA) to host physical address (HPA) shadow translation tables for translating GPAs of graphics data direct memory access (DMA) requests of a computing environment implementing a virtual machine monitor to support virtual machines. The requests may be sent through a render or display path of the computing environment from one or more virtual machines, transparently with respect to the virtual machine monitor. The creating, updating, and using may be performed by a memory controller detecting entries sent to existing global and page directory tables, forking off shadow table entries from the detected entries, and translating GPAs to HPAs for the shadow table entries.

    Abstract translation: 一种用于创建,更新和使用访客物理地址(GPA)以主机物理地址(HPA)影子转换表的方法和装置,用于将实现虚拟机监视器的计算环境的图形数据直接存储器访问(DMA)请求的GPA转换为 支持虚拟机。 可以通过虚拟机监视器透明地从一个或多个虚拟机通过计算环境的呈现或显示路径发送请求。 创建,更新和使用可以由存储器控制器执行,该存储器控制器检测发送到现有全局和页目录表的条目,从检测到的条目中分离影子表条目,以及将影子表条目的GPA转换为HPA。

    MODULAR GPU ARCHITECTURE FOR CLIENTS AND SERVERS

    公开(公告)号:US20230109990A1

    公开(公告)日:2023-04-13

    申请号:US17496467

    申请日:2021-10-07

    Abstract: One embodiment provides a graphics processor including an active base die including a fabric interconnect and a chiplet including a switched fabric, wherein the chiplet couples with the active base die via an array of interconnect structures, the array of interconnect structures couple the fabric interconnect with the switched fabric, and the chiplet includes a first modular interconnect configured to couple a block of graphics processing resources to the switched fabric and a second modular interconnect configured to couple a memory subsystem with the switched fabric and the block of graphics processing resources, the memory interconnect including a set of memory controllers and a set of physical interfaces.

    UNIFIED SUBMIT PORT FOR GRAPHICS PROCESSING

    公开(公告)号:US20230094002A1

    公开(公告)日:2023-03-30

    申请号:US17484711

    申请日:2021-09-24

    Abstract: Dynamic routing of texture-load in graphics processing is described. An example of an apparatus includes a graphics processor including a plurality of processing engines of a class of processing engines of the graphic processor; a set of queues for the plurality of processing engines; and a unified submit port for the plurality of processing engines, wherein the unified submit port is to notify a scheduler regarding availability of slots in the set of queues for receipt of workload contexts; and wherein, upon the unified submit port receiving a workload context for processing by the plurality of processing engines, the unified submit port is to detect an available processing engine of the plurality of processing engines and direct the received context to a slot of the set of queues for processing by the available processing engine.

    System, apparatus and method for multi-die distributed memory mapped input/output support

    公开(公告)号:US11157431B2

    公开(公告)日:2021-10-26

    申请号:US16447025

    申请日:2019-06-20

    Abstract: In one embodiment, a method includes: receiving, in a root tile of an accelerator device having a plurality of tiles, a message from a processor, the message comprising a register write request to a register of a first remote tile of the plurality of remote tiles; decoding, in an endpoint controller of the root tile, a system address of the message to identify a destination tile for the message, based at least in part on a base address register decode of the system address; and in response to identifying the first remote tile as the destination tile, updating a first portion of an address offset field of the system address to a predetermined value and directing the message to the first remote tile coupled to the root tile via a sideband interconnect. Other embodiments are described and claimed.

Patent Agency Ranking