FLASH MEMORY ACCESS SCHEDULING
    21.
    发明申请

    公开(公告)号:US20250036315A1

    公开(公告)日:2025-01-30

    申请号:US18359729

    申请日:2023-07-26

    Abstract: Various examples disclosed herein relate to trimming of system elements to prepare the elements for execution of boot code and application code. In an example embodiment, a system is provided. The system includes memory access circuitry and processing circuitry coupled to the memory access circuitry. The memory access circuitry is configured to receive a read request corresponding to a set of instructions for execution by processing circuitry stored in non-volatile memory, determine whether to preempt current access to the non-volatile memory corresponding to one or more access requests in favor of the read request based on a priority of the read request relative to the one or more access requests, obtain the set of instructions from the non-volatile memory, and supply the set of instructions to the processing circuitry. The processing circuitry executes the set of instructions.

    BATCH PROCESSING OF MULTI-CHANNEL DATA
    22.
    发明公开

    公开(公告)号:US20240338253A1

    公开(公告)日:2024-10-10

    申请号:US18361401

    申请日:2023-07-28

    CPC classification number: G06F9/5038

    Abstract: Various examples disclosed herein relate to digital signal processing, and more particularly, to processing stages of multi-channel processing pipelines in batches according to an order. A method of such processing is provided and includes retrieving multi-channel data from a memory and processing the multi-channel data with a hardware accelerator implementing a multi-stage processing pipeline for each channel of a plurality of channels. The multi-stage processing pipelines can be arranged in a cyclically descending order based on a total number of stages of each multi-stage processing pipeline. Processing the multi-channel data includes sequentially processing a plurality of batches each including one or more stages from different multi-stage processing pipelines adjacent to each other in the cyclically descending order. Processing the plurality of batches may include processing corresponding ones of the stages in parallel.

    INTEGRATED CIRCUIT WITH MULTI-APPLICATION IMAGE PROCESSING

    公开(公告)号:US20230388661A1

    公开(公告)日:2023-11-30

    申请号:US18194249

    申请日:2023-03-31

    CPC classification number: H04N23/81 H04N23/843 H04N23/88

    Abstract: In described examples, an integrated circuit includes first, second, third, and fourth image processing blocks, a data selection circuitry, and a pipeline memory. An input of the first image processing block receives raw image data. An input of the second image processing block is coupled to an output of the first image processing block. An input of the third image processing block is coupled to an output of the second image processing block. A first input of the data selection circuitry is coupled to an output of the first image processing block, and a second input of the data selection circuitry is coupled to an output of the second image processing block. A data input of the pipeline memory is coupled to an output of the data selection circuitry, and an output of the pipeline memory is coupled to an input of the fourth image processing block.

    HARDWARE EVENT TRIGGERED PIPELINE CONTROL
    24.
    发明公开

    公开(公告)号:US20230385102A1

    公开(公告)日:2023-11-30

    申请号:US18175364

    申请日:2023-02-27

    CPC classification number: G06F9/4881 G06F9/30189 G06F9/30079

    Abstract: Various embodiments disclosed herein relate to hardware enabled pipeline control. In a hardware acceleration system, pipelines are configured to include a hardware enable flag that allows hardware initiation of the pipeline based on triggering of a configurable event. The pipeline can be configured to set the event that triggers the initiation of the pipeline. For example, the end of pipeline of a first pipeline may trigger the initiation of a second pipeline. Accordingly, pipelines that are configured to allow hardware enable based on a specifically configured event are not subject to the extra processing required to initiate the pipeline via software in external memory and triggered by an external controller.

    METHODS AND APPARATUS TO EXTEND LOCAL BUFFER OF A HARDWARE ACCELERATOR

    公开(公告)号:US20230350819A1

    公开(公告)日:2023-11-02

    申请号:US18345098

    申请日:2023-06-30

    CPC classification number: G06F13/1668 G06F13/28

    Abstract: Methods and apparatus to extend local buffer of a hardware accelerator are disclosed herein. In some examples, an apparatus, including a local memory, a first hardware accelerator (HWA), a second HWA, the second HWA and the first HWA connected in a flexible data pipeline, and a spare scheduler to manage, in response to the spare scheduler inserted in the flexible data pipeline, data movement between the first HWA and the second HWA through the local memory and a memory. Local buffer extension may be performed by software to control data movement between local memory and other system memory. The other system memory may be on-chip memory and/or external memory. The HWA sub-system includes a set of spare schedulers to manage the data movement. Data aggregation may be performed in the other system memory. Additionally, the other system memory may be utilized for conversion between data line and data block.

    Package On Package Memory Interface and Configuration With Error Code Correction

    公开(公告)号:US20230258454A1

    公开(公告)日:2023-08-17

    申请号:US18306510

    申请日:2023-04-25

    CPC classification number: G01C21/206 G06F11/10 G06F2201/845

    Abstract: Information communication circuitry, including a first integrated circuit for coupling to a second integrated circuit in a package on package configuration. The first integrated circuit comprises processing circuitry for communicating information bits, and the information bits comprise data bits and error correction bits, where the error correction bits are for indicating whether data bits are received correctly. The second integrated circuit comprises a memory for receiving and storing at least some of the information bits. The information communication circuitry also includes interfacing circuitry for selectively communicating, along a number of conductors, between the package on package configuration. In a first instance, the interfacing circuitry selectively communicates only data bits along the number of conductors. In a second instance, the interfacing circuitry selectively communicates data bits along a first set of the number of conductors and error correction bits along a second set of the number of conductors.

    Methods and apparatus to extend local buffer of a hardware accelerator

    公开(公告)号:US11693795B2

    公开(公告)日:2023-07-04

    申请号:US17138740

    申请日:2020-12-30

    CPC classification number: G06F13/1668 G06F13/28

    Abstract: Methods and apparatus to extend local buffer of a hardware accelerator are disclosed herein. In some examples, an apparatus, including a local memory, a first hardware accelerator (HWA), a second HWA, the second HWA and the first HWA connected in a flexible data pipeline, and a spare scheduler to manage, in response to the spare scheduler inserted in the flexible data pipeline, data movement between the first HWA and the second HWA through the local memory and a memory. Local buffer extension may be performed by software to control data movement between local memory and other system memory. The other system memory may be on-chip memory and/or external memory. The HWA sub-system includes a set of spare schedulers to manage the data movement. Data aggregation may be performed in the other system memory. Additionally, the other system memory may be utilized for conversion between data line and data block.

    Image processing accelerator
    28.
    发明授权

    公开(公告)号:US11237991B2

    公开(公告)日:2022-02-01

    申请号:US16995364

    申请日:2020-08-17

    Abstract: A processing accelerator includes a shared memory, and a stream accelerator, a memory-to-memory accelerator, and a common DMA controller coupled to the shared memory. The stream accelerator is configured to process a real-time data stream, and to store stream accelerator output data generated by processing the real-time data stream in the shared memory. The memory-to-memory accelerator is configured to retrieve input data from the shared memory, to process the input data, and to store, in the shared memory, memory-to-memory accelerator output data generated by processing the input data. The common DMA controller is configured to retrieve stream accelerator output data from the shared memory and transfer the stream accelerator output data to memory external to the processing accelerator; and to retrieve the memory-to-memory accelerator output data from the shared memory and transfer the memory-to-memory accelerator output data to memory external to the processing accelerator.

    METHODS AND APPARATUS TO EXTEND LOCAL BUFFER OF A HARDWARE ACCELERATOR

    公开(公告)号:US20210326276A1

    公开(公告)日:2021-10-21

    申请号:US17138740

    申请日:2020-12-30

    Abstract: Methods and apparatus to extend local buffer of a hardware accelerator are disclosed herein. In some examples, an apparatus, including a local memory, a first hardware accelerator (HWA), a second HWA, the second HWA and the first HWA connected in a flexible data pipeline, and a spare scheduler to manage, in response to the spare scheduler inserted in the flexible data pipeline, data movement between the first HWA and the second HWA through the local memory and a memory. Local buffer extension may be performed by software to control data movement between local memory and other system memory. The other system memory may be on-chip memory and/or external memory. The HWA sub-system includes a set of spare schedulers to manage the data movement. Data aggregation may be performed in the other system memory. Additionally, the other system memory may be utilized for conversion between data line and data block.

    METHODS AND APPARATUS FOR IMAGE FRAME FREEZE DETECTION

    公开(公告)号:US20210136358A1

    公开(公告)日:2021-05-06

    申请号:US16669138

    申请日:2019-10-30

    Abstract: Methods, apparatus, systems, and articles of manufacture are disclosed for image frame freeze detection. An example hardware accelerator includes a core logic circuit to generate second image data based on first image data associated with a first image frame, the second image data corresponding to at least one of processed image data, transformed image data, or one or more image data statistics, a load/store engine (LSE) coupled to the core logic circuit, the LSE to determine a first CRC value based on the second image data obtained from the core logic circuit, and a first interface coupled to a second interface, the second interface coupled to memory, the first interface to transmit the first CRC value obtained from the memory to a host device.

Patent Agency Ranking