Speculative thread execution with hardware transactional memory
    22.
    发明授权
    Speculative thread execution with hardware transactional memory 有权
    使用硬件事务内存的推测线程执行

    公开(公告)号:US08881153B2

    公开(公告)日:2014-11-04

    申请号:US13801943

    申请日:2013-03-13

    Abstract: In an embodiment, if a self thread has more than one conflict, a transaction of the self thread is aborted and restarted. If the self thread has only one conflict and an enemy thread of the self thread has more than one conflict, the transaction of the self thread is committed. If the self thread only conflicts with the enemy thread and the enemy thread only conflicts with the self thread and the self thread has a key that has a higher priority than a key of the enemy thread, the transaction of the self thread is committed. If the self thread only conflicts with the enemy thread, the enemy thread only conflicts with the self thread, and the self thread has a key that has a lower priority than the key of the enemy thread, the transaction of the self thread is aborted.

    Abstract translation: 在一个实施例中,如果自线程具有多于一个冲突,则自线程的事务被中止并重新启动。 如果自线程只有一个冲突,并且自线程的敌方线程有多个冲突,则自线程的事务被提交。 如果自线程只与敌方线程冲突,敌方线程只与自线程冲突,自线程的密钥优先级高于敌方线程的密钥,则自线程的事务被提交。 如果自线程只与敌方线程相冲突,敌方线程只会与自身线程冲突,自线程的密钥优先级低于敌方线程的密钥,自身线程的事务中止。

    Debugging A High Performance Computing Program
    23.
    发明申请
    Debugging A High Performance Computing Program 有权
    调试高性能计算程序

    公开(公告)号:US20130179866A1

    公开(公告)日:2013-07-11

    申请号:US13780215

    申请日:2013-02-28

    CPC classification number: G06F11/3636 G06F11/3664

    Abstract: Methods, apparatus, and computer program products are disclosed for debugging a high performance computing program by gathering lists of addresses of calling instructions for a plurality of threads of execution of the program, assigning the threads to groups in dependence upon the addresses, and displaying the groups to identify defective threads.

    Abstract translation: 公开了用于调试高性能计算程序的方法,装置和计算机程序产品,其通过收集程序执行的多个线程的调用指令的地址列表,根据地址将线程分配给组,并且显示 组识别有缺陷的线程。

    Distributing an executable job load file to compute nodes in a parallel computer
    24.
    发明授权
    Distributing an executable job load file to compute nodes in a parallel computer 有权
    将可执行作业加载文件分配给并行计算机中的计算节点

    公开(公告)号:US09444908B2

    公开(公告)日:2016-09-13

    申请号:US14303208

    申请日:2014-06-12

    CPC classification number: H04L67/327 G06F9/5016 G06F9/5077 G06F9/52 H04L67/32

    Abstract: Distributing an executable job load file to compute nodes in a parallel computer, the parallel computer comprising a plurality of compute nodes, including: determining, by a compute node in the parallel computer, whether the compute node is participating in a job; determining, by the compute node in the parallel computer, whether a descendant compute node is participating in the job; responsive to determining that the compute node is participating in the job or that the descendant compute node is participating in the job, communicating, by the compute node to a parent compute node, an identification of a data communications link over which the compute node receives data from the parent compute node; constructing a class route for the job, wherein the class route identifies all compute nodes participating in the job; and broadcasting the executable load file for the job along the class route for the job.

    Abstract translation: 将可执行作业加载文件分发到并行计算机中的计算节点,所述并行计算机包括多个计算节点,包括:由所述并行计算机中的计算节点确定所述计算节点是否正在参与作业; 由并行计算机中的计算节点确定后代计算节点是否正在参与该作业; 响应于确定所述计算节点正在参与所述作业或者所述后代计算节点正在参与所述作业,所述计算节点向所述计算节点通信所述计算节点在其上接收数据的数据通信链路的标识 从父计算节点; 构建作业的类路由,其中​​类路由标识参与作业的所有计算节点; 并为作业的类路由广播作业的可执行加载文件。

    Core file limiter for abnormally terminating processes
    25.
    发明授权
    Core file limiter for abnormally terminating processes 有权
    用于异常终止进程的核心文件限制器

    公开(公告)号:US09003226B2

    公开(公告)日:2015-04-07

    申请号:US13677033

    申请日:2012-11-14

    CPC classification number: G06F11/0724 G06F11/0709 G06F11/0781

    Abstract: Computer program product and system to limit core file generation in a massively parallel computing system comprising a plurality of compute nodes each executing at least one task, of a plurality of tasks, by: upon determining that a first task executing on a first compute node has failed, performing an atomic load and increment operation on a core file count; generating a first core file upon determining that the core file count is below a predefined threshold; and not generating the first core file upon determining that the core file count is not below the predefined threshold.

    Abstract translation: 一种计算机程序产品和系统,用于通过以下步骤来限制大型并行计算系统中的核心文件生成:包括多个计算节点,每个计算节点通过以下步骤执行多个任务中的至少一个任务:在确定在第一计算节点上执行的第一任务具有 失败,对核心文件计数执行原子加载和递增操作; 在确定所述核心文件计数低于预定阈值时产生第一核心文件; 并且在确定核心文件计数不低于预定阈值时,不产生第一核心文件。

    Calculating a checksum with inactive networking components in a computing system
    26.
    发明授权
    Calculating a checksum with inactive networking components in a computing system 有权
    使用计算系统中的非活动网络组件计算校验和

    公开(公告)号:US08943199B2

    公开(公告)日:2015-01-27

    申请号:US13740525

    申请日:2013-01-14

    CPC classification number: H04L43/04 H04L1/00 H04L1/0061

    Abstract: Calculating a checksum utilizing inactive networking components in a computing system, including: identifying, by a checksum distribution manager, an inactive networking component, wherein the inactive networking component includes a checksum calculation engine for computing a checksum; sending, to the inactive networking component by the checksum distribution manager, metadata describing a block of data to be transmitted by an active networking component; calculating, by the inactive networking component, a checksum for the block of data; transmitting, to the checksum distribution manager from the inactive networking component, the checksum for the block of data; and sending, by the active networking component, a data communications message that includes the block of data and the checksum for the block of data.

    Abstract translation: 使用计算系统中的非活动网络组件来计算校验和,包括:由校验和分发管理器识别非活动网络组件,其中所述非活动网络组件包括用于计算校验和的校验和计算引擎; 由校验和分发管理器向不活动网络组件发送描述要由主动网络组件发送的数据块的元数据; 由非活动网络组件计算数据块的校验和; 从非活动网络组件向校验和分发管理器发送数据块的校验和; 以及由所述主动网络组件发送包括所述数据块和所述数据块的校验和的数据通信消息。

    Route planning using location/time encoded data from location aware snowplows

    公开(公告)号:US08880330B2

    公开(公告)日:2014-11-04

    申请号:US13682161

    申请日:2012-11-20

    CPC classification number: G01C21/3407 G01C21/3461 G01C21/3691

    Abstract: Techniques are described for using geographic location data of snowplows to update a snow depth associated with a section of a surface street. The route of the snowplow may be tracked continuously which allows the snow depth of the street to be reset in real-time. The updated snow depth may then be used to generate a cost of travelling on the street. The travelling cost may then be used to plan a route through the region. Alternatively, the cost may be graphically displayed on a map to permit a user to determine which surface streets to avoid. In another embodiment, the cost of travelling on a particular street may be changed by the attributes of the vehicle that will navigate the snow-covered roads, driver's experience, or the geographic terrain of a road. This new cost may then be used, for example, to plan a route or display roads to avoid.

    Remote direct memory access (‘RDMA’) in a parallel computer
    28.
    发明授权
    Remote direct memory access (‘RDMA’) in a parallel computer 有权
    并行计算机中的远程直接存储器访问('RDMA')

    公开(公告)号:US08874681B2

    公开(公告)日:2014-10-28

    申请号:US13688706

    申请日:2012-11-29

    CPC classification number: G06F15/167 G06F12/00 G06F12/1081

    Abstract: Remote direct memory access (‘RDMA’) in a parallel computer, the parallel computer including a plurality of nodes, each node including a messaging unit, including: receiving an RDMA read operation request that includes a virtual address representing a memory region at which to receive data to be transferred from a second node to the first node; responsive to the RDMA read operation request: translating the virtual address to a physical address; creating a local RDMA object that includes a counter set to the size of the memory region; sending a message that includes an DMA write operation request, the physical address of the memory region on the first node, the physical address of the local RDMA object on the first node, and a remote virtual address on the second node; and receiving the data to be transferred from the second node.

    Abstract translation: 并行计算机中的远程直接存储器访问(“RDMA”),并行计算机包括多个节点,每个节点包括消息单元,包括:接收RDMA读取操作请求,其包括虚拟地址,该虚拟地址表示存储区域 接收要从第二节点传送到第一节点的数据; 响应于RDMA读取操作请求:将虚拟地址转换为物理地址; 创建本地RDMA对象,其包括设置为存储器区域的大小的计数器; 发送包括DMA写入操作请求的消息,第一节点上的存储器区域的物理地址,第一节点上的本地RDMA对象的物理地址以及第二节点上的远程虚拟地址; 并从第二节点接收要传送的数据。

    Debugging a high performance computing program
    29.
    发明授权
    Debugging a high performance computing program 有权
    调试高性能计算程序

    公开(公告)号:US08813037B2

    公开(公告)日:2014-08-19

    申请号:US13780215

    申请日:2013-02-28

    CPC classification number: G06F11/3636 G06F11/3664

    Abstract: Methods, apparatus, and computer program products are disclosed for debugging a high performance computing program by gathering lists of addresses of calling instructions for a plurality of threads of execution of the program, assigning the threads to groups in dependence upon the addresses, and displaying the groups to identify defective threads.

    Abstract translation: 公开了用于调试高性能计算程序的方法,装置和计算机程序产品,其通过收集程序执行的多个线程的调用指令的地址列表,根据地址将线程分配给组,并且显示 组识别有缺陷的线程。

    Adaptive recovery for parallel reactive power throttling

    公开(公告)号:US08799696B2

    公开(公告)日:2014-08-05

    申请号:US13706882

    申请日:2012-12-06

    CPC classification number: G06F1/3234 G06F1/206 Y02D10/16

    Abstract: Power throttling may be used to conserve power and reduce heat in a parallel computing environment. Compute nodes in the parallel computing environment may be organized into groups based on, for example, whether they execute tasks of the same job or receive power from the same converter. Once one of compute nodes in the group detects that a parameter (i.e., temperature, current, power consumption, etc.) has exceeded a first threshold, power throttling on all the nodes in the group may be activated. However, before deactivating power throttling, a plurality of parameters associated with the group of compute nodes may be monitored to ensure they are all below a second threshold. If so, the power throttling for all of the compute nodes is deactivated.

Patent Agency Ranking