Remote Task Queuing by Networked Computing Devices
    2.
    发明申请
    Remote Task Queuing by Networked Computing Devices 有权
    网络计算设备的远程任务排队

    公开(公告)号:US20140331230A1

    公开(公告)日:2014-11-06

    申请号:US14164220

    申请日:2014-01-26

    Abstract: The described embodiments include a networking subsystem in a second computing device that is configured to receive a task message from a first computing device. Based on the task message, the networking subsystem updates an entry in a task queue with task information from the task message. A processing subsystem in the second computing device subsequently retrieves the task information from the task queue and performs the corresponding task. In these embodiments, the networking subsystem processes the task message (e.g., stores the task information in the task queue) without causing the processing subsystem to perform operations for processing the task message.

    Abstract translation: 所描述的实施例包括被配置为从第一计算设备接收任务消息的第二计算设备中的网络子系统。 基于任务消息,网络子系统使用来自任务消息的任务信息来更新任务队列中的条目。 第二计算设备中的处理子系统随后从任务队列检索任务信息并执行相应的任务。 在这些实施例中,网络子系统处理任务消息(例如,将任务信息存储在任务队列中)而不使处理子系统执行用于处理任务消息的操作。

    GPU remote communication with triggered operations

    公开(公告)号:US10936533B2

    公开(公告)日:2021-03-02

    申请号:US15297079

    申请日:2016-10-18

    Abstract: Methods, devices, and systems for transmitting data over a computer communications network are disclosed. A queue of communications commands can be pre-generated using a central processing unit (CPU) and stored in a device memory of a network interface controller (NIC). Thereafter, if a graphics processing unit (GPU) has data to communicate to a remote GPU, it can store the data in a send buffer, where the location in the buffer is pointed to by a pre-generated command. The GPU can then signal to the interface device that the data is ready, triggering execution of the pre-generated command to send the data.

    Conditional atomic operations in single instruction multiple data processors

    公开(公告)号:US10209990B2

    公开(公告)日:2019-02-19

    申请号:US14728643

    申请日:2015-06-02

    Abstract: A conditional fetch-and-phi operation tests a memory location to determine if the memory locations stores a specified value and, if so, modifies the value at the memory location. The conditional fetch-and-phi operation can be implemented so that it can be concurrently executed by a plurality of concurrently executing threads, such as the threads of wavefront at a GPU. To execute the conditional fetch-and-phi operation, one of the concurrently executing threads is selected to execute a compare-and-swap (CAS) operation at the memory location, while the other threads await the results. The CAS operation tests the value at the memory location and, if the CAS operation is successful, the value is passed to each of the concurrently executing threads.

    Conditional Notification Mechanism
    5.
    发明申请
    Conditional Notification Mechanism 审中-公开
    条件通知机制

    公开(公告)号:US20140250442A1

    公开(公告)日:2014-09-04

    申请号:US13782063

    申请日:2013-03-01

    CPC classification number: G06F9/542 G06F2209/543

    Abstract: The described embodiments include a computing device. In these embodiments, an entity in the computing device receives an identification of a memory location and a condition to be met by a value in the memory location. Upon a predetermined event occurring, the entity causes an operation to be performed when the value in the memory location meets the condition.

    Abstract translation: 所描述的实施例包括计算设备。 在这些实施例中,计算设备中的实体通过存储器位置中的值接收存储器位置的标识和要满足的条件。 当预定事件发生时,当存储器位置中的值满足条件时,实体导致执行操作。

    Network interface controller-based scheduling of processing tasks in a distributed computing system

    公开(公告)号:US10963309B2

    公开(公告)日:2021-03-30

    申请号:US15267936

    申请日:2016-09-16

    Abstract: Techniques for scheduling processing tasks in a device having multiple computing elements are disclosed. A network interface controller of the device receives processing tasks, for execution on the computing elements, from a network that is external to the device. The network interface controller schedules the tasks for execution on the computing devices based on policy data available to the network interface controller. A scheduler within the network interface controller, which can be implemented as a standalone processing unit (such as a microcontroller, a programmable processing core, or an application specific integrated circuit), performs such scheduling, thereby freeing the central processing unit of the device from the burden of performing scheduling operations. The scheduler schedules the tasks according to any technically feasible scheduling technique.

    GPU REMOTE COMMUNICATION WITH TRIGGERED OPERATIONS

    公开(公告)号:US20180107627A1

    公开(公告)日:2018-04-19

    申请号:US15297079

    申请日:2016-10-18

    CPC classification number: G06F15/17331 H04L47/50 H04L49/9063

    Abstract: Methods, devices, and systems for transmitting data over a computer communications network are disclosed. A queue of communications commands can be pre-generated using a central processing unit (CPU) and stored in a device memory of a network interface controller (NIC). Thereafter, if a graphics processing unit (GPU) has data to communicate to a remote GPU, it can store the data in a send buffer, where the location in the buffer is pointed to by a pre-generated command. The GPU can then signal to the interface device that the data is ready, triggering execution of the pre-generated command to send the data.

    Remote task queuing by networked computing devices
    8.
    发明授权
    Remote task queuing by networked computing devices 有权
    网络计算设备的远程任务排队

    公开(公告)号:US09582402B2

    公开(公告)日:2017-02-28

    申请号:US14164220

    申请日:2014-01-26

    Abstract: The described embodiments include a networking subsystem in a second computing device that is configured to receive a task message from a first computing device. Based on the task message, the networking subsystem updates an entry in a task queue with task information from the task message. A processing subsystem in the second computing device subsequently retrieves the task information from the task queue and performs the corresponding task. In these embodiments, the networking subsystem processes the task message (e.g., stores the task information in the task queue) without causing the processing subsystem to perform operations for processing the task message.

    Abstract translation: 所描述的实施例包括被配置为从第一计算设备接收任务消息的第二计算设备中的网络子系统。 基于任务消息,网络子系统使用来自任务消息的任务信息来更新任务队列中的条目。 第二计算设备中的处理子系统随后从任务队列检索任务信息并执行相应的任务。 在这些实施例中,网络子系统处理任务消息(例如,将任务信息存储在任务队列中)而不使处理子系统执行用于处理任务消息的操作。

    Method for memory consistency among heterogeneous computer components
    10.
    发明授权
    Method for memory consistency among heterogeneous computer components 有权
    异构计算机组件之间内存一致性的方法

    公开(公告)号:US09361118B2

    公开(公告)日:2016-06-07

    申请号:US14275271

    申请日:2014-05-12

    Abstract: A method, computer program product, and system is described that determines the correctness of using memory operations in a computing device with heterogeneous computer components. Embodiments include an optimizer based on the characteristics of a Sequential Consistency for Heterogeneous-Race-Free (SC for HRF) model that analyzes a program and determines the correctness of the ordering of events in the program. HRF models include combinations of the properties: scope order, scope inclusion, and scope transitivity. The optimizer can determine when a program is heterogeneous-race-free in accordance with an SC for HRF memory consistency model. For example, the optimizer can analyze a portion of program code, respect the properties of the SC for HRF model, and determine whether a value produced by a store memory event will be a candidate for a value observed by a load memory event. In addition, the optimizer can determine whether reordering of events is possible.

    Abstract translation: 描述了一种方法,计算机程序产品和系统,其确定在具有异构计算机组件的计算设备中使用存储器操作的正确性。 实施例包括基于用于异构无竞争(SC for HRF)的顺序一致性的特性的优化器,该模型分析程序并确定程序中的事件的顺序的正确性。 HRF模型包括属性的组合:范围顺序,范围包含和范围传递性。 优化器可以根据HR对HRF内存一致性模型的SC来确定程序何时是异构无竞争的。 例如,优化器可以分析程序代码的一部分,尊重SC的HRF模型的属性,并且确定由存储器存储器事件产生的值是否将是由加载存储器事件观察到的值的候选。 此外,优化器可以确定是否可能重新排序事件。

Patent Agency Ranking