-
公开(公告)号:US10983793B2
公开(公告)日:2021-04-20
申请号:US16369846
申请日:2019-03-29
Applicant: INTEL CORPORATION
Inventor: Joshua Fryman , Ankit More , Jason Howard , Robert Pawlowski , Yigit Demir , Nick Pepperling , Fabrizio Petrini , Sriram Aananthakrishnan , Shaden Smith
Abstract: The present disclosure is directed to systems and methods of performing one or more broadcast or reduction operations using direct memory access (DMA) control circuitry. The DMA control circuitry executes a modified instruction set architecture (ISA) that facilitates the broadcast distribution of data to a plurality of destination addresses in system memory circuitry. The broadcast instruction may include broadcast of a single data value to each destination address. The broadcast instruction may include broadcast of a data array to each destination address. The DMA control circuitry may also execute a reduction instruction that facilitates the retrieval of data from a plurality of source addresses in system memory and performing one or more operations using the retrieved data. Since the DMA control circuitry, rather than the processor circuitry performs the broadcast and reduction operations, system speed and efficiency is beneficially enhanced.
-
公开(公告)号:US12204901B2
公开(公告)日:2025-01-21
申请号:US17359305
申请日:2021-06-25
Applicant: Intel Corporation
Inventor: Robert Pawlowski , Sriram Aananthakrishnan , Jason Howard , Joshua Fryman
IPC: G06F9/30 , G06F9/38 , G06F12/0875
Abstract: Techniques for operating on an indirect memory access instruction, where the instruction accesses a memory location via at least one indirect address. A pipeline processes the instruction and a memory operation engine generates a first access to the at least one indirect address and a second access to a target address determined by the at least one indirect address. A cache memory used with the pipeline and the memory operation engine caches pointers. In response to a cache hit when executing the indirect memory access instruction, operations dereference a pointer to obtain the at least one indirect address, not set a cache bit, and return data for the instruction without storing the data in the cache memory; and in response to a cache miss, operations set the cache bit, obtain, and store a cache line for a missed pointer, and return data without storing the data in the cache memory.
-
3.
公开(公告)号:US20240119015A1
公开(公告)日:2024-04-11
申请号:US18458462
申请日:2023-08-30
Applicant: Intel Corporation
Inventor: Shruti Sharma , Robert Pawlowski
CPC classification number: G06F13/1673 , G06F9/526
Abstract: Systems, apparatuses and methods may provide for technology that detects a condition in which a plurality of atomic instructions target a common address and different bit positions in a mask, generates a combined read-lock request for the plurality of atomic instructions in response to the condition, and sends the combined read-lock request to a lock buffer coupled to a memory device associated with the common address.
-
4.
公开(公告)号:US20240020253A1
公开(公告)日:2024-01-18
申请号:US18477787
申请日:2023-09-29
Applicant: Intel Corporation
Inventor: Shruti Sharma , Robert Pawlowski , Fabio Checconi , Jesmin Jahan Tithi
IPC: G06F13/28
CPC classification number: G06F13/28 , G06F2213/28
Abstract: Systems, apparatuses and methods may provide for technology that detects a plurality of sub-instruction requests from a first memory engine in a plurality of memory engines, wherein the plurality of sub-instruction requests are associated with a direct memory access (DMA) data type conversion request from a first pipeline, wherein each sub-instruction request corresponds to a data element in the DMA data type conversion request, and wherein the first memory engine is to correspond to the first pipeline, decodes the plurality of sub-instruction requests to identify one or more arguments, loads a source array from a dynamic random access memory (DRAM) in a plurality of DRAMs, wherein the operation engine is to correspond to the DRAM, and conducts a conversion of the source array from a first data type to a second data type in accordance with the one or more arguments.
-
5.
公开(公告)号:US10795819B1
公开(公告)日:2020-10-06
申请号:US16453670
申请日:2019-06-26
Applicant: Intel Corporation
Inventor: Robert Pawlowski , Bharadwaj Krishnamurthy , Vincent Cave , Jason M. Howard , Ankit More , Joshua B. Fryman
IPC: G06F12/00 , G06F12/0817 , G06F12/0811 , G06F9/38 , G06F9/30 , G06F12/0891
Abstract: Disclosed embodiments relate to a system with configurable cache sub-domains and cross-die memory coherency. In one example, a system includes R racks, each rack housing N nodes, each node incorporating D dies, each die containing C cores and a die shadow tag, each core including P pipelines and a core shadow tag, each pipelines associated with a data cache and data cache tags and being either non-coherent or coherent and one of X coherency domains, wherein each pipeline, when needing to read a cache line, issues a read request to its associated data cache, then, if need be, issues a read request to its associated core-level cache, then, if need be, issues a read request to its associated die-level cache, then, if need be, issues a no-cache remote read request to a target die being mapped to hold the cache line.
-
公开(公告)号:US12158852B2
公开(公告)日:2024-12-03
申请号:US17358832
申请日:2021-06-25
Applicant: Intel Corporation
Inventor: Robert Pawlowski , Bharadwaj Krishnamurthy , Shruti Sharma , Byoungchan Oh , Jing Fang , Daniel Klowden , Jason Howard , Joshua Fryman
IPC: G06F13/28
Abstract: Systems, methods, and apparatuses for direct memory access instruction set architecture support for flexible dense compute using a reconfigurable spatial array are described. In one embodiment, a processor includes a first type of hardware processor core that includes a two-dimensional grid of compute circuits, a memory, and a direct memory access circuit coupled to the memory and the two-dimensional grid of compute circuits; and a second different type of hardware processor core that includes a decoder circuit to decode a single instruction into a decoded single instruction, the single instruction including a first field to identify a base address of two-dimensional data in the memory, a second field to identify a number of elements in each one-dimensional array of the two-dimensional data, a third field to identify a number of one-dimensional arrays of the two-dimensional data, a fourth field to identify an operation to be performed by the two-dimensional grid of compute circuits, and a fifth field to indicate the direct memory access circuit is to move the two-dimensional data indicated by the first field, the second field, and the third field into the two-dimensional grid of compute circuits and the two-dimensional grid of compute circuits is to perform the operation on the two-dimensional data according to the fourth field, and an execution circuit to execute the decoded single instruction according to the fields.
-
公开(公告)号:US12153932B2
公开(公告)日:2024-11-26
申请号:US17129555
申请日:2020-12-21
Applicant: Intel Corporation
Inventor: Ankit More , Fabrizio Petrini , Robert Pawlowski , Shruti Sharma , Sowmya Pitchaimoorthy
IPC: G06F9/4401 , G06F13/40
Abstract: Examples include techniques for an in-network acceleration of a parallel prefix-scan operation. Examples include configuring registers of a node included in a plurality of nodes on a same semiconductor package. The registers to be configured responsive to receiving an instruction that indicates a logical tree to map to a network topology that includes the node. The instruction associated with a prefix-scan operation to be executed by at least a portion of the plurality of nodes.
-
8.
公开(公告)号:US20230333998A1
公开(公告)日:2023-10-19
申请号:US18312752
申请日:2023-05-05
Applicant: Intel Corporation
Inventor: Shruti Sharma , Robert Pawlowski , Fabio Checconi , Jesmin Jahan Tithi
IPC: G06F13/28
CPC classification number: G06F13/28
Abstract: Systems, apparatuses and methods may provide for technology that includes a plurality of memory engines corresponding to a plurality of pipelines, wherein each memory engine in the plurality of memory engines is adjacent to a pipeline in the plurality of pipelines, and wherein a first memory engine is to request one or more direct memory access (DMA) operations associated with a first pipeline, and a plurality of operation engines corresponding to a plurality of dynamic random access memories (DRAMs), wherein each operation engine in the plurality of operation engines is adjacent to a DRAM in the plurality of DRAMs, and wherein one or more of the plurality of operation engines is to conduct the one or more DMA operations based on one or more bitmaps.
-
公开(公告)号:US11630691B2
公开(公告)日:2023-04-18
申请号:US17410818
申请日:2021-08-24
Applicant: Intel Corporation
Inventor: Robert Pawlowski , Ankit More , Jason M. Howard , Joshua B. Fryman , Tina C. Zhong , Shaden Smith , Sowmya Pitchaimoorthy , Samkit Jain , Vincent Cave , Sriram Aananthakrishnan , Bharadwaj Krishnamurthy
Abstract: Disclosed embodiments relate to an improved memory system architecture for multi-threaded processors. In one example, a system includes a system comprising a multi-threaded processor core (MTPC), the MTPC comprising: P pipelines, each to concurrently process T threads; a crossbar to communicatively couple the P pipelines; a memory for use by the P pipelines, a scheduler to optimize reduction operations by assigning multiple threads to generate results of commutative arithmetic operations, and then accumulate the generated results, and a memory controller (MC) to connect with external storage and other MTPCs, the MC further comprising at least one optimization selected from: an instruction set architecture including a dual-memory operation; a direct memory access (DMA) engine; a buffer to store multiple pending instruction cache requests; multiple channels across which to stripe memory requests; and a shadow-tag coherency management unit.
-
公开(公告)号:US11360809B2
公开(公告)日:2022-06-14
申请号:US16024343
申请日:2018-06-29
Applicant: Intel Corporation
Inventor: William Paul Griffin , Joshua Fryman , Jason Howard , Sang Phill Park , Robert Pawlowski , Michael Abbott , Scott Cline , Samkit Jain , Ankit More , Vincent Cave , Fabrizio Petrini , Ivan Ganev
Abstract: Embodiments of apparatuses, methods, and systems for scheduling tasks to hardware threads are described. In an embodiment, a processor includes a multiple hardware threads and a task manager. The task manager is to issue a task to a hardware thread. The task manager includes a hardware task queue to store a descriptor for the task. The descriptor is to include a field to store a value to indicate whether the task is a single task, a collection of iterative tasks, and a linked list of tasks.
-
-
-
-
-
-
-
-
-