-
公开(公告)号:US09329900B2
公开(公告)日:2016-05-03
申请号:US13730539
申请日:2012-12-28
Applicant: Intel Corporation
Inventor: Paolo Narvaez , Ganapati N. Srinivasa , Eugene Gorbatov , Dheeraj R. Subbareddy , Mishali Naik , Alon Naveh , Abirami Prabhakaran , Eliezer Weissmann , David A. Koufaty , Paul Brett , Scott D. Hahn , Andrew J. Herdrich , Ravishankar Iyer , Nagabhushan Chitlur , Inder M. Sodhi , Gaurav Khanna , Russell J. Fenger
CPC classification number: G06F9/5044 , G06F9/45533 , G06F9/5077 , G06F9/5094 , G06F15/80 , Y02D10/22
Abstract: A heterogeneous processor architecture is described. For example, a processor according to one embodiment of the invention comprises: a first set of one or more physical processor cores having first processing characteristics; a second set of one or more physical processor cores having second processing characteristics different from the first processing characteristics; virtual-to-physical (V-P) mapping logic to expose a plurality of virtual processors to software, the plurality of virtual processors to appear to the software as a plurality of homogeneous processor cores, the software to allocate threads to the virtual processors as if the virtual processors were homogeneous processor cores; wherein the V-P mapping logic is to map each virtual processor to a physical processor within the first set of physical processor cores or the second set of physical processor cores such that a thread allocated to a first virtual processor by software is executed by a physical processor mapped to the first virtual processor from the first set or the second set of physical processors.
Abstract translation: 描述异构处理器架构。 例如,根据本发明的一个实施例的处理器包括:具有第一处理特性的一个或多个物理处理器核心的第一组; 具有不同于所述第一处理特性的第二处理特性的第二组一个或多个物理处理器核; 虚拟到物理(VP)映射逻辑,以将多个虚拟处理器暴露给软件,所述多个虚拟处理器将软件呈现为多个同构的处理器核,所述软件将线程分配给虚拟处理器,如同 虚拟处理器是同类处理器核心; 其中所述VP映射逻辑将每个虚拟处理器映射到所述第一物理处理器核心集合或所述第二物理处理器核心集合内的物理处理器,使得通过软件分配给第一虚拟处理器的线程由物理处理器映射执行 从第一组或第二组物理处理器到第一虚拟处理器。
-
公开(公告)号:US12132790B2
公开(公告)日:2024-10-29
申请号:US17875672
申请日:2022-07-28
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Patrick Bohan , Kshitij Arun Doshi , Brinda Ganesh , Andrew J. Herdrich , Monica Kenguva , Karthik Kumar , Patrick G Kutch , Felipe Pastor Beneyto , Rashmin Patel , Suraj Prabhakaran , Ned M. Smith , Petar Torre , Alexander Vul
IPC: H04L67/148 , G06F9/48 , H04L41/5003 , H04L41/5019 , H04L43/0811 , H04L47/70 , H04L67/00 , H04L67/10 , H04W4/40 , H04W4/70
CPC classification number: H04L67/148 , G06F9/4856 , H04L41/5019 , H04L43/0811 , H04L47/82 , H04L67/10 , H04L67/34 , H04W4/40 , H04W4/70 , H04L41/5003
Abstract: An architecture to perform resource management among multiple network nodes and associated resources is disclosed. Example resource management techniques include those relating to: proactive reservation of edge computing resources; deadline-driven resource allocation; speculative edge QoS pre-allocation; and automatic QoS migration across edge computing nodes. In a specific example, a technique for service migration includes: identifying a service operated with computing resources in an edge computing system, involving computing capabilities for a connected edge device with an identified service level; identifying a mobility condition for the service, based on a change in network connectivity with the connected edge device; and performing a migration of the service to another edge computing system based on the identified mobility condition, to enable the service to be continued at the second edge computing apparatus to provide computing capabilities for the connected edge device with the identified service level.
-
公开(公告)号:US12106106B2
公开(公告)日:2024-10-01
申请号:US17134256
申请日:2020-12-25
Applicant: Intel Corporation
Inventor: Andrew J. Herdrich , Jason W. Brandt
CPC classification number: G06F9/3009 , G06F9/30101 , G06F9/3851 , G06F9/3885
Abstract: Embodiments for memory bandwidth monitoring extensible counters are described. In embodiments, an apparatus includes memory bandwidth monitoring hardware to monitor an event, a shared cache to be shared by multiple cores. At least one of the cores is to execute multiple threads and includes at least three registers. The first register is programmable by software to store a thread identifier of one of threads and an event identifier of the event during execution of the thread. At least one value of the event identifier corresponds to a shared cache miss. The second register is to provide to the software a second value corresponding to a number of bits available to represent the count. The third register is to provide to the software a count of occurrences of the event and an indicator to indicate whether the count reached a maximum count representable by the number of bits.
-
84.
公开(公告)号:US11816036B2
公开(公告)日:2023-11-14
申请号:US17738919
申请日:2022-05-06
Applicant: Intel Corporation
Inventor: Anil Vasudevan , Venkata Krishnan , Andrew J. Herdrich , Ren Wang , Robert G. Blankenship , Vedaraman Geetha , Shrikant M. Shah , Marshall A. Millier , Raanan Sade , Binh Q. Pham , Olivier Serres , Chyi-Chang Miao , Christopher B. Wilkerson
IPC: G06F12/0868 , G06F12/0897 , G06F3/06 , G06F12/0811 , G06F12/0871
CPC classification number: G06F12/0868 , G06F3/065 , G06F3/068 , G06F3/0619 , G06F12/0811 , G06F12/0871 , G06F12/0897 , G06F2212/1024 , G06F2212/283 , G06F2212/311 , G06F2212/6046
Abstract: Method and system for performing data movement operations is described herein. One embodiment of a method includes: storing data for a first memory address in a cache line of a memory of a first processing unit, the cache line associated with a coherency state indicating that the memory has sole ownership of the cache line; decoding an instruction for execution by a second processing unit, the instruction comprising a source data operand specifying the first memory address and a destination operand specifying a memory location in the second processing unit; and responsive to executing the decoded instruction, copying data from the cache line of the memory of the first processing unit as identified by the first memory address, to the memory location of the second processing unit, wherein responsive to the copy, the cache line is to remain in the memory and the coherency state is to remain unchanged.
-
公开(公告)号:US11442760B2
公开(公告)日:2022-09-13
申请号:US15200725
申请日:2016-07-01
Applicant: Intel Corporation
Inventor: Barry E. Huntley , Jr-Shian Tsai , Gilbert Neiger , Rajesh M. Sankaran , Mesut A. Ergin , Ravi L. Sahita , Andrew J. Herdrich , Wei Wang
Abstract: A processor of an aspect includes a decode unit to decode an aperture access instruction, and an execution unit coupled with the decode unit. The execution unit, in response to the aperture access instruction, is to read a host physical memory address, which is to be associated with an aperture that is to be in system memory, from an access protected structure, and access data within the aperture at a host physical memory address that is not to be obtained through address translation. Other processors are also disclosed, as are methods, systems, and machine-readable medium storing aperture access instructions.
-
86.
公开(公告)号:US20220261351A1
公开(公告)日:2022-08-18
申请号:US17738919
申请日:2022-05-06
Applicant: Intel Corporation
Inventor: Anil Vasudevan , Venkata Krishnan , Andrew J. Herdrich , Ren Wang , Robert G. Blankenship , Vedaraman Geetha , Shrikant M. Shah , Marshall A. Millier , Raanan Sade , Binh Q. Pham , Olivier Serres , Chyi-Chang Miao , Christopher B. Wilkerson
IPC: G06F12/0868 , G06F12/0897 , G06F3/06 , G06F12/0811 , G06F12/0871
Abstract: Method and system for performing data movement operations is described herein. One embodiment of a method includes: storing data for a first memory address in a cache line of a memory of a first processing unit, the cache line associated with a coherency state indicating that the memory has sole ownership of the cache line; decoding an instruction for execution by a second processing unit, the instruction comprising a source data operand specifying the first memory address and a destination operand specifying a memory location in the second processing unit; and responsive to executing the decoded instruction, copying data from the cache line of the memory of the first processing unit as identified by the first memory address, to the memory location of the second processing unit, wherein responsive to the copy, the cache line is to remain in the memory and the coherency state is to remain unchanged.
-
公开(公告)号:US20220206862A1
公开(公告)日:2022-06-30
申请号:US17134252
申请日:2020-12-25
Applicant: Intel Corporation
Inventor: Monica Gupta , Russell Fenger , Andrew J. Herdrich , Rajshree Chabukswar , Jumnit Hong , Sneha Gohad
IPC: G06F9/50
Abstract: Embodiments of apparatuses, methods, and systems for resource control based on software priority are described. In embodiments, an apparatus includes resource sharing hardware and multiple cores. The resource sharing hardware is to share the shared resource among the cores. A first core includes first execution circuitry to execute multiple threads. The first core also includes registers programmable by software. A first register is to store a first identifier of a first thread and a first priority tag to indicate a first priority of the first thread relative to a second priority of a second thread. A second register to store a second identifier of the second thread and a second priority tag to indicate the second priority of the second thread relative to the first priority of the first thread. The resource sharing hardware is to use the first priority and the second priority to control access to the shared resource by the first thread and the second thread.
-
公开(公告)号:US20220182284A1
公开(公告)日:2022-06-09
申请号:US17505262
申请日:2021-10-19
Applicant: Intel Corporation
Inventor: Rashmin Patel , Monica Kenguva , Francesc Guim Bernat , Edwin Verplanke , Andrew J. Herdrich
IPC: H04L41/0816 , H04L41/14
Abstract: Technologies for analyzing and optimizing workloads (e.g., virtual network functions) executing on edge resources are disclosed. According to one embodiment disclosed herein, a compute device launches a virtualized system including a virtual network function and a performance manager, the performance manager to monitor a current resource usage of the virtual network function as a function of a performance profile. The compute device determines, in response to a determination that one or more quality-of-service (QoS) requirements is not satisfied, whether one or more resources from the platform are available for satisfying the QoS requirements. The compute device receives, in response to a determination that the one or more resources are available for satisfying the QoS requirements, the one or more resources and updates the performance profile as a function of the received resources.
-
公开(公告)号:US20220166846A1
公开(公告)日:2022-05-26
申请号:US17390658
申请日:2021-07-30
Applicant: Intel Corporation
Inventor: Ramanathan Sethuraman , Timothy Verrall , Ned M. Smith , Thomas Willhalm , Brinda Ganesh , Francesc Guim Bernat , Karthik Kumar , Evan Custodio , Suraj Prabhakaran , Ignacio Astilleros Diez , Nilesh K. Jain , Ravi Iyer , Andrew J. Herdrich , Alexander Vul , Patrick G. Kutch , Kevin Bohan , Trevor Cooper
IPC: H04L67/303 , H04L9/40 , H04L9/08 , H04L67/12
Abstract: Technologies for managing telemetry and sensor data on an edge networking platform are disclosed. According to one embodiment disclosed herein, a device monitors telemetry data associated with multiple services provided in the edge networking platform. The device identifies, for each of the services and as a function of the associated telemetry data, one or more service telemetry patterns. The device generates a profile including the identified service telemetry patterns.
-
公开(公告)号:US11036531B2
公开(公告)日:2021-06-15
申请号:US15635124
申请日:2017-06-27
Applicant: Intel Corporation
Inventor: Patrick Connor , James R. Hearn , Scott P. Dubal , Andrew J. Herdrich , Kapil Sood
IPC: G06F9/455
Abstract: Examples may include techniques to live migrate a virtual machine (VM) using disaggregated computing resources including compute and memory resources. Examples include copying data between allocated memory resources that serve as near or far memory for compute resources supporting the VM at a source or destination server in order to initiate and complete the live migration of the VM.
-
-
-
-
-
-
-
-
-