-
公开(公告)号:US11412052B2
公开(公告)日:2022-08-09
申请号:US16235137
申请日:2018-12-28
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Patrick Bohan , Kshitij Arun Doshi , Brinda Ganesh , Andrew J. Herdrich , Monica Kenguva , Karthik Kumar , Patrick G Kutch , Felipe Pastor Beneyto , Rashmin Patel , Suraj Prabhakaran , Ned M. Smith , Petar Torre , Alexander Vul
IPC: H04L67/148 , H04L43/0811 , H04L67/10 , H04L41/5019 , H04L67/00 , H04L41/5003 , H04L47/70 , H04W4/40 , H04W4/70 , G06F9/48
Abstract: An architecture to perform resource management among multiple network nodes and associated resources is disclosed. Example resource management techniques include those relating to: proactive reservation of edge computing resources; deadline-driven resource allocation; speculative edge QoS pre-allocation; and automatic QoS migration across edge computing nodes. In a specific example, a technique for service migration includes: identifying a service operated with computing resources in an edge computing system, involving computing capabilities for a connected edge device with an identified service level; identifying a mobility condition for the service, based on a change in network connectivity with the connected edge device; and performing a migration of the service to another edge computing system based on the identified mobility condition, to enable the service to be continued at the second edge computing apparatus to provide computing capabilities for the connected edge device with the identified service level.
-
公开(公告)号:US11379342B2
公开(公告)日:2022-07-05
申请号:US16827410
申请日:2020-03-23
Applicant: Intel Corporation
Inventor: Ren Wang , Bin Li , Andrew J. Herdrich , Tsung-Yuan C. Tai , Ramakrishna Huggahalli
IPC: G06F11/30 , G06F11/34 , G06F12/0811 , G06F12/121 , G06F13/16 , G06F13/42 , G06F12/128 , G06F12/084 , G06F12/0888 , H04L67/1097 , G06F13/28
Abstract: There is disclosed in one example a computing apparatus, including: a processor; a multilevel cache including a plurality of cache levels; a peripheral device configured to write data directly to a selected cache level; and a cache monitoring circuit, including a cache counter to track cache lines evicted from the selected cache level without being processed; and logic to provide a direct write policy according to the cache counter.
-
公开(公告)号:US10817425B2
公开(公告)日:2020-10-27
申请号:US14583389
申请日:2014-12-26
Applicant: Intel Corporation
Inventor: Ren Wang , Andrew J. Herdrich , Yen-cheng Liu , Herbert H. Hum , Jong Soo Park , Christopher J. Hughes , Namakkal N. Venkatesan , Adrian C. Moga , Aamer Jaleel , Zeshan A. Chishti , Mesut A. Ergin , Jr-shian Tsai , Alexander W. Min , Tsung-yuan C. Tai , Christian Maciocco , Rajesh Sankaran
IPC: G06F12/0842 , G06F12/0831 , G06F12/0893 , G06F12/109 , G06F12/0813 , G06F9/455
Abstract: Methods and apparatus implementing Hardware/Software co-optimization to improve performance and energy for inter-VM communication for NFVs and other producer-consumer workloads. The apparatus include multi-core processors with multi-level cache hierarchies including and L1 and L2 cache for each core and a shared last-level cache (LLC). One or more machine-level instructions are provided for proactively demoting cachelines from lower cache levels to higher cache levels, including demoting cachelines from L1/L2 caches to an LLC. Techniques are also provided for implementing hardware/software co-optimization in multi-socket NUMA architecture system, wherein cachelines may be selectively demoted and pushed to an LLC in a remote socket. In addition, techniques are disclosure for implementing early snooping in multi-socket systems to reduce latency when accessing cachelines on remote sockets.
-
44.
公开(公告)号:US10606755B2
公开(公告)日:2020-03-31
申请号:US15640060
申请日:2017-06-30
Applicant: Intel Corporation
Inventor: Anil Vasudevan , Venkata Krishnan , Andrew J. Herdrich , Ren Wang , Robert G. Blankenship , Vedaraman Geetha , Shrikant M. Shah , Marshall A. Millier , Raanan Sade , Binh Q. Pham , Olivier Serres , Chyi-Chang Miao , Christopher B. Wilkerson
IPC: G06F12/0868 , G06F12/0897 , G06F3/06 , G06F12/0811 , G06F12/0871
Abstract: Method and system for performing data movement operations is described herein. One embodiment of a method includes: storing data for a first memory address in a cache line of a memory of a first processing unit, the cache line associated with a coherency state indicating that the memory has sole ownership of the cache line; decoding an instruction for execution by a second processing unit, the instruction comprising a source data operand specifying the first memory address and a destination operand specifying a memory location in the second processing unit; and responsive to executing the decoded instruction, copying data from the cache line of the memory of the first processing unit as identified by the first memory address, to the memory location of the second processing unit, wherein responsive to the copy, the cache line is to remain in the memory and the coherency state is to remain unchanged.
-
公开(公告)号:US10554505B2
公开(公告)日:2020-02-04
申请号:US13630545
申请日:2012-09-28
Applicant: Intel Corporation
Inventor: Mrittika Ganguli , Muthuvel M. I , Ananth S. Narayan , Jaideep Moses , Andrew J. Herdrich , Rahul Khanna
IPC: H04L12/24
Abstract: In accordance with some embodiments, a cloud service provider may operate a data center in a way that dynamically reallocates resources across nodes within the data center based on both utilization and service level agreements. In other words, the allocation of resources may be adjusted dynamically based on current conditions. The current conditions in the data center may be a function of the nature of all the current workloads. Instead of simply managing the workloads in a way to increase overall execution efficiency, the data center instead may manage the workload to achieve quality of service requirements for particular workloads according to service level agreements.
-
公开(公告)号:US10339023B2
公开(公告)日:2019-07-02
申请号:US14496216
申请日:2014-09-25
Applicant: Intel Corporation
Inventor: Ren Wang , Tsung-Yuan C. Tai , Paul S. Diefenbaugh , Andrew J. Herdrich
IPC: G06F12/084 , G06F11/30 , G06F11/34 , G06F12/0895 , G06F1/3206 , G06F1/3234 , G06F1/3287
Abstract: In one embodiment, a processor includes: a plurality of cores each to independently execute instructions; a shared cache memory coupled to the plurality of cores and having a plurality of clusters each associated with one or more of the plurality of cores; a plurality of cache activity monitors each associated with one of the plurality of clusters, where each cache activity monitor is to monitor one or more performance metrics of the corresponding cluster and to output cache metric information; a plurality of thermal sensors each associated with one of the plurality of clusters and to output thermal information; and a logic coupled to the plurality of cores to receive the cache metric information from the plurality of cache activity monitors and the thermal information and to schedule one or more threads to a selected core based at least in part on the cache metric information and the thermal information for the cluster associated with the selected core. Other embodiments are described and claimed.
-
公开(公告)号:US10268580B2
公开(公告)日:2019-04-23
申请号:US15282483
申请日:2016-09-30
Applicant: Intel Corporation
Inventor: Kshitij A. Doshi , Namakkal N. Venkatesan , Ren Wang , Andrew J. Herdrich
IPC: G06F9/30 , G06F12/128 , G06F12/0811
Abstract: Processors and methods implementing a machine instruction to perform cache line demotion on multiple cache lines to enable efficient sharing of cache lines between processor cores. One general aspect includes a processor comprising: a plurality of hardware processor cores, where each of the hardware processor cores to include a first cache. The processor also includes a second cache, communicatively coupled to and shared by the plurality of hardware processor cores. The processor to support a first machine instruction, the first machine instruction to include a vector register operand identifying a vector register which contains a plurality of data elements each used to identify a cache line. An execution of the first machine instruction by one of the plurality of hardware processor cores to cause a plurality of identified cache lines to be demoted, such that the demoted cache lines are moved from the first cache to the second cache.
-
公开(公告)号:US20190042295A1
公开(公告)日:2019-02-07
申请号:US16024554
申请日:2018-06-29
Applicant: Intel Corporation
Inventor: Liang Ma , John Joseph Browne , Xuebin Yang , Tomasz Kantecki , Andrew J. Herdrich
IPC: G06F9/455
Abstract: Particular embodiments described herein provide for an electronic device that can be configured to receive a request for a timestamp associated with a virtual machine, determine a current time from a timestamp counter, and subtract a timing compensation from the current time from the timestamp counter to create the timestamp, where the timing compensation includes an amount of time that execution of the virtual machine was suspended. In an example, a VM_EXIT instruction was used to suspend execution of the virtual machine and the timestamp counter was read before the VM_EXIT instruction was processed by a hypervisor.
-
公开(公告)号:US10178054B2
公开(公告)日:2019-01-08
申请号:US15088910
申请日:2016-04-01
Applicant: INTEL CORPORATION
Inventor: Stephen T. Palermo , Iosif Gasparakis , Scott P. Dubal , Kapil Sood , Trevor Cooper , Jr-Shian Tsai , Jesse C. Brandeburg , Andrew J. Herdrich , Edwin Verplanke
IPC: H04L12/861 , H04L12/715 , H04L12/931 , G06F15/173
Abstract: Methods and apparatus for accelerating VM-to-VM Network Traffic using CPU cache. A virtual queue manager (VQM) manages data that is to be kept in VM-VM shared data buffers in CPU cache. The VQM stores a list of VM-VM allow entries identifying data transfers between VMs that may use VM-VM cache “fast-path” forwarding. Packets are sent from VMs to the VQM for forwarding to destination VMs. Indicia in the packets (e.g., in a tag or header) is inspected to determine whether a packet is to be forwarded via a VM-VM cache fast path or be forwarded via a virtual switch. The VQM determines the VM data already in the CPU cache domain while concurrently coordinating with the data to and from the external shared memory, and also ensures data coherency between data kept in cache and that which is kept in shared memory.
-
50.
公开(公告)号:US20180373633A1
公开(公告)日:2018-12-27
申请号:US15634785
申请日:2017-06-27
Applicant: Intel Corporation
Inventor: Andrew J. Herdrich , Edwin Verplanke , Stephen R. Van Doren , Ravishankar Iyer , Eric R. Wehage , Rupin H. Vakharwala , Rajesh M. Sankaran , Jeffrey D. Chamberlain , Julius Mandelblat , Yen-Cheng Liu , Stephen T. Palermo , Tsung-Yuan C. Tai
IPC: G06F12/0811 , G06F13/42 , G06F9/455 , G06F9/50 , G06F12/1009 , G06F13/16
Abstract: Method and apparatus for per-agent control and quality of service of shared resources in a chip multiprocessor platform is described herein. One embodiment of a system includes: a plurality of core and non-core requestors of shared resources, the shared resources to be provided by one or more resource providers, each of the plurality of core and non-core requestors to be associated with a resource-monitoring tag and a resource-control tag; a mapping table to store the resource monitoring and control tags associated with each non-core requestor; and a tagging circuitry to receive a resource request sent from a non-core requestor to a resource provider, the tagging circuitry to responsively modify the resource request to include the resource-monitoring and resource-control tags associated with the non-core requestor in accordance to the mapping table and send the modified resource request to the resource provider.
-
-
-
-
-
-
-
-
-