-
公开(公告)号:US20200026514A1
公开(公告)日:2020-01-23
申请号:US16526147
申请日:2019-07-30
Applicant: Intel Corporation
Inventor: Abhishek R. Appu , Altug Koker , Joydeep Ray , Kamal Sinha , Kiran C. Veernapu , Subramaniam Maiyuran , Prasoonkumar Surti , Guei-Yuan Lueh , David Puffer , Supratim Pal , Eric J. Hoekstra , Travis T. Schluessler , Linda L. Hurd
Abstract: In an example, an apparatus comprises a plurality of execution units, and a first general register file (GRF) communicatively couple to the plurality of execution units, wherein the first GRF is shared by the plurality of execution units. Other embodiments are also disclosed and claimed.
-
82.
公开(公告)号:US20200005515A1
公开(公告)日:2020-01-02
申请号:US16023647
申请日:2018-06-29
Applicant: Intel Corporation
Inventor: Pradeep Ramani , Karthik Vaidyanathan , Prasoonkumar Surti
Abstract: A mechanism is described for facilitating efficient prediction of most commonly occurring values in data blocks in computing environments. An apparatus of embodiments, as described herein, includes one or more processors to perform parallel calculations on values associated with multiple sub-blocks of a data block, and predict, based on the parallel calculations, a most commonly-occurring value in the data block. The apparatus if further to classify the most commonly-occurring value as a mode value for one or more data types to be used with one or more applications.
-
公开(公告)号:US10521349B2
公开(公告)日:2019-12-31
申请号:US16277267
申请日:2019-02-15
Applicant: Intel Corporation
Inventor: Chandrasekaran Sakthivel , Prasoonkumar Surti , John C. Weast , Sara S. Baghsorkhi , Justin E. Gottschlich , Abhishek R. Appu , Nicolas C. Galoppo Von Borries , Joydeep Ray , Narayan Srinivasa , Feng Chen , Ben J. Ashbaugh , Rajkishore Barik , Tsung-Han Lin , Kamal Sinha , Eriko Nurvitadhi , Balaji Vembu , Altug Koker
IPC: G06F12/0837 , G06N3/08 , G06N20/00 , G06T1/20 , G06F12/0815 , G06N3/04 , G06N3/063
Abstract: In an example, an apparatus comprises a plurality of processing unit cores, a plurality of cache memory modules associated with the plurality of processing unit cores, and a machine learning model communicatively coupled to the plurality of processing unit cores, wherein the plurality of cache memory modules share cache coherency data with the machine learning model. Other embodiments are also disclosed and claimed.
-
公开(公告)号:US10452397B2
公开(公告)日:2019-10-22
申请号:US15477022
申请日:2017-04-01
Applicant: Intel Corporation
Inventor: Joydeep Ray , Altug Koker , Balaji Vembu , Abhishek R. Appu , Kamal Sinha , Prasoonkumar Surti , Kiran C. Veernapu
Abstract: Methods and apparatus relating to techniques for avoiding cache lookup for cold cache. In an example, an apparatus comprises logic, at least partially comprising hardware logic, to determine a first number of threads to be scheduled for each context of a plurality of contexts in a multi-context processing system, allocate a second number of streaming multiprocessors (SMs) to the respective plurality of contexts, and dispatch threads from the plurality of contexts only to the streaming multiprocessor(s) allocated to the respective plurality of contexts. Other embodiments are also disclosed and claimed.
-
公开(公告)号:US10403024B2
公开(公告)日:2019-09-03
申请号:US15988157
申请日:2018-05-24
Applicant: Intel Corporation
Inventor: Bimal Poddar , Prasoonkumar Surti , Rahul P. Sathe
Abstract: Embodiments provide for a graphics processing apparatus comprising render logic to detect rendering operations that will result in framebuffer having the same data as the initial clear color value and morphing such rendering operations to optimizations that are typically done for initial clearing of the framebuffer.
-
公开(公告)号:US20190265765A1
公开(公告)日:2019-08-29
申请号:US16286166
申请日:2019-02-26
Applicant: Intel Corporation
Inventor: Sanjeev S. Jahagirdar , Eric J. Asperheim , Subramaniam Maiyuran , Abhishek R. Appu , Joydeep Ray , Altug Koker , Prasoonkumar Surti , Kiran C. Veernapu
Abstract: Methods and apparatus relating to techniques for dynamic control of liquid cooling pumps to provide thermal cooling uniformity are described. In an embodiment, modification is made to operation of one or more of: one or more cooling pumps or one or more fans, based at least in part on comparison of one or more detected temperature or noise values at one or more components of a processor with one or more corresponding threshold values. The processor may include the logic that causes the modification and one or more sensors. The sensors are thermally or acoustically coupled to the one or more components of the processor to determine the detected temperature or noise values. Other embodiments are also disclosed and claimed.
-
公开(公告)号:US20190251033A1
公开(公告)日:2019-08-15
申请号:US16277114
申请日:2019-02-15
Applicant: Intel Corporation
Inventor: Abhishek R. Appu , Altug Koker , Joydeep Ray , Prasoonkumar Surti , Kamal Sinha , Kiran C. Veernapu , Balaji Vembu
IPC: G06F12/0888 , G06F13/40 , G06T1/60 , G06T1/20 , G06F13/42
CPC classification number: G06F12/0888 , G06F12/0895 , G06F13/4022 , G06F13/4282 , G06F2212/1024 , G06F2212/1028 , G06F2212/6032 , G06F2213/0026 , G06T1/20 , G06T1/60
Abstract: Methods and apparatus relating to techniques for avoiding cache lookup for cold cache. In an example, an apparatus comprises logic, at least partially comprising hardware logic, to receive, in a read/modify/write (RMW) pipeline, a cache access request from a requestor, wherein the cache request comprises a cache set identifier associated with requested data in the cache set, determine whether the cache set associated with the cache set identifier is in an inaccessible invalid state, and in response to a determination that the cache set is in an inaccessible state or an invalid state, to terminate the cache access request. Other embodiments are also disclosed and claimed.
-
公开(公告)号:US20190236026A1
公开(公告)日:2019-08-01
申请号:US16287781
申请日:2019-02-27
Applicant: Intel Corporation
Inventor: Prasoonkumar Surti , Abhishek R. Appu , Kiran C. Veernapu
IPC: G06F12/1045 , G06T1/20 , G06T1/60 , G06F12/1027 , G06F12/1009
CPC classification number: G06F12/1063 , G06F12/1009 , G06F12/1027 , G06F2212/455 , G06F2212/657 , G06F2212/68 , G06T1/20 , G06T1/60
Abstract: One embodiment provides for a graphics processor comprising a translation lookaside buffer (TLB) to cache a first page table entry for a virtual to physical address mapping for use by the graphics processor, the first page table entry to indicate that a first virtual page is cleared to a clear color and a graphics pipeline to bypass a memory access for the first virtual page based on the first page table entry, wherein the graphics pipeline is to read a field in the first page table entry to determine a value of the clear color.
-
公开(公告)号:US20190206090A1
公开(公告)日:2019-07-04
申请号:US15859408
申请日:2017-12-30
Applicant: Intel Corporation
Inventor: Joydeep Ray , Ben Ashbaugh , Prasoonkumar Surti , Pradeep Ramani , Rama Harihara , Jerin C. Justin , Jing Huang , Xiaoming Cui , Timothy B. Costa , Ting Gong , Elmoustapha Ould-Ahmed-Vall , Kumar Balasubramanian , Anil Thomas , Oguz H. Elibol , Jayaram Bobba , Guozhong Zhuang , Bhavani Subramanian , Gokce Keskin , Chandrasekaran Sakthivel , Rajesh Poornachandran
CPC classification number: G06T9/002 , G06F12/023 , G06F2212/302 , G06F2212/401 , G06T15/005
Abstract: Embodiments are generally directed to compression in machine learning and deep learning processing. An embodiment of an apparatus for compression of untyped data includes a graphical processing unit (GPU) including a data compression pipeline, the data compression pipeline including a data port coupled with one or more shader cores, wherein the data port is to allow transfer of untyped data without format conversion, and a 3D compression/decompression unit to provide for compression of untyped data to be stored to a memory subsystem and decompression of untyped data from the memory subsystem.
-
公开(公告)号:US10332320B2
公开(公告)日:2019-06-25
申请号:US15488914
申请日:2017-04-17
Applicant: Intel Corporation
Inventor: Barath Lakshamanan , Linda L. Hurd , Ben J. Ashbaugh , Elmoustapha Ould-Ahmed-Vall , Liwei Ma , Jingyi Jin , Justin E. Gottschlich , Chandrasekaran Sakthivel , Michael S. Strickland , Brian T. Lewis , Lindsey Kuper , Altug Koker , Abhishek R. Appu , Prasoonkumar Surti , Joydeep Ray , Balaji Vembu , Javier S. Turek , Naila Farooqui
IPC: G01C22/00 , G07C5/00 , G05D1/00 , G01C21/34 , G08G1/01 , H04W28/08 , G06N20/00 , G06F9/50 , G08G1/052 , G01S19/13 , G05D1/02 , H04L29/08 , H04L12/26
Abstract: One embodiment provides for a computing device within an autonomous vehicle, the compute device comprising a wireless network device to enable a wireless data connection with an autonomous vehicle network, a set of multiple processors including a general-purpose processor and a general-purpose graphics processor, the set of multiple processors to execute a compute manager to manage execution of compute workloads associated with the autonomous vehicle, the compute workload associated with autonomous operations of the autonomous vehicle, and offload logic configured to execute on the set of multiple processors, the offload logic to determine to offload one or more of the compute workloads to one or more autonomous vehicles within range of the wireless network device.
-
-
-
-
-
-
-
-
-