Apparatus and method for acceleration data structure refit

    公开(公告)号:US12229870B2

    公开(公告)日:2025-02-18

    申请号:US17982766

    申请日:2022-11-08

    Abstract: Apparatus and method for acceleration data structure refit. For example, one embodiment of an apparatus comprises: a ray generator to generate a plurality of rays in a first graphics scene; a hierarchical acceleration data structure generator to construct an acceleration data structure comprising a plurality of hierarchically arranged nodes including inner nodes and leaf nodes stored in a memory in a depth-first search (DFS) order; traversal hardware logic to traverse one or more of the rays through the acceleration data structure; intersection hardware logic to determine intersections between the one or more rays and one or more primitives within the hierarchical acceleration data structure; a node refit unit comprising circuitry and/or logic to read consecutively through at least the inner nodes in the memory in reverse DFS order to perform a bottom-up refit operation on the hierarchical acceleration data structure.

    Tile assignment to processing cores within a graphics processing unit

    公开(公告)号:US12229851B2

    公开(公告)日:2025-02-18

    申请号:US18385265

    申请日:2023-10-30

    Abstract: A graphics processing unit configured to process graphics data using a rendering space which is sub-divided into a plurality of tiles, the graphics processing unit comprising: a plurality of processing cores configured to render graphics data; cost indication logic configured to obtain a cost indication for each of a plurality of sets of one or more tiles of the rendering space, wherein the cost indication for a set of one or more tiles is suggestive of a cost of processing the set of one or more tiles; similarity indication logic configured to obtain similarity indications between sets of one or more tiles of the rendering space, wherein the similarity indication between two sets of one or more tiles is indicative of a level of similarity between the two sets of tiles according to at least one processing metric; and scheduling logic configured to assign the sets of one or more tiles to the processing cores for rendering in dependence on the cost indications and the similarity indications.

    Virtualization and multi-tenancy support in graphics processors

    公开(公告)号:US12229581B2

    公开(公告)日:2025-02-18

    申请号:US18239489

    申请日:2023-08-29

    Abstract: Graphics processing systems and methods are described. A graphics processing apparatus may comprise one or more graphics processing engines, a memory, a memory management unit (MMU) including a GPU second level page table and GPU dirty bit tracking, and a provisioning agent to receive a request from a virtual machine monitor (VMM) to provision a subcluster of graphics processing apparatuses, the subcluster including a plurality of graphics processing engines from a plurality of graphics processing apparatuses connected using a scale-up fabric, provision the scale-up fabric to route data within the subcluster of graphics processing apparatuses, and provision a plurality of resources on the graphics processing apparatus for the subcluster based on the request from the VMM.

    SYSTEM AND METHOD FOR ACCELERATED RAY TRACING WITH ASYNCHRONOUS OPERATION AND RAY TRANSFORMATION

    公开(公告)号:US20250054222A1

    公开(公告)日:2025-02-13

    申请号:US18779641

    申请日:2024-07-22

    Inventor: Mark Evan Cerny

    Abstract: A graphics processing unit (GPU) includes one or more processor cores adapted to execute a software-implemented shader program, and one or more hardware-implemented ray tracing units (RTU) adapted to traverse an acceleration structure to calculate intersections of rays with bounding volumes and graphics primitives asynchronously with shader operation. The RTU implements traversal logic to traverse the acceleration structure including transformation of rays as needed to account for variations in coordinate space between levels, stack management, and other tasks to relieve burden on the shader, communicating intersections to the shader which then calculates whether the intersection hit a transparent or opaque portion of the object intersected. Thus, one or more processing cores within the GPU perform accelerated ray tracing by offloading aspects of processing to the RTU, which traverses the acceleration structure within which the 3D environment is represented.

    DATA TRANSMISSION METHOD, APPARATUS AND SYSTEM, DEVICE AND STORAGE MEDIUM

    公开(公告)号:US20250054109A1

    公开(公告)日:2025-02-13

    申请号:US18929660

    申请日:2024-10-29

    Abstract: A data transmission method includes: acquiring data of a display image layer and data of a transparency layer and determining, based on the data of the display image layer and the data of the transparency layer, multiple sets of data to be superimposed; constructing a target layer based on the display image layer and the transparency layer for each set of data to be superimposed; determining first placement data and second placement data of the corresponding target layer based on the data of the display image layer and the data of the transparency layer for each set of data to be superimposed; placing the first placement data to a first position of the corresponding target layer and placing the second placement data to a second position of the corresponding target layer, to obtain data of the corresponding target layer; and transmitting data of multiple target layers.

    Variable width interleaved coding for graphics processing

    公开(公告)号:US12223682B2

    公开(公告)日:2025-02-11

    申请号:US17357038

    申请日:2021-06-24

    Abstract: Variable width interleaved coding for graphics processing is described. An example of an apparatus includes one or more processors including a graphic processor; and memory for storage of data including data for graphics processing, wherein the graphics processor includes an encoder pipeline to provide variable width interleaved coding and a decoder pipeline to decode the variable width interleaved coding, and wherein the encoder pipeline is to receive a plurality of bitstreams from workgroups; perform parallel entropy encoding on the bitstreams to generate a plurality of encoded bitstreams for each of the workgroups; perform variable interleaving of the bitstreams for each workgroup based at least in part on data requirements for decoding received from the decoder pipeline; and compact outputs for each of the workgroups into a contiguous stream of interleaved data.

Patent Agency Ranking