Abstract:
Provided is a multi network and method to transmit packets. The multi network includes a mesh network, a tree network, and a network interface connected to the mesh network and the tree network and configured to transmit, through the mesh network and the tree network, a packet generated by a processing unit, of a processing system having plural processing units, at a starting point to a destination point for another processing unit of the processing system and configured to selectively inject the packet into one of the mesh network and the tree network to transmit the packet to the other processing unit.
Abstract:
A processor and corresponding method are described including cores having a thread set allocated based on a pre-set implementation order, and a controller configured to receive scheduling information determined based on an implementation pattern regarding the allocated thread set from one of the cores and transmit the scheduling information to another of the cores. The one of cores determines the scheduling information according to characteristics of an application when implementation of the thread set is completed. Each of the cores re-determines an implementation order regarding the allocated thread set based on the determined scheduling information.
Abstract:
A method of managing a cache includes storing first data of an upper level cache in a lower level cache, predicting a reuse distance level of second data having a same signature as the first data based on access information about the first data, and storing the second data in one of the lower level cache and a main memory based on the predicted reuse distance level of the second data.
Abstract:
A method of transceiving data over an on-chip network includes determining whether a packet is received by a first router among a plurality of first routers constituting a ring network; determining a transmission destination of the packet received by the first router; and transmitting the packet to a second router among a plurality of second routers constituting a bus network connected to the first router in response to the determined transmission destination being located on the bus network connected to the first router.
Abstract:
A tile-based rendering method includes performing binning of a current frame; generating an identification code of binning information or property information of a tile to be rendered in the current frame; comparing the identification code of the tile to be rendered to an identification code of a previous tile, wherein a location of the previous tile in a previous frame is the same as a location of the tile to be rendered in the current frame; and rendering the current frame by either re-using an image stored in a frame buffer or performing pixel processing of the tile to be rendered depending on a result of the comparing.
Abstract:
Embodiments include a processor capable of supporting multi-mode and corresponding methods. The processor includes front end units, a number of processing elements more than a number of the front end units; and a controller configured to determine if thread divergence occurs due to conditional branching. If there is thread divergence, the processor may set control information to control processing elements using currently activated front end units. If there is not, the processor may set control information to control processing elements using a currently activated front end unit.
Abstract:
A swizzle pattern generator is provided to reduce an overhead due to execution of a swizzle instruction in vector processing. The swizzle pattern generator is configured to provide swizzle patterns with respect to data sets of at least one vector register or vector processing unit. The swizzle pattern generator may be reconfigurable to generate various swizzle patterns for different vector operations.
Abstract:
Provided is a processor including a plurality of devices. The processor includes a source processing device configured to identify data to request from another device, and a destination processing device configured to, in response to a request for the identified data from the source processing device using credit-based flow control, transmit the identified data to the source processing device using the credit-based flow control. The source processing device includes a credit buffer used for the credit-based flow control, the credit buffer being allocable to include a cache region configured to cache the transmitted identified data received by the source processing device.
Abstract:
A cache memory apparatus includes a tag comparator configured to compare upper bits of each of pieces of tag data included in a set indicated by a set address that is received with upper bits of a tag address that is received, compare other bits of each of the pieces of the tag data with other bits of the tag address, and determine whether there is a cache hit or a cache miss based on results of the comparisons, and an update controller configured to, in response to the cache miss being determined, determine, as an update candidate, a piece among pieces of cache data included in the set and corresponding to the pieces of the tag data, based on the result of the comparison of the upper bits of each of the pieces of the tag data and the upper bits of the tag address, and update the update candidate with new data.
Abstract:
A method and corresponding apparatus are configured to generate a mipmap are configured to allocate a mipmap status register of a mipmap level generated with respect to a texture, receive a request for the texture, and calculate a mipmap level with respect to the texture. The method and corresponding apparatus are also configured to determine whether a mipmap of the calculated mipmap level exists using the mipmap status register and outputting a result indicative thereof, and determine whether to generate the mipmap of the mipmap level based on the result.