Abstract:
A database system may include a memory device that includes a least a portion to serve as a buffer cache and an array of persistent storage devices configured to store data of a database. The database system may monitor a frequency of data value associated with a first portion of data of the database stored in the buffer cache. The database system may maintain the first portion of data in the buffer cache in response to the frequency of data value associated with the first portion of data being greater than a frequency of data value associated with at least a portion of the data of the database stored in the array of persistent storage devices.
Abstract:
Technology for an apparatus is described. The apparatus can include a plurality of cache memories and a cache controller. The cache controller can allocate a cache entry to store data across the plurality of cache memories. The cache entry can include a value in a metadata field indicating an interleave policy. The cache controller can selectively assign the interleave policy to be applied based on a type of data stored in the plurality of cache memories.
Abstract:
A method for using an access triggered architecture for a computer implemented application is provided. The method receives a set of data at a designated functional block associated with a system memory location; performs an operation at the designated functional block, using the set of data, to generate a result, wherein the operation is performed each time information is received at the designated functional block; and returns the generated result to the system memory location.
Abstract:
A device supporting big data in a process plant includes an interface to a communications network, a cache configured to store data observed by the device, and a multi-processing element processor to cause the data to be cached and transmitted (e.g., streamed) for historization at a unitary, logical centralized data storage area. The data storage area stores multiple types of process control or plant data using a common format. The device time-stamps the cached data, and, in some cases, all data that is generated or created by or received at the device may be cached and/or streamed. The device may be a field device, a controller, an input/output device, a network management device, a user interface device, or a historian device, and the device may be a node of a network supporting big data in the process plant. Multiple devices in the network may support layered or leveled caching of data.
Abstract:
A cache controller to configure a portion of a first memory as cache for a second memory responsive to an indicator of locality of memory access requests to the second memory. The indicator of locality determines a probability that a location of a memory access request to the second memory is predictable based upon at least one previous memory access request. The cache controller may determine a size of the cache based on a value of the indicator of locality or modify the size of the cache in response to changes in the value of the indicator of locality.
Abstract:
An approach for management of memory in a programmable integrated circuit (IC) includes configuring a memory map of the programmable IC with an association of a first subset of addresses of memory address space of the programmable IC and physical memory of the programmable IC. The memory map is further configured with an association of a second subset of addresses of the memory address space and a virtual memory block. At least a portion of a cache memory of the programmable IC is locked to the second subset of addresses.
Abstract:
A set of techniques is described for enabling profile-driven compiler optimization based on cloud-specific information. A service provider may host applications on behalf of multiple users by providing a set of shared resources in a multi-tenant computing environment, where the resources are shared by the various applications hosted thereon. The service provider can collect runtime conditions, resource contention data and other environment-specific information of the shared resources. This gathered information can be provided a profile-driven compiler. The profile-driven compiler can use the information to recompile the source code of the application to produce an optimized version the application that is specifically tuned to run on the shared resources. The running version of the application can then be replaced by the optimized version.
Abstract:
Using an “optimized” test case for testing hardware and/or software of a computer. The optimized test case is designed to be run on a data storage device including multiple read locations and multiple write locations. Initialization data is written, on the data storage device, only to the write locations of the data storage device. The optimized test case is run on the data storage device in a manner so that the optimized test case will only write data to each write location after that write location has had initialization data written to that write location. The optimized test case defines read locations and write locations so that, during running of the optimized test case, all read locations which are also write locations will be written by a write instruction of the test case before being read by a read instruction of the test case.
Abstract:
A memory system comprises a master control module, a memory control module, a nonvolatile memory and a cache, wherein the memory control module is connected with the master control module, the nonvolatile memory and the cache are respectively connected with the memory control module; and the memory control module is configured to, when the master control module sends a write command for the nonvolatile memory, store data to be written in the nonvolatile memory in the cache according to the write command, and release the cache used for storing the data to be written in the nonvolatile memory after finish of the write operation to the nonvolatile memory.
Abstract:
One embodiment of the present invention sets forth a technique for addressing data in a hierarchical graphics processing unit cluster. A hierarchical address is constructed based on the location of a storage circuit where a target unit of data resides. The hierarchical address comprises a level field indicating a hierarchical level for the unit of data and a node identifier that indicates which GPU within the GPU cluster currently stores the unit of data. The hierarchical address may further comprise one or more identifiers that indicate which storage circuit in a particular hierarchical level currently stores the unit of data. The hierarchical address is constructed and interpreted based on the level field. The technique advantageously enables programs executing within the GPU cluster to efficiently access data residing in other GPUs using the hierarchical address.