Abstract:
In a large-scale transaction such as the bulk loading of new records into an ordered, distributed database, a transaction limit such as an insert limit may be chosen, partitions on overfull storage servers may be designated to be moved to underfull storage servers, and the move assignments may be based, at least in part on the degree to which a storage server is underfull and the move and insertion costs of the partitions to be moved.
Abstract:
In one embodiment, a first one of a plurality of message processors receives a request with respect to a topic, and determines whether the first message processor itself is currently responsible for the topic. If so, then the first message processor services the request. If not, then the first message processor determines whether any of the other ones of the message processors is currently responsible for the topic. If so, then the request is redirected to another message processor to be serviced. If not, then the first message processor selects one of the message processors to be responsible for the topic and to service the request.
Abstract:
Methods, systems and computer program products are provided for managing at least one cache for a database. In one method, an operation record is inserted in an operations log for each write operation to the database. The operation record includes a Log Sequence Number (LSN). At least one cache is updated using operation records in the operations log. To update a cache, unapplied operations in the operations log for the cache are identified using an applied LSN associated with the cache. The applied LSN denotes an LSN of a last operation applied to the cache. The unapplied operations are then applied to the cache and the applied LSN is updated with an LSN corresponding to the last unapplied operation of the unapplied operations that were applied to the cache during updating the cache.
Abstract:
Methods and apparatus are described for partitioning native tables in a database cluster into logical tables. Each logical table is mapped into a unique portion of the native table by an intermediary server. Clients access a logical table as an ordinary, full-fledged database table through the intermediary server, which translates queries on the logical table into queries on the corresponding portion of the native table. The mapping may use the application name, logical table name, and a version number to create a native table key for each key in the logical table. A data structure storing these mappings may be stored at the intermediary server or in a native table in the database. This approach affords clients quick and flexible access to the database with better data integrity and security than native tables allow.
Abstract:
In one embodiment, for each one of one or more content producers, access a content-producing rate at which the content producer produces one or more items of content; and for each one of one or more content consumers, access a content-consuming rate at which the content consumer consumes one or more items of content. For each one of a plurality of consumer-producer pairs, wherein the consumer-producer pair comprises one of the content consumers and one of the content producers, and the content consumer follows the content producer, select between a push strategy and a pull strategy for delivering one or more items of content from the content producer to the content consumer based on the content-consuming rate of the content consumer and the content-producing rate of the content producer.
Abstract:
In one embodiment, for each one of one or more content producers, access a content-producing rate at which the content producer produces one or more items of content; and for each one of one or more content consumers, access a content-consuming rate at which the content consumer consumes one or more items of content. For each one of a plurality of consumer-producer pairs, wherein the consumer-producer pair comprises one of the content consumers and one of the content producers, and the content consumer follows the content producer, select between a push strategy and a pull strategy for delivering one or more items of content from the content producer to the content consumer based on the content-consuming rate of the content consumer and the content-producing rate of the content producer.
Abstract:
In one embodiment, a first one of a plurality of message processors receives a request with respect to a topic, and determines whether the first message processor itself is currently responsible for the topic. If so, then the first message processor services the request. If not, then the first message processor determines whether any of the other ones of the message processors is currently responsible for the topic. If so, then the request is redirected to another message processor to be serviced. If not, then the first message processor selects one of the message processors to be responsible for the topic and to service the request.
Abstract:
In one embodiment, for each one of one or more content producers, access a content-producing rate at which the content producer produces one or more items of content; and for each one of one or more content consumers, access a content-consuming rate at which the content consumer consumes one or more items of content. For each one of a plurality of consumer-producer pairs, wherein the consumer-producer pair comprises one of the content consumers and one of the content producers, and the content consumer follows the content producer, select between a push strategy and a pull strategy for delivering one or more items of content from the content producer to the content consumer based on the content-consuming rate of the content consumer and the content-producing rate of the content producer.
Abstract:
A system includes a leader server capable of communicating with a plurality of follower servers and a network. The leader server is configured to determine when a message received from a client connected to the network has been committed to main memory in a majority of the leader and follower servers. The leader server and each of the follower servers that committed the message to main memory retain the message in main memory until requested to remove the message from main memory by a delivery agent or an archiver.
Abstract:
A system includes a leader server capable of communicating with a plurality of follower servers and a network. The leader server is configured to determine when a message received from a client connected to the network has been committed to main memory in a majority of the leader and follower servers. The leader server and each of the follower servers that committed the message to main memory retain the message in main memory until requested to remove the message from main memory by a delivery agent or an archiver.