Abstract:
A release orchestration mechanism for cloud services. According to some implementations, while an app aware proxy routes production traffic to a first application (app) version that runs in a plurality of container orchestration system (COS) pods having first app version containers, configuration information is received. When a threshold number of the COS pods having the second app version containers are live, a validation of the second app version is caused. Then a transition to sending production traffic to the second app version containers is performed. After causing the transition, timers are started based on a time period indicated in the configuration information and the first app version containers are instructed to gracefully shut down. Based on expiration of the timers, any of the COS pods having the first app version containers that are not yet shut down are forced to shut down.
Abstract:
Systems and techniques for utilizing resource aware queues and/or service sharing in a multi-server environment. Requests directed to an application server are received into one or more queues in front of the application server. An acknowledgement of the requests is provided in response to the requests being received by the one or more queues. Metadata associated with the requests is utilized to manage the one or more queues. The requests from the one or more queues are processed based on the metadata.
Abstract:
Techniques and architectures to provide multiple application servers each having at least one connection agent and at least one enabler agent and at least one database organized as multiple nodes. An error condition between a first connection agent and a first database node is detected. The first connection agent indicates that the first database node is unavailable to stop attempts to access the first database node by a first application server. A first enabler agent periodically checks availability of the first database node. The first database node is indicated as available in response to a pre-selected number of successful attempts to contact the first database node with the first enabler agent. The first database node is accessed with the application server in response to the indication that the first database node is available.
Abstract:
Techniques are disclosed relating to an in-memory cache. In some embodiments, in response to determining that data for a requested entry is not present in the cache (e.g., because it has been evicted), a computing system is configured to invoke cached program code associated with the entry. In some embodiments, the computing system is configured to provide data generated by the program code in response to requests that indicate the entry. In some embodiments, the computing system is configured to store the generated data in the cache. In various embodiments, this may avoid cache misses and provide configurability in responding to requests to access the cache.
Abstract:
Techniques are disclosed relating to an in-memory, software-managed cache configured to store web application data. In some embodiments, operations to cache data specify a visibility parameter for the data, among a plurality of namespaces. In some embodiments, requests to access cached data are checked, based on a request's namespace and the visibility parameter for the cached data, to determine whether they are allowed to proceed. In some embodiments, this may facilitate caching data using shared computing systems and data structures while maintaining configurable privacy for cached data.
Abstract:
In accordance with embodiments, there are provided mechanisms and methods for facilitating dynamic workload scheduling and routing of message queues for fair management of the resources for application servers in an on-demand services environment. In one embodiment and by way of example, a method includes detecting an organization of a plurality of organization that is starving for resources. The organization may be seeking performance of a job request at a computing system within a multi-tenant database system. The method may further include consulting, based on a routing policy, a routing table for a plurality of queues available for processing the job request, selecting a queue of the plurality of queues for the organization based on a fair usage analysis obtained from the routing policy, and routing the job request to the selected queue.
Abstract:
Systems and methods are provided for receiving a request at a database system that includes a primary node and one or more standby nodes. When the request is determined to be a read-only request and the previous transaction has been replicated to the standby node, the request may be transmitted to one of the one or more standby nodes to perform the read-only operation to retrieve data of the request and transmitting the retrieved data. When the request is determined to be a read-only request and the previous transaction has not been replicated to the standby node, the read-only operation is performed to retrieve data of the request from the primary node and transmitting the retrieved data.
Abstract:
According to some implementations, while a proxy routes production traffic to a first application (app) version that runs in a plurality of container orchestration system (cos) pods having first app version containers, configuration information is received including an identification of a second app version container image for a second app version. The second app version is an updated version of the first app version. Cos pods having second app version containers are brought up based on the second app version container image identified in the configuration information. Test and/or warmup traffic is caused to be routed to the second app version containers. Responsive to an indication regarding the routing of the test and/or warmup traffic to the second app version, causing a transition to sending production traffic to the second app version containers instead of to the first app version.
Abstract:
Techniques and structures to prevent exhaustion of a database connection pool, including retrieving data from the database connection pool, monitoring the data to determine whether the connection pool is at risk of an exhaustion condition, analyzing the data to determine whether one or more clients accessing the database connection pool are offenders upon determining that the connection pool is at risk and throttling access to the one or more clients accessing the database connection pool upon determining the one or more clients to be offenders.
Abstract:
Systems and techniques for utilizing resource aware queues and/or service sharing in a multi-server environment. According to an example, an application server employs a traffic light metaphor to represent a utilization level of resources of the application server by associating a traffic light with each resource. A mapping is maintained that associates service requests with corresponding sets of affected traffic lights. A deferred queue is maintained for each traffic light to facilitate throttling of service requests directed to the application server that involve a resource that is under pressure. Responsive to receiving a service request directed to the application server, the service request is added directly or indirectly to one of multiple queues maintained in front of the application server based on a priority associated with the service request. Service requests are serviced from the queues in accordance with a priority associated with the queues.