Abstract:
The fail-over computer cluster enables multiple computing devices to operate using adaptive quorum rules to dictate which nodes are in the fail-over cluster at any given time. The adaptive quorum rules provide requirements for communications between nodes and connections with voting file systems. The adaptive quorum rules include particular recovery rules for unplanned changes in node configuration, such as due to a disruptive event. Such recovery quorum rules enable the fail-over cluster to continuing to operate with various changed configurations of its node members as a result of the disruptive event. In the changed configuration, access to voting file systems may not be required for a majority-group subset of nodes. If no majority-group subset remains, nodes may need direct or indirect access to voting file systems.
Abstract:
The fail-over computer cluster enables multiple computing devices to operate using adaptive quorum rules to dictate which nodes are in the fail-over cluster at any given time. The adaptive quorum rules provide requirements for communications between nodes and connections with voting file systems. The adaptive quorum rules include particular recovery rules for unplanned changes in node configuration, such as due to a disruptive event. Such recovery quorum rules enable the fail-over cluster to continuing to operate with various changed configurations of its node members as a result of the disruptive event. In the changed configuration, access to voting file systems may not be required for a majority-group subset of nodes. If no majority-group subset remains, nodes may need direct or indirect access to voting file systems.
Abstract:
The embodiments disclosed herein relate to using SNMP functionality for fast and accurate detection of node failure in a computer cluster. When cluster nodes are connected via a switch that provides SNMP capability, a trap mechanism can be used to immediately and proactively notify cluster nodes that a particular cluster node is disconnected from the network, potentially even before the remaining cluster nodes have noticed a lack of response. In addition, when a cluster node notices a lack of response from a particular cluster node, the switch can be queried for link status and traffic information that can be used for fast determination of node status.
Abstract:
The fail-over computer cluster enables multiple computing devices to operate using adaptive quorum rules to dictate which nodes are in the fail-over cluster at any given time. The adaptive quorum rules provide requirements for communications between nodes and connections with voting file systems. The adaptive quorum rules include particular recovery rules for unplanned changes in node configuration, such as due to a disruptive event. Such recovery quorum rules enable the fail-over cluster to continuing to operate with various changed configurations of its node members as a result of the disruptive event. In the changed configuration, access to voting file systems may not be required for a majority-group subset of nodes. If no majority-group subset remains, nodes may need direct or indirect access to voting file systems.
Abstract:
The fail-over computer cluster enables multiple computing devices to operate using adaptive quorum rules to dictate which nodes are in the fail-over cluster at any given time. The adaptive quorum rules provide requirements for communications between nodes and connections with voting file systems. The adaptive quorum rules include particular recovery rules for unplanned changes in node configuration, such as due to a disruptive event. Such recovery quorum rules enable the fail-over cluster to continuing to operate with various changed configurations of its node members as a result of the disruptive event. In the changed configuration, access to voting file systems may not be required for a majority-group subset of nodes. If no majority-group subset remains, nodes may need direct or indirect access to voting file systems.
Abstract:
A cluster manager manages copies of a mid-tier database as a mid-tier database cluster. The cluster manager may concurrently manage a backend database system. The cluster manager is configured to monitor for and react to failures of mid-tier database nodes. The cluster manager may react to a mid-tier database failure by, for example, assigning a new active node, creating a new standby node, creating new copies of the mid-tier databases, implementing new replication or backup schemes, reassigning the node's virtual address to another node, or relocating applications that were directly linked to the mid-tier database to another host. Each node or an associated agent may configure the cluster manager during initialization, based on common cluster configuration information. Each copy of the mid-tier database may be, for example, a memory resident database. Thus, a node must reload the entire database into memory to recover a copy of the database.
Abstract:
The embodiments disclosed herein relate to using SNMP functionality for fast and accurate detection of node failure in a computer cluster. When cluster nodes are connected via a switch that provides SNMP capability, a trap mechanism can be used to immediately and proactively notify cluster nodes that a particular cluster node is disconnected from the network, potentially even before the remaining cluster nodes have noticed a lack of response. In addition, when a cluster node notices a lack of response from a particular cluster node, the switch can be queried for link status and traffic information that can be used for fast determination of node status.
Abstract:
Described are improved systems, computer program products, and methods for a new local rolling online patching solution to a database server with minimized disruption to other instances during patching. For a one-node database server, it achieves single rolling online patching by starting a new instance of the same database from the newly patched home before shutting down the collocated instance running out of the old home.
Abstract:
The embodiments disclosed herein relate to using SNMP functionality for fast and accurate detection of node failure in a computer cluster. When cluster nodes are connected via a switch that provides SNMP capability, a trap mechanism can be used to immediately and proactively notify cluster nodes that a particular cluster node is disconnected from the network, potentially even before the remaining cluster nodes have noticed a lack of response. In addition, when a cluster node notices a lack of response from a particular cluster node, the switch can be queried for link status and traffic information that can be used for fast determination of node status.
Abstract:
A cluster manager manages copies of a mid-tier database as a mid-tier database cluster. The cluster manager may concurrently manage a backend database system. The cluster manager is configured to monitor for and react to failures of mid-tier database nodes. The cluster manager may react to a mid-tier database failure by, for example, assigning a new active node, creating a new standby node, creating new copies of the mid-tier databases, implementing new replication or backup schemes, reassigning the node's virtual address to another node, or relocating applications that were directly linked to the mid-tier database to another host. Each node or an associated agent may configure the cluster manager during initialization, based on common cluster configuration information. Each copy of the mid-tier database may be, for example, a memory resident database. Thus, a node must reload the entire database into memory to recover a copy of the database.