Glossary #
- active/active, active/passive #
A concept of how services are running on nodes. An active-passive scenario means that one or more services are running on the active node and the passive node waits for the active node to fail. Active-active means that each node is active and passive at the same time. For example, it has some services running, but can take over other services from the other node. Compare with primary/secondary and dual-primary in DRBD speak.
- arbitrator #
Additional instance in a Geo cluster that helps to reach consensus about decisions such as failover of resources across sites. Arbitrators are single machines that run one or more booth instances in a special mode.
In regular clusters, “arbitrator” can also refer to QNetd. QNetd can be configured along with QDevice to participate in quorum decisions in clusters with an even number of nodes.
- AutoYaST #
AutoYaST is a system for installing one or more SUSE Linux Enterprise systems automatically and without user intervention.
- bindnetaddr (bind network address) #
The network address the Corosync executive should bind to.
- booth #
The instance that manages the failover process between the sites of a Geo cluster. It aims to get multi-site resources active on one and only one site. This is achieved by using so-called tickets that are treated as failover domain between cluster sites, in case a site should be down.
- boothd (booth daemon) #
Each of the participating clusters and arbitrators in a Geo cluster runs a service, the
boothd
. It connects to the booth daemons running at the other sites and exchanges connectivity details.- CIB (cluster information base) #
A representation of the whole cluster configuration and status (cluster options, nodes, resources, constraints and the relationship to each other). It is written in XML and resides in memory. A primary CIB is kept and maintained on the DC (designated coordinator) and replicated to the other nodes. Normal read and write operations on the CIB are serialized through the primary CIB.
- clone #
A clone can refer to an identical copy of an existing node. Cloning nodes can make deploying multiple nodes simpler.
In the context of a cluster resource, a clone is a resource that can be active on multiple nodes. Any resource can be cloned, provided the respective resource agent supports it. A promotable clone (also known as a multi-state resource) is a special type of clone resource that can be promoted.
- cluster #
A high-performance cluster is a group of computers (real or virtual) sharing the application load to achieve faster results. A high-availability cluster is designed primarily to secure the highest possible availability of services.
- Cluster logical volume manager (Cluster LVM) #
The term
Cluster LVM
indicates that LVM is being used in a cluster environment. This needs some configuration adjustments to protect the LVM metadata on shared storage.- cluster partition #
Whenever communication fails between one or more nodes and the rest of the cluster, a cluster partition occurs. The nodes of a cluster are split into partitions but still active. They can only communicate with nodes in the same partition and are unaware of the separated nodes. As the loss of the nodes on the other partition cannot be confirmed, a split-brain scenario develops (see also split-brain).
- cluster site #
In Geo clustering, a cluster site (or just “site”) is a group of nodes in the same physical location, managed by booth.
- cluster stack #
The ensemble of software technologies and components that compose a cluster.
- colocation constraint #
Colocation constraints tell the cluster which resources may or may not run together on a node. See also resource constraint.
- concurrency violation #
A resource that should be running on only one node in the cluster is running on several nodes.
- conntrack tools #
Allow interaction with the in-kernel connection tracking system for enabling stateful packet inspection for iptables. Used by SUSE Linux Enterprise High Availability to synchronize the connection status between cluster nodes.
- Corosync #
Corosync provides reliable messaging, membership and quorum information about the cluster. This is handled by the Corosync Cluster Engine, a group communication system.
- CRM (cluster resource manager) #
The management entity responsible for coordinating all non-local interactions in a High Availability cluster. SUSE Linux Enterprise High Availability uses Pacemaker as CRM. The CRM is implemented as
pacemaker-controld
. It interacts with several components: local resource managers, both on its own node and on the other nodes, non-local CRMs, administrative commands, the fencing functionality, and the membership layer.crm
shell (crmsh) #The command-line utility crmsh manages your cluster, nodes and resources.
- Csync2 #
A synchronization tool that can be used to replicate configuration files across all nodes in the cluster, and even across Geo clusters.
- DC (designated coordinator) #
The DC is elected from all nodes in the cluster. This happens if there is no DC yet or if the current DC leaves the cluster for any reason. The DC is the only entity in the cluster that can decide that a cluster-wide change needs to be performed, such as fencing a node or moving resources around. All other nodes get their configuration and resource allocation information from the current DC.
- Disaster #
Unexpected interruption of critical infrastructure induced by nature, humans, hardware failure, or software bugs.
- Disaster Recover Plan #
A strategy to recover from a disaster with minimum impact on IT infrastructure.
- Disaster Recovery #
Disaster recovery is the process by which a business function is restored to the normal, steady state after a disaster.
- DLM (distributed lock manager) #
DLM coordinates disk access for clustered file systems and administers file locking to increase performance and availability.
- DRBD #
DRBD® is a block device designed for building high availability clusters. The whole block device is mirrored via a dedicated network and is seen as a network RAID-1.
- existing cluster #
The term “existing cluster” is used to refer to any cluster that consists of at least one node. Existing clusters have a basic Corosync configuration that defines the communication channels, but they do not necessarily have resource configuration yet.
- failover #
Occurs when a resource or node fails on one machine and the affected resources are started on another node.
- failover domain #
A named subset of cluster nodes that are eligible to run a cluster service if a node fails.
- fencing #
Describes the concept of preventing access to a shared resource by isolated or failing cluster members. There are two classes of fencing: resource level fencing and node level fencing. Resource level fencing ensures exclusive access to a given resource. Node level fencing prevents a failed node from accessing shared resources entirely and prevents resources from running on a node whose status is uncertain. This is usually done in a simple and abrupt way: reset or power off the node.
- Geo cluster (geographically dispersed cluster) #
Consists of multiple, geographically dispersed sites with a local cluster each. The sites communicate via IP. Failover across the sites is coordinated by a higher-level entity, the booth. Geo clusters need to cope with limited network bandwidth and high latency. Storage is replicated asynchronously.
- GFS2 #
Global File System 2 or GFS2 is a shared disk file system for Linux computer clusters. GFS2 allows all nodes to have direct concurrent access to the same shared block storage. GFS2 has no disconnected operating mode, and no client or server roles. All nodes in a GFS2 cluster function as peers. GFS2 supports up to 32 cluster nodes. Using GFS2 in a cluster requires hardware to allow access to the shared storage, and a lock manager to control access to the storage.
- group #
Resource groups contain multiple resources that need to be located together, started sequentially and stopped in the reverse order.
- Hawk2 #
A user-friendly Web-based interface with which you can monitor and administer your High Availability clusters from Linux or non-Linux machines alike. Hawk2 can be accessed from any machine inside or outside of the cluster by using a (graphical) Web browser.
- heuristics #
QDevice supports a set of commands (“heuristics”). The commands are executed locally on start-up of cluster services, cluster membership change, successful connection to the QNetd server, or, optionally, at regular times. Only if all commands are executed successfully are the heuristics considered to have passed; otherwise, they failed. The heuristics' result is sent to the QNetd server, where it is used in calculations to determine which partition should be quorate.
- load balancing #
The ability to make several servers participate in the same service and do the same work.
- local cluster #
A single cluster in one location (for example, all nodes are located in one data center). Network latency can be neglected. Storage is typically accessed synchronously by all nodes.
- location #
In the context of a whole cluster, “location” can refer to the physical location of nodes (for example, all nodes might be located in the same data center).
In the context of a location constraint, “location” refers to the nodes on which a resource can or cannot run.
- location constraint #
Location constraints define on which nodes a resource may be run, may not be run or is preferred to be run. See also resource constraint.
- LRM (local resource manager) #
The local resource manager is located between the Pacemaker layer and the resources layer on each node. It is implemented as
pacemaker-execd
daemon. Through this daemon, Pacemaker can start, stop, and monitor resources.- mcastaddr (multicast address) #
IP address to be used for multicasting by the Corosync executive. The IP address can either be IPv4 or IPv6.
- mcastport (multicast port) #
The port to use for cluster communication.
- meta attributes (resource options) #
Parameters that tell the CRM how to treat a specific resource. For example, the priority or the target role.
- metro cluster #
A single cluster that can stretch over multiple buildings or data centers, with all sites connected by fibre channel. Network latency is usually low (<5 ms for distances of approximately 20 miles). Storage is frequently replicated (mirroring or synchronous replication).
- multicast #
A technology used for a one-to-many communication within a network that can be used for cluster communication. Corosync supports both multicast and unicast.
- network device bonding #
Network device bonding combines two or more network interfaces into a single bonded device to increase bandwidth and/or provide redundancy. When using Corosync, the bonding device is not managed by the cluster software. Therefore, the bonding device must be configured on every cluster node that might need to access the bonding device.
- node #
Any computer (real or virtual) that is a member of a cluster and invisible to the user.
- OCFS2 #
Oracle Cluster File System 2 (OCFS2) is a general-purpose journaling file system that allows you to store application binary files, data files and databases on devices on shared storage. All nodes in a cluster have concurrent read and write access to the file system. A user space control daemon, managed via a clone resource, provides the integration with the HA stack, in particular with Corosync and the Distributed Lock Manager (DLM).
- order constraint #
Order constraints define the sequence of actions. See also resource constraint.
- pacemaker-controld (cluster controller daemon) #
The CRM is implemented as the
pacemaker-controld
daemon. It has an instance on each cluster node. All cluster decision-making is centralized by electing one of the pacemaker-controld instances to act as a primary. If the elected pacemaker-controld process fails (or the node it ran on), a new one is established.- parameters (instance attributes) #
Parameters determine which instance of a service the resource controls.
- PE (policy engine) #
The policy engine is implemented as the
pacemaker-schedulerd
daemon. When a cluster transition is needed, based on the current state and configuration,pacemaker-schedulerd
calculates the expected next state of the cluster. It determines what actions need to be scheduled to achieve the next state.- primitive #
A primitive resource is the most basic type of cluster resource.
- promotable clone #
Promotable clones (also known as multi-state resources) are a special type of clone resource that can be promoted. Active instances of these resources are divided into two states: active and passive. These are also sometimes called primary and secondary.
- QDevice (
corosync-qdevice
) # QDevice and QNetd participate in quorum decisions. With assistance from the arbitrator
corosync-qnetd
,corosync-qdevice
provides a configurable number of votes, allowing a cluster to sustain more node failures than the standard quorum rules allow.- QNetd (
corosync-qnetd
) # A systemd service (a daemon, the “QNetd server”) which is not part of the cluster. QNetd provides a vote to the
corosync-qdevice
daemon to help it participate in quorum decisions.- quorum #
In a cluster, a cluster partition is defined to have quorum (be “quorate”) if it has the majority of nodes (or votes). Quorum distinguishes exactly one partition. It is part of the algorithm to prevent several disconnected partitions or nodes from proceeding and causing data and service corruption (split brain). Quorum is a prerequisite for fencing, which then ensures that quorum is indeed unique.
- RA (resource agent) #
A script acting as a proxy to manage a resource (for example, to start, stop, or monitor a resource). SUSE Linux Enterprise High Availability supports different kinds of resource agents. For details, see Section 10.2, “Supported resource agent classes”.
- ReaR (Relax and Recover) #
An administrator tool set for creating disaster recovery images.
- resource #
Any type of service or application that is known to Pacemaker. Examples include an IP address, a file system, or a database.
The term “resource” is also used for DRBD, where it names a set of block devices that are using a common connection for replication.
- resource constraint #
Resource constraints let you specify which cluster nodes resources can run on, what order resources load in, and what other resources a specific resource is dependent on.
See also colocation constraint, location constraint and order constraint.
- resource set #
As an alternative format for defining location, colocation or order constraints, you can use resource sets, where primitives are grouped together in one set. When creating a constraint, you can specify multiple resources for the constraint to apply to.
- resource template #
If you want to create lots of resources with similar configurations, defining a resource template is the easiest way. After being defined, it can be referenced in primitives or in certain types of constraints. If a template is referenced in a primitive, the primitive inherits all operations, instance attributes (parameters), meta attributes and utilization attributes defined in the template.
- RRP (redundant ring protocol) #
Allows the use of multiple redundant local area networks for resilience against partial or total network faults. This way, cluster communication can still be kept up as long as a single network is operational. Corosync supports the Totem Redundant Ring Protocol.
- SBD (STONITH Block Device) #
Provides a node fencing mechanism through the exchange of messages via shared block storage (SAN, iSCSI, FCoE, etc.). Can also be used in diskless mode. Needs a hardware or software watchdog on each node to ensure that misbehaving nodes are really stopped.
- SFEX (shared disk file exclusiveness) #
SFEX provides storage protection over SAN.
- split-brain #
A scenario in which the cluster nodes are divided into two or more groups that do not know of each other (either through a software or hardware failure). STONITH prevents a split-brain situation from badly affecting the entire cluster. Also known as a “partitioned cluster” scenario.
The term
split brain
is also used in DRBD but means that the two nodes contain different data.- SPOF (single point of failure) #
Any component of a cluster that, should it fail, triggers the failure of the entire cluster.
- STONITH #
The acronym for “Shoot the other node in the head”. It refers to the fencing mechanism that shuts down a misbehaving node to prevent it from causing trouble in a cluster. In a Pacemaker cluster, the implementation of node level fencing is STONITH. For this, Pacemaker comes with a fencing subsystem,
pacemaker-fenced
.- switchover #
Planned, on-demand moving of services to other nodes in a cluster. See failover.
- ticket #
A component used in Geo clusters. A ticket grants the right to run certain resources on a specific cluster site. A ticket can only be owned by one site at a time. Resources can be bound to a certain ticket by dependencies. Only if the defined ticket is available at a site, the respective resources are started. Vice versa, if the ticket is removed, the resources depending on that ticket are automatically stopped.
- unicast #
A technology for sending messages to a single network destination. Corosync supports both multicast and unicast. In Corosync, unicast is implemented as UDP-unicast (UDPU).
- utilization #
Tells the CRM what capacity a certain resource requires from a node.
- YaST #
A graphical user interface for general system installation and administration. Use it to install SUSE Linux Enterprise High Availability on top of SUSE Linux Enterprise Server as described in the Installation and Setup Quick Start.