This chapter covers two different scenarios: upgrading a cluster to another version of SUSE Linux Enterprise High Availability Extension (either a major release or a service pack) as opposed to updating individual packages on cluster nodes. See Section 5.2, “Upgrading your Cluster to the Latest Product Version” versus Section 5.3, “Updating Software Packages on Cluster Nodes”.
If you want to upgrade your cluster, check Section 5.2.1, “Supported Upgrade Paths for SLE HA and SLE HA Geo” and Section 5.2.2, “Required Preparations Before Upgrading” before starting to upgrade.
In the following, find definitions of the most important terms used in this chapter:
The Major Release of SUSE Linux Enterprise (or any software product) is a new version that brings new features and tools, decommissions previously deprecated components and comes with backward incompatible changes.
If a new product version includes major changes that are backward incompatible, the cluster needs to be upgraded by an offline migration: you need to take all nodes offline and upgrade the cluster as a whole, before you can bring all nodes back online.
In a rolling upgrade one cluster node at a time is upgraded while the rest of the cluster is still running: you take the first node offline, upgrade it and bring it back online to join the cluster. Then you continue one by one until all cluster nodes are upgraded to a major version.
Combines several patches into a form that is easy to install or deploy. Service packs are numbered and usually contain security fixes, updates, upgrades, or enhancements of programs.
Installation of a newer minor version of a package.
Installation of a newer major version of a package or distribution, which brings new features. See also Offline Migration versus Rolling Upgrade.
Which upgrade path is supported, and how to perform the upgrade depends both on the current product version and on the target version you want to migrate to.
Rolling upgrades are only supported from the GA of a product version to the next service pack, and from one service pack to the next.
Offline migrations are required to upgrade from one major version to the next major version (for example, from SLE HA 11 to SLE HA 12) or from a service pack belonging to one major version to the next major version (for example, from SLE HA 11 SP3 to SLE HA 12).
For information regarding the upgrade of the base system (SUSE Linux Enterprise Server), see the SUSE Linux Enterprise Server Deployment Guide of the target version you want to upgrade to. The guide is available at https://documentation.suse.com/#sles/.
Section 5.2.1 gives an overview of the supported upgrade paths for SLE HA (Geo) and the additional documentation to refer to.
Mixed clusters running on SUSE Linux Enterprise High Availability Extension 11/SUSE Linux Enterprise High Availability Extension 12 are not supported.
After the upgrade process to product version 12, reverting back to product version 11 is not supported.
Upgrade From ... To |
Upgrade Path |
For Details See |
---|---|---|
SLE HA 11 SP3 to SLE HA (Geo) 12 |
Offline Migration |
|
SLE HA (Geo) 11 SP4 to SLE HA (Geo) 12 SP1 |
Offline Migration |
|
SLE HA (Geo) 12 to SLE HA (Geo) 12 SP1 |
Rolling Upgrade |
|
SLE HA (Geo) 12 SP1 to SLE HA (Geo) 12 SP2 |
Rolling Upgrade |
|
SLE HA (Geo) 12 SP2 to SLE HA (Geo) 12 SP3 |
Rolling Upgrade |
|
SLE HA (Geo) 12 SP3 to SLE HA (Geo) 12 SP4 |
Rolling Upgrade |
|
SLE HA (Geo) 12 SP4 to SLE HA (Geo) 12 SP5 |
Rolling Upgrade |
|
All documentation listed in the column For Details See is available from https://documentation.suse.com.
Ensure that your system backup is up to date and restorable.
Test the upgrade procedure on a staging instance of your cluster setup first, before performing it in a production environment.
This gives you an estimation of the time frame required for the maintenance window. It also helps to detect and solve any unexpected problems that might arise.
This section applies to the following scenarios:
Upgrading from SLE HA 11 SP3 to SLE HA 12
Upgrading from SLE HA 11 SP4 to SLE HA 12 SP1
If your cluster is still based on an older product version than the ones listed above, first upgrade it to a version of SUSE Linux Enterprise Server and SUSE Linux Enterprise High Availability Extension that can be used as a source for upgrading to the desired target version.
The High Availability Extension 12 cluster stack comes with major changes in various
components (for example, /etc/corosync/corosync.conf
, disk formats of OCFS2).
Therefore, a rolling upgrade
from any SUSE Linux Enterprise High Availability Extension
11 version is not supported. Instead, all cluster nodes must be offline
and the cluster needs to be migrated as a whole as described in
Procedure 5.1, “Performing a Cluster-wide Offline Migration”.
Log in to each cluster node and stop the cluster stack with:
root #
rcopenais
stop
For each cluster node, perform an upgrade to the desired target version of SUSE Linux Enterprise Server and SUSE Linux Enterprise High Availability Extension. If you have an existing Geo cluster setup and want to upgrade it, see the additional instructions in the Geo Clustering for SUSE Linux Enterprise High Availability Extension Geo Clustering Quick Start. To find the details for the individual upgrade processes, see Section 5.2.1, “Supported Upgrade Paths for SLE HA and SLE HA Geo”.
After the upgrade process has finished, reboot each node with the upgraded version of SUSE Linux Enterprise Server and SUSE Linux Enterprise High Availability Extension.
If you use OCFS2 in your cluster setup, update the on-device structure by executing the following command:
root #
o2cluster
--update PATH_TO_DEVICE
It adds additional parameters to the disk which are needed for the updated OCFS2 version that is shipped with SUSE Linux Enterprise High Availability Extension 12 and 12 SPx.
To update /etc/corosync/corosync.conf
for Corosync version 2:
Log in to one node and start the YaST cluster module.
Switch to the Procedure 4.1, “Defining the First Communication Channel (Multicast)” or Procedure 4.2, “Defining the First Communication Channel (Unicast)”, respectively.
category and enter values for the following new parameters: and . For details, seeIf YaST should detect any other options that are invalid or missing according to Corosync version 2, it will prompt you to change them.
Confirm your changes in YaST. YaST will write them to
/etc/corosync/corosync.conf
.
If Csync2 is configured for your cluster, use the following command to push the updated Corosync configuration to the other cluster nodes:
root #
csync2
-xv
For details on Csync2, see Section 4.5, “Transferring the Configuration to All Nodes”.
Alternatively, synchronize the updated Corosync configuration by
manually copying /etc/corosync/corosync.conf
to all cluster nodes.
Log in to each node and start the cluster stack with:
root #
systemctl
start pacemaker
Check the cluster status with crm status
or with
Hawk2.
Configure the following services to start at boot time:
root #
systemctl enable pacemakerroot #
systemctl enable hawkroot #
systemctl enable sbd
Tags (for grouping resources) and some ACL features only work with the
CIB syntax version pacemaker-2.0
or higher. (To
check your version, use the cibadmin -Q |grep
validate-with
command.) If you have upgraded from SUSE Linux Enterprise High Availability Extension 11
SPx, your CIB version will not be upgraded by
default. To manually upgrade to the latest CIB version use one of the
following commands:
root #
cibadmin
--upgrade --force
or
root #
crm
configure upgrade force
This section applies to the following scenarios:
Upgrading from SLE HA 12 to SLE HA 12 SP1
Upgrading from SLE HA 12 SP1 to SLE HA 12 SP2
Upgrading from SLE HA (Geo) 12 SP2 to SLE HA (Geo) 12 SP3
Upgrading from SLE HA (Geo) 12 SP3 to SLE HA (Geo) 12 SP4
Upgrading from SLE HA (Geo) 12 SP4 to SLE HA (Geo) 12 SP5
Before starting an upgrade for a node, stop the cluster stack on that node.
If the cluster resource manager on a node is active during the software update, this can lead to unpredictable results like fencing of active nodes.
Log in as root
on the node that you want to upgrade and stop the
cluster stack:
root #
systemctl
stop pacemaker
Perform an upgrade to the desired target version of SUSE Linux Enterprise Server and SUSE Linux Enterprise High Availability Extension. To find the details for the individual upgrade processes, see Section 5.2.1, “Supported Upgrade Paths for SLE HA and SLE HA Geo”.
Restart the cluster stack on the upgraded node to make the node rejoin the cluster:
root #
systemctl
start pacemaker
Take the next node offline and repeat the procedure for that node.
Check the cluster status with crm status
or with
Hawk2.
The new features shipped with the latest product version will only be available after all cluster nodes have been upgraded to the latest product version. Mixed version clusters are only supported for a short time frame during the rolling upgrade. Complete the rolling upgrade within one week.
The Hawk2
screen also shows a warning if different CRM versions are detected for your cluster nodes.Before starting an update for a node, either stop the cluster stack on that node or put the node into maintenance mode, depending on whether the cluster stack is affected or not. See Step 1 for details.
If the cluster resource manager on a node is active during the software update, this can lead to unpredictable results like fencing of active nodes.
Before installing any package updates on a node, check the following:
Does the update affect any packages belonging to SUSE Linux Enterprise High Availability Extension or the
Geo clustering extension? If yes
: Stop the cluster stack on
the node before starting the software update:
root #
systemctl
stop pacemaker
Does the package update require a reboot? If yes
:
Stop the cluster stack on the node before starting the software
update:
root #
systemctl
stop pacemaker
If none of the situations above apply, you do not need to stop the cluster stack. In that case, put the node into maintenance mode before starting the software update:
root #
crm
node maintenance NODE_NAME
For more details on maintenance mode, see Section 16.2, “Different Options for Maintenance Tasks”.
Install the package update using either YaST or Zypper.
After the update has been successfully installed:
Either start the cluster stack on the respective node (if you stopped it in Step 1):
root #
systemctl
start pacemaker
or remove the maintenance flag to bring the node back to normal mode:
root #
crm
node ready NODE_NAME
Check the cluster status with crm status
or with
Hawk2.
For detailed information about any changes and new features of the product you are upgrading to, refer to its release notes. They are available from https://www.suse.com/releasenotes/.