SAP Data Intelligence 3 on Rancher Kubernetes Engine 2 #
Containerization
SAP
SAP Data Intelligence 3 is the tool set to govern big amounts of data, and it runs fully containerized. This document describes the installation and configuration of SAP Data Intelligence 3 deployed on Rancher Kubernetes Engine 2.
Disclaimer: Documents published as part of the SUSE Best Practices series have been contributed voluntarily by SUSE employees and third parties. They are meant to serve as examples of how particular actions can be performed. They have been compiled with utmost attention to detail. However, this does not guarantee complete accuracy. SUSE cannot verify that actions described in these documents do what is claimed or whether actions described have unintended consequences. SUSE LLC, its affiliates, the authors, and the translators may not be held liable for possible errors or the consequences thereof.
1 Introduction #
This guide describes the on-premises installation of SAP Data Intelligence 3.3 on Rancher Kubernetes Engine (RKE) 2. In a nutshell, the installation of SAP Data Intelligence 3.3 consists of the following steps:
Installing SUSE Rancher
Installing RKE 2 Kubernetes cluster on the dedicated nodes
Deploying SAP Data Intelligence 3.3 on RKE 2 Kubernetes cluster
Performing post-installation steps for SAP Data Intelligence 3.3
Testing the installation of SAP Data Intelligence 3.3
To have a fully supported setup, there are two Kubernetes clusters required. One runs SUSE Rancher Management server and the other runs the actual workload, which for the purpose of this guide is SAP Data Intelligence.
2 Requirements #
2.1 Hardware requirements #
This chapter describes the hardware requirements for installing SAP Data Intelligence 3.3 on RKE 2 on top of SUSE Linux Enterprise Server 15 SP4. Only the AMD64/Intel 64 architecture is applicable for our use case.
2.1.1 Hardware Sizing #
Correct hardware sizing is very important for setting up SAP Data Intelligence 3.3 on RKE 2.
2.1.1.1 Development systems #
Minimal hardware requirements for a generic SAP Data Intelligence 3 deployment:
At least 7 nodes are needed for the Kubernetes cluster.
Minimum sizing of the nodes needs to be as shown below:
Server Role | Count | RAM | CPU | Disk space |
---|---|---|---|---|
Management Workstation | 1 | 16 GiB | 4 | >100 GiB |
Master Node | 3 | 16 GiB | 4 | >120 GiB |
Worker Node | 4 | 32 GiB | 8 | >120 GiB |
2.1.1.2 Production systems #
Minimal hardware requirements for an SAP Data Intelligence 3 deployment for production use:
At least seven nodes are needed for the Kubernetes cluster.
Minimum sizing of the nodes needs to be as shown below:
Server Role Count RAM CPU Disk space Management Workstation
1
16 GiB
4
>100 GiB
Master Node
3
16 GiB
4
>120 GiB
Worker Node
4
64 GiB
16
>120 GiB
2.2 Software requirements #
The following list contains the software components needed to install SAP Data Intelligence 3.3 on RKE:
SUSE Linux Enterprise Server 15 SP4
Rancher Kubernetes Engine 2
SAP Software Lifecycle Bridge
SAP Data Intelligence 3.3
Secure private registry for container images, for example https://documentation.suse.com/sbp/all/single-html/SBP-Private-Registry/index.html
Access to a storage solution providing dynamically physical volumes
If it is planned to use Vora’s streaming tables checkpoint store, an S3 bucket like object store is needed
If it is planned to enable backup of SAP Data Intelligence 3.3 during installation access to an S3-compatible object store is needed
2.3 Installation on top of Harvester #
When using Harvester to provision the virtual machines for an SAP Data Intelligence installation, the hardware requirements for Harvester need to be added to the requirements of SAP Data Intelligence described at Section 2.1.1, “Hardware Sizing”.
2.3.1 Harvester hardware requirements #
A full list of requirements for Harvester can be found at https://docs.harvesterhci.io/v1.0/install/requirements
2.3.2 Development systems #
The recommended setup for a Harvester cluster to be used for development has the following requirements:
Server Role | Count | RAM | CPU | Disk space |
---|---|---|---|---|
Harvester Node | 3 | 32 GiB | 8 | >140 GiB |
Adding the hardware requirements of SAP Data Intelligence as described in chapter Section 2.1.1.1, “Development systems”, the following hardware is required to run an SAP Data Intelligence cluster on top of Harvester for development:
Server Role | Count | RAM | CPU | Disk space |
---|---|---|---|---|
Harvester Node | 3 | 288 GiB | 72 | >1360 GiB |
2.3.3 Production systems #
The recommended setup for a Harvester cluster to be used in production has the following requirements:
Server Role | Count | RAM | CPU | "Disk space" | Disk speed |
---|---|---|---|---|---|
Harvester Node | 3 | 64 GiB | 16 | >500 GiB | >5000 IOPs |
Adding the hardware requirements of SAP Data Intelligence as described in chapter Section 2.1.1.2, “Production systems”, the following hardware is required to run an SAP Data Intelligence cluster on top of Harvester in production:
Server Role | Count | RAM | CPU | Disk space |
---|---|---|---|---|
Harvester Node | 3 | 512 GiB | 128 | >2440 GiB |
3 Installing SUSE Rancher #
3.1 Preparation #
To have a highly available SUSE Rancher setup, you need a load balancer for your SUSE Rancher nodes.
This section describes how to set up a custom load balancer using haproxy
. If you already have a load balancer, you can use that to make SUSE Rancher highly available.
If you do not plan to set up a highly available SUSE Rancher cluster, you can skip this section.
3.1.1 Installing an haproxy
-based load balancer #
Set up a virtual machine or a bare metal server with SUSE Linux Enterprise Server and SUSE Linux Enterprise High Availability or use SUSE Linux Enterprise Server for SAP Applications.
Install the haproxy
package.
$ zypper in haproxy
Create the configuration for haproxy
.
Find an example configuration file for haproxy
below and adapt for the actual environment.
# cat <<EOF > /etc/haproxy/haproxy.cfg
global
log /dev/log daemon
maxconn 32768
chroot /var/lib/haproxy
user haproxy
group haproxy
daemon
tune.bufsize 32768
tune.ssl.default-dh-param 2048
ssl-default-bind-ciphers ALL:!aNULL:!eNULL:!EXPORT:!DES:!3DES:!MD5:!PSK:!RC4:!ADH:!LOW@STRENGTH
defaults
log global
mode tcp
option log-health-checks
option log-separate-errors
option dontlog-normal
option dontlognull
option tcplog
retries 3
option redispatch
maxconn 10000
timeout connect 5s
timeout client 50s
timeout server 450s
listen stats
bind 0.0.0.0:80
bind :::80 v6only
stats enable
stats uri /
stats refresh 5s
# access the kubernetes api
frontend kubeapi
bind *:6443
mode tcp
default_backend kubeapibackend
# address to register new nodes
frontend rke2server
bind *:9345
mode tcp
default_backend rke2serverbackend
backend kubeapibackend
balance roundrobin
server mynode1 192.168.122.20:6443 check
server mynode2 192.168.122.30:6443 check
server mynode3 192.168.122.40:6443 check
backend rke2serverbackend
balance roundrobin
server mynode1 192.168.122.20:9345 check
EOF
Check the configuration file:
$ haproxy -f /path/to/your/haproxy.conf -c
Enable and start the haproxy
load balancer:
$ systemctl enable haproxy $ systemctl start haproxy
Do not forget to restart or reload haproxy
if any changes are made to the haproxy configuration file.
3.1.2 Installing RKE2 #
To install RKE2, the script provided at https://get.rke2.io can be used as follows:
$ curl -sfL https://get.rke2.io | sh -
For HA setups, it is necessary to create RKE2 cluster configuration files in advance. On the first master node:
$ mkdir -p /etc/rancher/rke2
$ cat <<EOF > /etc/rancher/rke2/config.yaml
token: 'your cluster token'
system-default-registry: registry.rancher.com
tls-san:
- FQDN of fixed registration address on load balancer
- other hostname
- IP v4 address
EOF
Create configuration files for additional cluster nodes:
$ cat <<EOF > /etc/rancher/rke2/config.yaml
server: https://"FQDN of registration address":9345
token: 'your cluster token'
system-default-registry: registry.rancher.com
tls-san:
- FQDN of fixed registration address on load balancer
- other hostname
- IP v4 address
EOF
You also need take about ETCD Snapshots and to perfom backups of your Rancher instance. This is not part of this Document and you can find more information in our Documentation.
For security reasons, we generally recommend activating the CIS profile when installing RKE2. This is currently still being validated and will be included in the documentation at a later date.
Now enable and start the RKE2 components and run the following command on each cluster node:
$ systemctl enable rke2-server --now
To verify the installation, run the following command:
$ /var/lib/rancher/rke2/bin/kubectl --kubeconfig /etc/rancher/rke2/rke2.yaml get nodes
For convenience, the kubectl
binary can be added to the $PATH and the given kubeconfig
can be set via an environment variable:
$ export PATH=$PATH:/var/lib/rancher/rke2/bin/
$ export KUBECONFIG=/etc/rancher/rke2/rke2.yaml
3.1.3 Installing Helm #
To install SUSE Rancher and some of its required components, you need to use Helm.
One way to install Helm is to run:
$ curl https://raw.githubusercontent.com/helm/helm/main/scripts/get-helm-3 | bash
3.1.4 Installing cert-manager #
To install the cert-manager
package, do the following:
$ kubectl create namespace cert-manager
How to create the imagePullSecret is described in the SAP-EIC-ImagePullSecrets.xml.
3.1.4.1 Installing the application #
You will need to login to the Rancher Application Collection:
$ helm registry login dp.apps.rancher.io/charts -u <yourUser> -p <your-token>
$ helm install cert-manager oci://dp.apps.rancher.io/charts/cert-manager \
--set crds.enabled=true \
--set-json 'global.imagePullSecrets=[{"name":"application-collection"}]' \
--namespace=cert-manager \
--version 1.15.2
3.2 Installing SUSE Rancher #
To install SUSE Rancher, you need to add the related Helm repository. To achieve that, use the following command:
$ helm repo add rancher-prime https://charts.rancher.com/server-charts/prime
Next, create the cattle-system
namespace in Kubernetes as follows:
$ kubectl create namespace cattle-system
The Kubernetes cluster is now ready for the installation of SUSE Rancher:
$ helm install rancher rancher-prime/rancher \
--namespace cattle-system \
--set hostname=<your.domain.com> \
--set replicas=3
During the rollout of SUSE Rancher, you can monitor the progress using the following command:
$ kubectl -n cattle-system rollout status deploy/rancher-prime
When the deployment is done, you can access the SUSE Rancher cluster at https://<your.domain.com>. Here you will also find a description about how to log in for the first time.
4 Installing Longhorn #
This chapter details the minimum requirements to install Longhorn and describes three different ways for the installation. For more details, visit https://longhorn.io/docs/1.6.2/deploy/install/
4.1 Requirements #
Before Longhorn can be installed on a Kubernetes cluster,
all nodes must have the open-iscsi
package installed, and the ISCSI daemon needs to be started. To do so, run:
# zypper in -y open-iscsi
# systemctl enable iscsid --now
To esure a node is prepared for Longhorn, you can use the following script to check:
$ curl -sSfL https://raw.githubusercontent.com/longhorn/longhorn/v1.6.2/scripts/environment_check.sh | bash
4.2 Installing Longhorn using SUSE Rancher #
Up-to-date and detailed instructions how to install Longhorn using SUSE Rancher can be found at https://longhorn.io/docs/1.6.2/deploy/install/install-with-rancher/
4.3 Installing Longhorn using Helm #
To install Longhorn using Helm, run the following commands:
$ helm repo add longhorn https://charts.longhorn.io
$ helm repo update
$ helm install longhorn longhorn/longhorn --namespace longhorn-system --create-namespace
These commands will add the Longhorn Helm charts to the list of Helm repositories, update the Helm repository, and execute the installation of Longhorn.
=== Installing Longhorn using kubectl
You can install Longhorn using kubectl
with the following command:
$ kubectl apply -f https://raw.githubusercontent.com/longhorn/longhorn/v1.6.2/deploy/longhorn.yaml
4.4 Exposing Longhorn UI by creating an Ingress with Basic Authentication #
Create a basic auth file:
$ USER=<USERNAME_HERE>; \ PASSWORD=<PASSWORD_HERE>; \ echo "${USER}:$(openssl passwd -stdin -apr1 <<< ${PASSWORD})" >> auth
Create a Secret from the file auth:
$ kubectl -n longhorn-system create secret generic basic-auth --from-file=auth
Create the Ingress with basic authentication:
$ cat <<EOF > longhorn-ingress.yaml apiVersion: networking.k8s.io/v1beta1 kind: Ingress metadata: name: longhorn-ingress namespace: longhorn-system annotations: # type of authentication nginx.ingress.kubernetes.io/auth-type: basic # prevent the controller from redirecting (308) to HTTPS nginx.ingress.kubernetes.io/ssl-redirect: 'false' # name of the secret that contains the user/password definitions nginx.ingress.kubernetes.io/auth-secret: basic-auth # message to display with an appropriate context why the authentication is required nginx.ingress.kubernetes.io/auth-realm: 'Authentication Required ' spec: rules: - http: paths: - path: / backend: serviceName: longhorn-frontend servicePort: 80 EOF $ kubectl -n longhorn-system apply -f longhorn-ingress.yaml
For more details, visit https://longhorn.io/docs/1.6.2/deploy/accessing-the-ui/longhorn-ingress/.
5 Installation of RKE 2 on top of VMware vSphere and VMware vSAN #
5.1 Prerequisites: #
A running VMware vSphere / vSAN installation.
NOTE: The installation of the VMware vSphere / vSAN environment is not in the scope of this document.
Create the virtual machines for the RKE 2 cluster with SUSE Linux Enterprise Server 15 SP4 as operating system in the vSphere environment. Make sure these virtual machines are sized according to the recommendations given above in this guide.
Make sure that
uuid
creation for disks is enabled in the settings for the virtual machines.
5.2 Install RKE 2 cluster on top of the VMware virtual machines. #
Before you start the installation of RKE 2, create the configuration below for the RKE 2 cluster. This is neccessary to use the vSAN as backing storage for the storage class in RKE 2. You will need the following data:
user on vSphere/vSAN with the necessary access rights
vCenter hostname
datacenter ID
ClusterID
vSAN url / datastorage url
You should obtain this information from the VMware vSphere/vSAN administrator.
These data will be used to configure the helm manifests for the vsphere CPI and CSI provider and to access the resources in the vSphere installation.
To use the vSphere CPI and CSI, RKE2 must be configured to use the rancher-vsphere cloud provider.
$ sudo mkdir -p /etc/rancher/rke2 $ sudo echo "cloud-provider-name: rancher-vsphere" > /etc/rancher/rke2/config.yaml"
This enables the deployment of the vSphere CPI and CSI from pre-packaged Helm charts in RKE 2. It will also deploy a storage class that makes use of the vSphere CPI/CSI drivers.
Create the configuration for the CPI vSphere provider Helm chart:
Create the directory structure on first the master node
$ sudo mkdir -p /var/lib/rancher/rke2/server/manifests $ cd /var/lib/rancher/rke2/server/manifests
Then create the file rancher-vsphere-cpi-config.yaml
in the directory.
/var/lib/rancher/rke2/server/manifests
$ cat <<EOF > apiVersion: helm.cattle.io/v1 kind: HelmChartConfig metadata: name: rancher-vsphere-cpi labels: namespace: kube-system spec: valuesContent: |- vCenter: host: "vcenterhostname" datacenters: "datacentername" username: "xxxxxxxxxxx" password: "xxxxxxxxxxxx" insecure: true credentialsSecret: generate: true cloudControllerManager: nodeSelector: node-role.kubernetes.io/control-plane: "true" EOF
In the same directory, the file rancher-vsphere-csi-config.yaml
will be created.
$ cat <<EOF > /var/lib/rancher/rke2/server/manifests/rancher-vsphere-csi-config.yaml apiVersion: helm.cattle.io/v1 kind: HelmChartConfig metadata: name: rancher-vsphere-csi namespace: kube-system spec: valuesContent: |- vCenter: host: "vcenter host" datacenters: "datacenter" username: "xxxxxxx" password: "xxxxxxxxx" clusterId: "vSANClusterID" insecure: true configSecret: configTemplate: | [Global] cluster-id = {{ required ".Values.vCenter.clusterId must be provided" (default .Values.vCenter.clusterId .Values.global.cattle.clusterId) | quote }} user = {{ .Values.vCenter.username | quote }} password = {{ .Values.vCenter.password | quote }} port = {{ .Values.vCenter.port | quote }} insecure-flag = {{ .Values.vCenter.insecureFlag | quote }} [VirtualCenter {{ .Values.vCenter.host | quote }}] datacenters = {{ .Values.vCenter.datacenters | quote }} [Labels] storageClass: datastoreURL: "ds:///vmfs/volumes/vsan:XXXXXXXXXXX/" csiController: nodeSelector: node-role.kubernetes.io/control-plane: "true" EOF
See the RKE 2 documentation here:
Now you can deploy the RKE 2 cluster on the dedicated virtual machines.
Connect to the nodes dedicated as master for the RKE 2 cluster
Download and install RKE 2
$ export INSTALL_RKE2_TYPE=server $ export INSTALL_RKE2_VERSION=<wanted version here> $ curl -sfL https://get.rke2.io | sh - $ systemctl enable --now rke2-server.service
Connect to the nodes dedicated as workers of the RKE 2 cluster:
$ export INSTALL_RKE2_TYPE=agent $ export INSTALL_RKE2_VERSION=<wanted version here> $ curl -sfL https://get.rke2.io | sh - $ systemctl enable --now rke2-agent.service
More details can be found in the RKE 2 documentation:
After the deployment of the RKE 2 cluster, check the availability of the storage class vsphere-csi-sc which should have been created.
$ kubectl get sc NAME PROVISIONER RECLAIMPOLICY VOLUMEBINDINGMODE ALLOWVOLUMEEXPANSION AGE vsphere-csi-sc (default) csi.vsphere.vmware.com Delete Immediate false 17m
Now you can proceed with installing SAP Data Intelligence.
6 Installing SAP Data Intelligence 3.3 #
This section describes the installation of SAP Data Intelligence 3.3 on an RKE 2-powered Kubernetes cluster.
6.1 Preparation #
The following steps need to be executed before the deployment of SAP Data Intelligence 3.3 can start:
Create a namespace for SAP Data Intelligence 3.3.
Create an access to a secure private registry.
Create a default storage class.
Download and install SAP SLC Bridge.
Download the stack.xml file for provisioning the DI 3.3 installation.
Check if the
nfsd
andnfsv4
kernel modules are loaded and/or loadable on the Kubernetes nodes.
6.1.1 Creating namespace for SAP Data Intelligence 3.3 in the Kubernetes cluster #
Log in to your management workstation and create the namespace in the Kubernetes cluster where DI 3.3 will be deployed.
$ kubectl create ns <NAMESPACE for DI 31> $ kubectl get ns
6.1.2 Creating cert file to access the secure private registry #
Create a file named cert that contains the SSL certificate chain for the secure private registry. This imports the certificates into SAP Data Intelligence 3.3.
$ cat CA.pem > cert $ kubectl -n <NAMESPACE for DI 31> create secret generic cmcertificates --from-file=cert
6.2 Creating default storage class #
To install SAP Data Intelligence 3.3, a default storage class is needed to provision the installation with physical volumes (PV).
Below find an example for a ceph
/rbd
based storage class that uses the CSI.
Create the yaml files for the storage class. Contact your storage admin to get the required information.
Create
config-map
:$ cat << EOF > csi-config-map.yaml --- apiVersion: v1 kind: ConfigMap data: config.json: |- [ { "clusterID": "<ID of your ceph cluster>", "monitors": [ "<IP of Monitor 1>:6789", "<IP of Monitor 2>:6789", "<IP of Monitor 3>:6789" ] } ] metadata: name: ceph-csi-config EOF
Create a secret to access the storage:
$ cat << EOF > csi-rbd-secret.yaml --- apiVersion: v1 kind: Secret metadata: name: csi-rbd-secret namespace: default stringData: userID: admin userKey: AQCR7htglvJzBxAAtPN0YUeSiDzyTeQe0lveDQ== EOF
Download the file:
$ curl -LO https://raw.githubusercontent.com/ceph/ceph-csi/master/deploy/rbd/kubernetes/csi-rbdplugin-provisioner.yaml
Download the file:
$ curl -LO https://raw.githubusercontent.com/ceph/ceph-csi/master/deploy/rbd/kubernetes/csi-rbdplugin.yaml
Create a pool on the Ceph storage where the PVs will be created, and insert the pool name and the Ceph cluster ID:
$ cat << EOF > csi-rbd-sc.yaml --- apiVersion: storage.k8s.io/v1 kind: StorageClass metadata: name: csi-rbd-sc provisioner: rbd.csi.ceph.com parameters: clusterID: <your ceph cluster id> pool: <your pool> csi.storage.k8s.io/provisioner-secret-name: csi-rbd-secret csi.storage.k8s.io/provisioner-secret-namespace: default csi.storage.k8s.io/node-stage-secret-name: csi-rbd-secret csi.storage.k8s.io/node-stage-secret-namespace: default reclaimPolicy: Delete mountOptions: - discard EOF
Create config for encryption. This is needed, else the deployment of the CSI driver for
ceph
/rbd
will fail.$ cat << EOF > kms-config.yaml --- apiVersion: v1 kind: ConfigMap data: config.json: |- { }, "vault-tokens-test": { "encryptionKMSType": "vaulttokens", "vaultAddress": "http://vault.default.svc.cluster.local:8200", "vaultBackendPath": "secret/", "vaultTLSServerName": "vault.default.svc.cluster.local", "vaultCAVerify": "false", "tenantConfigName": "ceph-csi-kms-config", "tenantTokenName": "ceph-csi-kms-token", "tenants": { "my-app": { "vaultAddress": "https://vault.example.com", "vaultCAVerify": "true" }, "an-other-app": { "tenantTokenName": "storage-encryption-token" } } } } metadata: name: ceph-csi-encryption-kms-config EOF
Deploy the
ceph
/rbd
CSI and storage class:$ kubectl apply -f csi-config-map.yaml $ kubectl apply -f csi-rbd-secret.yaml $ kubectl apply -f \ https://raw.githubusercontent.com/ceph/ceph-csi/master/deploy/rbd/kubernetes/csi-provisioner-rbac.yaml $ kubectl apply -f \ https://raw.githubusercontent.com/ceph/ceph-csi/master/deploy/rbd/kubernetes/csi-nodeplugin-rbac.yaml $ kubectl apply -f csi-rbdplugin-provisioner.yaml $ kubectl apply -f csi-rbdplugin.yaml $ kubectl apply -f csi-rbd-sc.yaml $ kubectl apply -f kms-config.yaml $ kubectl patch storageclass csi-rbd-sc \ -p '{"metadata": {"annotations":{"storageclass.kubernetes.io/is-default-class":"true"}}}'
Check your storage class:
$ kubectl get sc NAME PROVISIONER RECLAIMPOLICY VOLUMEBINDINGMODE ALLOWVOLUMEEXPANSION AGE csi-rbd-sc (default) rbd.csi.ceph.com Delete Immediate false 103m
6.3 Downloading the SLC Bridge #
The SLC Bridge can be obtained:
from the SAP software center at https://support.sap.com/en/tools/software-logistics-tools.html#section_622087154. Choose "Download SLC Bridge".
via the information in the release notes of the SLC Bridge at https://launchpad.support.sap.com/#/notes/2589449.
Download the SLC Bridge software to the management workstation.
6.4 Installing the SLC Bridge #
Rename the SLC Bridge binary to slcb
and make it executable. Deploy the SLC Bridge to the Kubernetes cluster.
$ mv SLCB01_XX-70003322.EXE slcb $ chmod 0700 slcb $ export KUBECONFIG=<KUBE_CONFIG> $ ./slcb init
During the interactive installation, the following information is needed:
URL of secure private registry
Choose expert mode
Choose NodePort for the service
Take a note of the service port of the SLC Bridge. It is needed for the installation of SAP Data Intelligence 3.3 or for the reconfiguration of DI 3.3, for example to enable backup. If you forgot to note it down, the following command will list the service port:
$ kubectl -n sap-slcbridge get svc
6.5 Creating and downloading Stack XML for the SAP Data Intelligence installation #
Follow the steps described in the chapter Install SAP Data Intelligence with SLC Bridge in a Cluster with Internet Access of the SAP Data Intelligence 3.3 Installation Guide.
6.5.1 Creating Stack XML #
You can create the Stack XML via the SAP Maintenance Planner. Access the tool via https://support.sap.com/en/alm/solution-manager/processes-72/maintenance-planner.html. Go to the Maintenance Planner at https://apps.support.sap.com/sap/support/mp published on the SAP Web site and generate a Stack XML file with the container image definitions of the SAP Data Intelligence release that you want to install. Download the Stack XML file to a local directory. Copy stack.xml to the management workstation.
6.6 Running the installation of SAP Data Intelligence #
The installation of SAP Data Intelligence 3.3 is invoked by:
$ export KUBECONFIG=<path to kubeconfig> $ ./slcb execute --useStackXML MP_Stack_XXXXXXXXXX_XXXXXXXX_.xml --url https://<node>:<service port>/docs/index.html
This starts an interactive process for configuring and deploying SAP Data Intelligence 3.3.
The table below lists some parameters available for an SAP Data Intelligence 3.3 installation:
Parameter | Condition | Recommendation |
---|---|---|
Kubernetes Namespace | Always | set to namespace created beforehand |
Installation Type | installation or update | either |
Container Registry | Always | add the uri for the secure private registry |
Checkpoint Store Configuration | installation | whether to enable Checkpoint Store |
Checkpoint Store Type | if Checkpoint Store is enabled | use S3 object store from SES |
Checkpoint Store Validation | if Checkpoint is enabled | Object store access will be verified |
Container Registry Settings for Pipeline Modeler | optional | used if a second container registry is used |
StorageClass Configuration | optional, needed if a different StorageClass is used for some components | leave the default |
Default StorageClass | detected by SAP Data Intelligence installer | The Kubernetes cluster shall have a storage class annotated as default SC |
Enable Kaniko Usage | optional if running on Docker | enable |
Container Image Repository Settings for SAP Data Intelligence Modeler | mandatory | |
Container Registry for Pipeline Modeler | optional | Needed if a different container registry is used for the pipeline modeler images |
Loading NFS Modules | optional | Make sure that nfsd and nfsv4 kernel modules are loaded on all worker nodes |
Additional Installer Parameters | optional |
For more details about input parameters for an SAP Data Intelligence 3.3 installation, visit the section Required Input Parameters of the SAP Data Intelligence Installation Guide.
6.7 Post-installation tasks #
After the installation workflow is successfully finished, you need to carry out some additional tasks:
Obtain or create an SSL certificate to securely access the SAP Data Intelligence installation:
Create a certificate request using
openssl
, for example:$ openssl req -newkey rsa:2048 -keyout <hostname>.key -out <hostname>.csr
Decrypt the key:
$ openssl rsa -in <hostname>.key -out decrypted-<hostname>.key
Let a CA sign the <hostname>.csr You will receive a <hostname>.crt.
Create a secret from the certificate and the key in the SAP Data Intelligence 3 namespace:
$ export NAMESPACE=<{di} 3 namespace> $ kubectl -n $NAMESPACE create secret tls vsystem-tls-certs --key decrypted-<hostname>.key--cert <hostname>.crt
Deploy an
nginx-ingress
controller:For more information, see https://kubernetes.github.io/ingress-nginx/deploy/#bare-metal.
Create the
nginx-ingress
controller as a nodePort service according to the Ingressnginx
documentation:$ kubectl apply -f https://raw.githubusercontent.com/kubernetes/ingress-nginx/controller-v0.46.0/deploy/static/provider/baremetal/deploy.yaml
Determine the port the
nginx
controller is redirecting HTTPS to:$ kubectl -n ingress-nginx get svc ingress-nginx-controller
The output should be similar to the below:
kubectl -n ingress-nginx get svc ingress-nginx-controller NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE ingress-nginx-controller NodePort 10.43.86.90 <none> 80:31963/TCP,443:{di_version}06/TCP 53d
In our example here, the TLS port is be 3.306. Note the port IP down as you will need it to access the SAP Data Intelligence installation from the outside.
Create an Ingress to access the SAP Data Intelligence installation:
$ cat <<EOF > ingress.yaml apiVersion: networking.k8s.io/v1 kind: Ingress metadata: annotations: kubernetes.io/ingress.class: nginx nginx.ingress.kubernetes.io/force-ssl-redirect: "true" nginx.ingress.kubernetes.io/secure-backends: "true" nginx.ingress.kubernetes.io/backend-protocol: HTTPS nginx.ingress.kubernetes.io/proxy-body-size: "0" nginx.ingress.kubernetes.io/proxy-buffer-size: 16k nginx.ingress.kubernetes.io/proxy-connect-timeout: "30" nginx.ingress.kubernetes.io/proxy-read-timeout: "1800" nginx.ingress.kubernetes.io/proxy-send-timeout: "1800" name: vsystem spec: rules: - host: "<hostname FQDN must match SSL certificate" http: paths: - backend: serviceName: vsystem servicePort: 8797 path: / tls: - hosts: - "<hostname FQDN must match SSL certificate>" secretName: vsystem-tls-certs EOF $ kubectl apply -f ingress.yaml
Connecting to https://hostname:<ingress service port> brings up the SAP Data Intelligence login dialog.
6.8 Testing the SAP Data Intelligence 3 installation #
Finally, the SAP Data Intelligence 3 installation should be verified with some very basic tests:
Log in to SAP Data Intelligence’s launchpad
Create example pipeline
Create ML Scenario
Test machine learning
Download
vctl
For details, see the SAP Data Intelligence 3 Installation Guide
7 Maintenance tasks #
This section provides some tips about what should and could be done to maintain the Kubernetes cluster, the operating system and the SAP Data Intelligence 3 deployment.
7.1 Backup #
It is good practice to keep backups of all relevant data to be able to restore the environment in case of a failure. To perform regular backups, follow the instructions as outlined in the respective documentation below:
For RKE 2, consult section Backups and Disaster Recovery
SAP Data Intelligence 3 can be configured to create regular backups. For more information, visit help.sap.com:
7.2 Upgrade or update #
This section explains how you can keep your installation of SAP Data Intelligence, RKE 2 and SUSE Linux Enterprise Server up-to-date.
7.2.1 Updating the operating system #
To obtain updates for SUSE Linux Enterprise Server 15 SP4, the installation must be registered either to SUSE Customer Center, an SMT or RMT server, or SUSE Manager with a valid subscription.
SUSE Linux Enterprise Server 15 SP4 can be updated on the command line using
zypper
:$ sudo zypper ref -s $ sudo zypper lu $ sudo zypper patch
Other methods for updating SUSE Linux Enterprise Server 15 SP4 are described in the product documentation.
If an update requires a reboot of the server, make sure that this can be done safely.
For example, block access to SAP Data Intelligence, and drain and cordon the Kubernetes node before rebooting:
$ kubectl edit ingress <put in some dummy port> $ kubectl drain <node>
Check the status of the node:
$kubectl get node <node>
The node should be marked as not schedulable.
On RKE 2 master nodes, run the command:
$ sudo systemctl stop rke2-server
On RKE 2 worker nodes, run the command:
$ sudo systemctl stop rke2-agent
Update SUSE Linux Enterprise Server 15 SP4:
$ ssh node $ sudo zypper patch
Reboot the nodes if necessary or start the appropriate RKE 2 service.
On master nodes, run the command:
$ sudo systemctl start rke2-server
On worker nodes, run the command:
$ sudo systemctl start rke2-agent
Check if the respective nodes are back and uncordon them.
$ kubectl get nodes $ kubectl uncordon <node>
8 Legal notice #
Copyright © 2006–2024 SUSE LLC and contributors. All rights reserved.
Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or (at your option) version 1.3; with the Invariant Section being this copyright notice and license. A copy of the license version 1.2 is included in the section entitled "GNU Free Documentation License".
SUSE, the SUSE logo and YaST are registered trademarks of SUSE LLC in the United States and other countries. For SUSE trademarks, see https://www.suse.com/company/legal/.
Linux is a registered trademark of Linus Torvalds. All other names or trademarks mentioned in this document may be trademarks or registered trademarks of their respective owners.
Documents published as part of the SUSE Best Practices series have been contributed voluntarily by SUSE employees and third parties. They are meant to serve as examples of how particular actions can be performed. They have been compiled with utmost attention to detail. However, this does not guarantee complete accuracy. SUSE cannot verify that actions described in these documents do what is claimed or whether actions described have unintended consequences. SUSE LLC, its affiliates, the authors, and the translators may not be held liable for possible errors or the consequences thereof.
Below we draw your attention to the license under which the articles are published.
9 GNU Free Documentation License #
Copyright © 2000, 2001, 2002 Free Software Foundation, Inc. 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. Everyone is permitted to copy and distribute verbatim copies of this license document, but changing it is not allowed.
0. PREAMBLE#
The purpose of this License is to make a manual, textbook, or other functional and useful document "free" in the sense of freedom: to assure everyone the effective freedom to copy and redistribute it, with or without modifying it, either commercially or noncommercially. Secondarily, this License preserves for the author and publisher a way to get credit for their work, while not being considered responsible for modifications made by others.
This License is a kind of "copyleft", which means that derivative works of the document must themselves be free in the same sense. It complements the GNU General Public License, which is a copyleft license designed for free software.
We have designed this License in order to use it for manuals for free software, because free software needs free documentation: a free program should come with manuals providing the same freedoms that the software does. But this License is not limited to software manuals; it can be used for any textual work, regardless of subject matter or whether it is published as a printed book. We recommend this License principally for works whose purpose is instruction or reference.
1. APPLICABILITY AND DEFINITIONS#
This License applies to any manual or other work, in any medium, that contains a notice placed by the copyright holder saying it can be distributed under the terms of this License. Such a notice grants a world-wide, royalty-free license, unlimited in duration, to use that work under the conditions stated herein. The "Document", below, refers to any such manual or work. Any member of the public is a licensee, and is addressed as "you". You accept the license if you copy, modify or distribute the work in a way requiring permission under copyright law.
A "Modified Version" of the Document means any work containing the Document or a portion of it, either copied verbatim, or with modifications and/or translated into another language.
A "Secondary Section" is a named appendix or a front-matter section of the Document that deals exclusively with the relationship of the publishers or authors of the Document to the Document’s overall subject (or to related matters) and contains nothing that could fall directly within that overall subject. (Thus, if the Document is in part a textbook of mathematics, a Secondary Section may not explain any mathematics.) The relationship could be a matter of historical connection with the subject or with related matters, or of legal, commercial, philosophical, ethical or political position regarding them.
The "Invariant Sections" are certain Secondary Sections whose titles are designated, as being those of Invariant Sections, in the notice that says that the Document is released under this License. If a section does not fit the above definition of Secondary then it is not allowed to be designated as Invariant. The Document may contain zero Invariant Sections. If the Document does not identify any Invariant Sections then there are none.
The "Cover Texts" are certain short passages of text that are listed, as Front-Cover Texts or Back-Cover Texts, in the notice that says that the Document is released under this License. A Front-Cover Text may be at most 5 words, and a Back-Cover Text may be at most 25 words.
A "Transparent" copy of the Document means a machine-readable copy, represented in a format whose specification is available to the general public, that is suitable for revising the document straightforwardly with generic text editors or (for images composed of pixels) generic paint programs or (for drawings) some widely available drawing editor, and that is suitable for input to text formatters or for automatic translation to a variety of formats suitable for input to text formatters. A copy made in an otherwise Transparent file format whose markup, or absence of markup, has been arranged to thwart or discourage subsequent modification by readers is not Transparent. An image format is not Transparent if used for any substantial amount of text. A copy that is not "Transparent" is called "Opaque".
Examples of suitable formats for Transparent copies include plain ASCII without markup, Texinfo input format, LaTeX input format, SGML or XML using a publicly available DTD, and standard-conforming simple HTML, PostScript or PDF designed for human modification. Examples of transparent image formats include PNG, XCF and JPG. Opaque formats include proprietary formats that can be read and edited only by proprietary word processors, SGML or XML for which the DTD and/or processing tools are not generally available, and the machine-generated HTML, PostScript or PDF produced by some word processors for output purposes only.
The "Title Page" means, for a printed book, the title page itself, plus such following pages as are needed to hold, legibly, the material this License requires to appear in the title page. For works in formats which do not have any title page as such, "Title Page" means the text near the most prominent appearance of the work’s title, preceding the beginning of the body of the text.
A section "Entitled XYZ" means a named subunit of the Document whose title either is precisely XYZ or contains XYZ in parentheses following text that translates XYZ in another language. (Here XYZ stands for a specific section name mentioned below, such as "Acknowledgements", "Dedications", "Endorsements", or "History".) To "Preserve the Title" of such a section when you modify the Document means that it remains a section "Entitled XYZ" according to this definition.
The Document may include Warranty Disclaimers next to the notice which states that this License applies to the Document. These Warranty Disclaimers are considered to be included by reference in this License, but only as regards disclaiming warranties: any other implication that these Warranty Disclaimers may have is void and has no effect on the meaning of this License.
2. VERBATIM COPYING#
You may copy and distribute the Document in any medium, either commercially or noncommercially, provided that this License, the copyright notices, and the license notice saying this License applies to the Document are reproduced in all copies, and that you add no other conditions whatsoever to those of this License. You may not use technical measures to obstruct or control the reading or further copying of the copies you make or distribute. However, you may accept compensation in exchange for copies. If you distribute a large enough number of copies you must also follow the conditions in section 3.
You may also lend copies, under the same conditions stated above, and you may publicly display copies.
3. COPYING IN QUANTITY#
If you publish printed copies (or copies in media that commonly have printed covers) of the Document, numbering more than 100, and the Document’s license notice requires Cover Texts, you must enclose the copies in covers that carry, clearly and legibly, all these Cover Texts: Front-Cover Texts on the front cover, and Back-Cover Texts on the back cover. Both covers must also clearly and legibly identify you as the publisher of these copies. The front cover must present the full title with all words of the title equally prominent and visible. You may add other material on the covers in addition. Copying with changes limited to the covers, as long as they preserve the title of the Document and satisfy these conditions, can be treated as verbatim copying in other respects.
If the required texts for either cover are too voluminous to fit legibly, you should put the first ones listed (as many as fit reasonably) on the actual cover, and continue the rest onto adjacent pages.
If you publish or distribute Opaque copies of the Document numbering more than 100, you must either include a machine-readable Transparent copy along with each Opaque copy, or state in or with each Opaque copy a computer-network location from which the general network-using public has access to download using public-standard network protocols a complete Transparent copy of the Document, free of added material. If you use the latter option, you must take reasonably prudent steps, when you begin distribution of Opaque copies in quantity, to ensure that this Transparent copy will remain thus accessible at the stated location until at least one year after the last time you distribute an Opaque copy (directly or through your agents or retailers) of that edition to the public.
It is requested, but not required, that you contact the authors of the Document well before redistributing any large number of copies, to give them a chance to provide you with an updated version of the Document.
4. MODIFICATIONS#
You may copy and distribute a Modified Version of the Document under the conditions of sections 2 and 3 above, provided that you release the Modified Version under precisely this License, with the Modified Version filling the role of the Document, thus licensing distribution and modification of the Modified Version to whoever possesses a copy of it. In addition, you must do these things in the Modified Version:
Use in the Title Page (and on the covers, if any) a title distinct from that of the Document, and from those of previous versions (which should, if there were any, be listed in the History section of the Document). You may use the same title as a previous version if the original publisher of that version gives permission.
List on the Title Page, as authors, one or more persons or entities responsible for authorship of the modifications in the Modified Version, together with at least five of the principal authors of the Document (all of its principal authors, if it has fewer than five), unless they release you from this requirement.
State on the Title page the name of the publisher of the Modified Version, as the publisher.
Preserve all the copyright notices of the Document.
Add an appropriate copyright notice for your modifications adjacent to the other copyright notices.
Include, immediately after the copyright notices, a license notice giving the public permission to use the Modified Version under the terms of this License, in the form shown in the Addendum below.
Preserve in that license notice the full lists of Invariant Sections and required Cover Texts given in the Document’s license notice.
Include an unaltered copy of this License.
Preserve the section Entitled "History", Preserve its Title, and add to it an item stating at least the title, year, new authors, and publisher of the Modified Version as given on the Title Page. If there is no section Entitled "History" in the Document, create one stating the title, year, authors, and publisher of the Document as given on its Title Page, then add an item describing the Modified Version as stated in the previous sentence.
Preserve the network location, if any, given in the Document for public access to a Transparent copy of the Document, and likewise the network locations given in the Document for previous versions it was based on. These may be placed in the "History" section. You may omit a network location for a work that was published at least four years before the Document itself, or if the original publisher of the version it refers to gives permission.
For any section Entitled "Acknowledgements" or "Dedications", Preserve the Title of the section, and preserve in the section all the substance and tone of each of the contributor acknowledgements and/or dedications given therein.
Preserve all the Invariant Sections of the Document, unaltered in their text and in their titles. Section numbers or the equivalent are not considered part of the section titles.
Delete any section Entitled "Endorsements". Such a section may not be included in the Modified Version.
Do not retitle any existing section to be Entitled "Endorsements" or to conflict in title with any Invariant Section.
Preserve any Warranty Disclaimers.
If the Modified Version includes new front-matter sections or appendices that qualify as Secondary Sections and contain no material copied from the Document, you may at your option designate some or all of these sections as invariant. To do this, add their titles to the list of Invariant Sections in the Modified Version’s license notice. These titles must be distinct from any other section titles.
You may add a section Entitled "Endorsements", provided it contains nothing but endorsements of your Modified Version by various parties—for example, statements of peer review or that the text has been approved by an organization as the authoritative definition of a standard.
You may add a passage of up to five words as a Front-Cover Text, and a passage of up to 25 words as a Back-Cover Text, to the end of the list of Cover Texts in the Modified Version. Only one passage of Front-Cover Text and one of Back-Cover Text may be added by (or through arrangements made by) any one entity. If the Document already includes a cover text for the same cover, previously added by you or by arrangement made by the same entity you are acting on behalf of, you may not add another; but you may replace the old one, on explicit permission from the previous publisher that added the old one.
The author(s) and publisher(s) of the Document do not by this License give permission to use their names for publicity for or to assert or imply endorsement of any Modified Version.
5. COMBINING DOCUMENTS#
You may combine the Document with other documents released under this License, under the terms defined in section 4 above for modified versions, provided that you include in the combination all of the Invariant Sections of all of the original documents, unmodified, and list them all as Invariant Sections of your combined work in its license notice, and that you preserve all their Warranty Disclaimers.
The combined work need only contain one copy of this License, and multiple identical Invariant Sections may be replaced with a single copy. If there are multiple Invariant Sections with the same name but different contents, make the title of each such section unique by adding at the end of it, in parentheses, the name of the original author or publisher of that section if known, or else a unique number. Make the same adjustment to the section titles in the list of Invariant Sections in the license notice of the combined work.
In the combination, you must combine any sections Entitled "History" in the various original documents, forming one section Entitled "History"; likewise combine any sections Entitled "Acknowledgements", and any sections Entitled "Dedications". You must delete all sections Entitled "Endorsements".
6. COLLECTIONS OF DOCUMENTS#
You may make a collection consisting of the Document and other documents released under this License, and replace the individual copies of this License in the various documents with a single copy that is included in the collection, provided that you follow the rules of this License for verbatim copying of each of the documents in all other respects.
You may extract a single document from such a collection, and distribute it individually under this License, provided you insert a copy of this License into the extracted document, and follow this License in all other respects regarding verbatim copying of that document.
7. AGGREGATION WITH INDEPENDENT WORKS#
A compilation of the Document or its derivatives with other separate and independent documents or works, in or on a volume of a storage or distribution medium, is called an "aggregate" if the copyright resulting from the compilation is not used to limit the legal rights of the compilation’s users beyond what the individual works permit. When the Document is included in an aggregate, this License does not apply to the other works in the aggregate which are not themselves derivative works of the Document.
If the Cover Text requirement of section 3 is applicable to these copies of the Document, then if the Document is less than one half of the entire aggregate, the Document’s Cover Texts may be placed on covers that bracket the Document within the aggregate, or the electronic equivalent of covers if the Document is in electronic form. Otherwise they must appear on printed covers that bracket the whole aggregate.
8. TRANSLATION#
Translation is considered a kind of modification, so you may distribute translations of the Document under the terms of section 4. Replacing Invariant Sections with translations requires special permission from their copyright holders, but you may include translations of some or all Invariant Sections in addition to the original versions of these Invariant Sections. You may include a translation of this License, and all the license notices in the Document, and any Warranty Disclaimers, provided that you also include the original English version of this License and the original versions of those notices and disclaimers. In case of a disagreement between the translation and the original version of this License or a notice or disclaimer, the original version will prevail.
If a section in the Document is Entitled "Acknowledgements", "Dedications", or "History", the requirement (section 4) to Preserve its Title (section 1) will typically require changing the actual title.
9. TERMINATION#
You may not copy, modify, sublicense, or distribute the Document except as expressly provided for under this License. Any other attempt to copy, modify, sublicense or distribute the Document is void, and will automatically terminate your rights under this License. However, parties who have received copies, or rights, from you under this License will not have their licenses terminated so long as such parties remain in full compliance.
10. FUTURE REVISIONS OF THIS LICENSE#
The Free Software Foundation may publish new, revised versions of the GNU Free Documentation License from time to time. Such new versions will be similar in spirit to the present version, but may differ in detail to address new problems or concerns. See http://www.gnu.org/copyleft/.
Each version of the License is given a distinguishing version number. If the Document specifies that a particular numbered version of this License "or any later version" applies to it, you have the option of following the terms and conditions either of that specified version or of any later version that has been published (not as a draft) by the Free Software Foundation. If the Document does not specify a version number of this License, you may choose any version ever published (not as a draft) by the Free Software Foundation.
ADDENDUM: How to use this License for your documents#
Copyright (c) YEAR YOUR NAME. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation; with no Invariant Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license is included in the section entitled “GNU Free Documentation License”.
If you have Invariant Sections, Front-Cover Texts and Back-Cover Texts, replace the “ with…Texts.” line with this:
with the Invariant Sections being LIST THEIR TITLES, with the Front-Cover Texts being LIST, and with the Back-Cover Texts being LIST.
If you have Invariant Sections without Cover Texts, or some other combination of the three, merge those two alternatives to suit the situation.
If your document contains nontrivial examples of program code, we recommend releasing these examples in parallel under your choice of free software license, such as the GNU General Public License, to permit their use in free software.