Jump to contentJump to page navigation: previous page [access key p]/next page [access key n]
Applies to SUSE Linux Enterprise Server 15-SP2, K3s 1.21.2

1 Introduction

On the digital transformation journey to a full cloud native landscape, utilization of microservices becomes the main approach with the dominant technology for such container orchestration being Kubernetes.⁠[1] With its large community of developers and abundant features and capabilities, Kubernetes has become the de-facto standard and is included across most container-as-a-service platforms. With all of these technologies in place, both developer and operation teams can effectively deploy, manage and deliver functionality to their end users in a resilient and agile manner.

1.1 Motivation

As developers and organizations continue their journey from simple, containerized microservices towards having these workloads orchestrated and deployed where ever they need, being able to install, monitor and use such Kubernetes infrastructures is a core need. Such deployments, being Cloud Native Computing Foundation (CNCF⁠[2]) conformant and certified⁠[3] are essential for both development and production workloads.

For simplified scenarios, like edge, remote or IoT, this is where K3s leads the industry, being simple and secure.

1.2 Scope

The scope of this document is to provide a general reference implementation of K3s. This can be done in a variety of scenarios to create an edge-oriented, lightweight Kubernetes cluster deployment.

1.3 Audience

This document is intended for IT decision makers, architects, system administrators and technicians who are implementing a flexible, software-defined Kubernetes platform. One should still be familiar with the traditional IT infrastructure pillars — networking, computing and storage — along with the local use cases for sizing, scaling and limitations within each pillars' environments.

Print this page