Introduction to SUSE AI
- WHAT?
SUSE AI is a locally hosted AI service.
- WHY?
To learn more about AI and the benefits of running an AI service inside your company network instead of relying on public cloud AI solutions.
- EFFORT
To understand the basics of AI and benefits of using AI service locally requires less than 30 minutes of your time.
- GOAL
To make you realize that SUSE AI is the right choice to run local and secure AI workloads.
1 Introduction #
This topic describes what SUSE AI (SUSE AI) is, of which components it consists, and what the benefits of using SUSE AI are.
1.1 What is SUSE AI? #
The SUSE AI service is a locally hosted generative AI. It provides secure auditable AI capabilities and allows organizations to deploy AI models on their own infrastructure using SUSE Linux Micro. The service ensures complete control over data and compliance with regulatory requirements.
1.2 What are typical scenarios for using SUSE AI #
After you deploy and configure SUSE AI inside your company and create a language model based on your documentation, it can:
Implement AI-driven chatbots to handle customer inquiries, providing continuous support and reducing the load on human agents.
Generate reports and summaries on business performance or sales metrics with minimal manual input.
Automate the generation of blog posts or social media content, ensuring consistency and saving time.
Generate personalized e-mail content tailored to individual customer segments.
Offer personalized product or content recommendations based on customer preferences.
Forecast trends, customer behaviors, and market shifts, enabling more informed strategic decisions.
1.3 How does SUSE AI work? #
This section describes individual components of SUSE AI and what happens after you query the AI-driven chatbot.
1.3.1 Structure of SUSE AI #
SUSE AI is designed to run on a cluster of nodes. It consists of two separate components—the SUSE AI Foundation and the SUSE AI Apps.
The SUSE AI Foundation includes:
SUSE Linux Micro as the underlying operating system with the optional NVIDIA driver installed.
RKE2 cluster managed by Rancher Manager ensuring container and application lifecycle management.
NVIDIA GPU Operator to utilize the aNVIDIA GPU computing power and capabilities for processing AI-related tasks.
NeuVector for security and compliance.
providing advanced performance and data monitoring.
The SUSE AI Apps collection includes:
- Ollama (https://ollama.com)
A platform that simplifies installation and management of large language models (LLM) on local devices.
- OpenWebUI (https://openwebui.com)
An extensible Web user interface for the Ollama LLM runner.
- Milvus (https://milvus.io)
A vector database built for generative AI applications with minimal performance loss.
1.3.2 Processing queries by an AI-driven chatbot #
When you ask AI a question, several processes happen in the background to generate a response.
Input processing. The AI first processes the text of your question to understand its meaning. This involves identifying the subject, intent and any details or context provided. This process is called Natural Language Understanding (NLU).
Contextual analysis. If you are interacting with AI in a session where you have already asked previous questions, the AI considers the context of the conversation. This results in more relevant and coherent answers.
Knowledge retrieval. The AI retrieves information from its pre-trained knowledge base. This database includes facts, data and concepts that the AI has been trained on. If the AI has access to real-time data, it can search for the latest information online to provide an up-to-date response.
Response generation. Using natural language generation (NLG) techniques, the AI constructs a coherent and grammatically correct response based on the information it retrieved.
Output. The AI delivers the response to you in a human-readable format. This might be a straightforward answer, a detailed explanation or a step-by-step guide, depending on the complexity of your question.
Feedback Loop (optional). In specific AI systems, your feedback or follow-up questions can help refine future responses, allowing the AI to improve its answers over time.
1.4 Benefits of using SUSE AI #
Running the SUSE AI service brings the following benefits:
Non-local AI services raise concerns about data security and accessibility. Locally hosted AI, however, increases data privacy, security and compliance with regulatory standards.
SUSE AI provides a user-friendly interface for managing and deploying AI models.
Pre-built NVIDIA drivers and operator provide increased AI performance.
SUSE AI components are dynamically scalable, from the Web user interface to the underlying data store.
SUSE AI cares about security—it uses TLS certificates for Web UI and API access. Data stored in a database are encrypted as well.
2 Legal Notice #
Copyright© 2006–2024 SUSE LLC and contributors. All rights reserved.
Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or (at your option) version 1.3; with the Invariant Section being this copyright notice and license. A copy of the license version 1.2 is included in the section entitled “GNU Free Documentation License”.
For SUSE trademarks, see https://www.suse.com/company/legal/. All other third-party trademarks are the property of their respective owners. Trademark symbols (®, ™ etc.) denote trademarks of SUSE and its affiliates. Asterisks (*) denote third-party trademarks.
All information found in this book has been compiled with utmost attention to detail. However, this does not guarantee complete accuracy. Neither SUSE LLC, its affiliates, the authors, nor the translators shall be held liable for possible errors or the consequences thereof.