Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
48 changes: 36 additions & 12 deletions articles/ai-deployment-airgapped.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -83,28 +83,52 @@ include::../glues/ai-ssecurity-intro.adoc[leveloffset=+2]
include::../tasks/Security-installation-airgapped.adoc[leveloffset=+3]
include::../tasks/observability-settingup-ai.adoc[leveloffset=+2]
// AI library

include::../glues/ai-library-intro.adoc[leveloffset=+1]
:override-title: Installation procedure

include::../tasks/AI-deployment-ailibrary-installing.adoc[leveloffset=+2]

include::../tasks/cert-manager-installing.adoc[leveloffset=+2]
// include::../tasks/opensearch-installing.adoc[leveloffset=+2]

include::../tasks/opensearch-installing.adoc[leveloffset=+2]

include::../tasks/milvus-installing.adoc[leveloffset=+2]

include::../tasks/ollama-installing.adoc[leveloffset=+2]

include::../references/ollama-helmchart.adoc[leveloffset=+3]

include::../tasks/owui-installing.adoc[leveloffset=+2]

include::../references/owui-helm-overrides.adoc[leveloffset=+3]

include::../references/owui-helmchart.adoc[leveloffset=+3]
// include::../tasks/vllm-installing.adoc[leveloffset=+2]
// include::../references/vllm-helm-overrides.adoc[leveloffset=+3]
// include::../tasks/mcpo-installing.adoc[leveloffset=+2]
// include::../tasks/pytorch-installing.adoc[leveloffset=+2]
// include::../references/pytorch-helm-overrides.adoc[leveloffset=+3]
// include::../references/pytorch-helmchart.adoc[leveloffset=+3]
// include::../tasks/mlflow-installing.adoc[leveloffset=+2]
// include::../tasks/ai-deployment-ailibrary-deployer.adoc[leveloffset=+2]
// steps after deployment
// :override-title: Steps after the installation is complete
// include::../tasks/AI-deployment-steps-after.adoc[leveloffset=+1]

include::../tasks/vllm-installing.adoc[leveloffset=+2]

include::../references/vllm-helm-overrides.adoc[leveloffset=+3]

include::../tasks/mcpo-installing.adoc[leveloffset=+2]

include::../tasks/pytorch-installing.adoc[leveloffset=+2]

include::../references/pytorch-helm-overrides.adoc[leveloffset=+3]

include::../references/pytorch-helmchart.adoc[leveloffset=+3]

include::../tasks/qdrant-installing.adoc[leveloffset=+2]

include::../references/qdrant-helm-overrides.adoc[leveloffset=+3]

include::../tasks/litellm-installing.adoc[leveloffset=+2]

include::../references/litellm-helm-overrides.adoc[leveloffset=+3]

include::../references/litellm-helmchart.adoc[leveloffset=+3]

include::../tasks/mlflow-installing.adoc[leveloffset=+2]
//include::../tasks/ai-deployment-ailibrary-deployer.adoc[leveloffset=+2]

[appendix]
include::../references/AI-glossary.adoc[leveloffset=+1]
Expand Down
4 changes: 2 additions & 2 deletions concepts/AI-air-gap-stack.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -28,5 +28,5 @@ The following simplified workflow outlines the intended usage:
Transfer the downloaded content to an air-gapped _local_ host and add it as a {zypper} repository to install {nvidia} drivers on _local_ GPU nodes.
. Use `SUSE-AI-get-images.sh` on a _remote_ host to download {docker} images of required {productname} components.
Transfer them to an air-gapped _local_ host.
. Use `SUSE-AI-load-images.sh` to load the transferred {docker} images of {productname} components into a custom _local_ ${docker} image registry.
. Install {ailibrary} components on the _local_ {kube} cluster from the _local_ custom {docker} registry.
. Use `SUSE-AI-load-images.sh` to load the transferred {docker} images of {productname} components into a custom _local_ {docker} image registry.
. Install {ailibrary} applications on the _local_ {kube} cluster from the _local_ custom {docker} registry.
12 changes: 10 additions & 2 deletions references/litellm-helm-overrides.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -18,21 +18,29 @@ endif::[]
:override-abstract!:
:override-title!:

:revdate: 2026-02-10
:revdate: 2026-03-25
:page-revdate: {revdate}

include::../snippets/helm-chart-overrides-intro.adoc[]

ifeval::["{PROF_DEPLOYMENT}" == "standard"]
:imgRegistry: {empty}
endif::[]
ifeval::["{PROF_DEPLOYMENT}" == "airgapped"]
:imgRegistry: imageRegistry: <LOCAL_DOCKER_REGISTRY_URL>:5043
endif::[]

[#litellm-chart-example-basic]
.Basic override file with {postgresql} deployment and master key automatically generated.
====
[source,yaml]
[source,yaml,subs="+attributes"]
----
# litellm_custom_overrides.yaml
global:
imagePullSecrets:
- application-collection <.>
- suse-ai-registry <.>
{imgRegistry}
postgresql:
persistence:
storageClassName: "local-path" <.>
Expand Down
2 changes: 1 addition & 1 deletion references/litellm-helmchart.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -57,4 +57,4 @@ include::../snippets/helm-chart-overrides-intro.adoc[]
| `pdb.annotations` | Extra metadata annotations to add to the PDB | `{}`
| `pdb.labels` | Extra metadata labels to add to the PDB | `{}`

|===
|===
14 changes: 11 additions & 3 deletions references/ollama-helmchart.adoc
Original file line number Diff line number Diff line change
@@ -1,8 +1,15 @@
[#ollama-helmchart]
= Values for the {ollama} {helm} chart
:revdate: 2025-12-22
:revdate: 2026-03-24
:page-revdate: {revdate}

ifeval::["{PROF_DEPLOYMENT}" == "standard"]
:imgRegistry: {empty}
endif::[]
ifeval::["{PROF_DEPLOYMENT}" == "airgapped"]
:imgRegistry: imageRegistry: <LOCAL_DOCKER_REGISTRY_URL>:5043
endif::[]

include::../snippets/helm-chart-overrides-intro.adoc[]

[IMPORTANT]
Expand All @@ -28,11 +35,12 @@ If you do not want to use the {nvidia} GPU, remove the `gpu` section from `ollam

.Basic override file with GPU and two models pulled at startup
====
[source,yaml]
[source,yaml,subs="+attributes"]
----
global:
imagePullSecrets:
- application-collection
{imgRegistry}
ingress:
enabled: false
defaultModel: "gemma:2b"
Expand All @@ -54,7 +62,7 @@ persistentVolume: <.>
enabled: true
storageClass: local-path <.>
----
<.> Without the `persistentVolume` option enabled, changes made to {ollama}--such as downloading other LLM-- are lost when the container is restarted.
<.> Without the `persistentVolume` option enabled, changes made to {ollama}, such as downloading other LLM, are lost when the container is restarted.
<.> Use `local-path` storage only for testing purposes.
For production use, we recommend using a storage solution suitable for persistent storage, such as {sstorage}.
====
Expand Down
Loading