Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
18 changes: 10 additions & 8 deletions docs/runtime.md
Original file line number Diff line number Diff line change
Expand Up @@ -212,14 +212,16 @@ helm repo add nvidia https://helm.ngc.nvidia.com/nvidia
helm repo update
helm install nvidia-gpu-operator -n nvidia-gpu-operator \
--create-namespace \
--set operator.defaultRuntime=containerd \
--set toolkit.env[0].name=CONTAINERD_CONFIG \
--set toolkit.env[0].value=/etc/k0s/containerd.d/nvidia.toml \
--set toolkit.env[1].name=CONTAINERD_SOCKET \
--set toolkit.env[1].value=/run/k0s/containerd.sock \
--set toolkit.env[2].name=CONTAINERD_RUNTIME_CLASS \
--set toolkit.env[2].value=nvidia \
nvidia/gpu-operator
--set "operator.defaultRuntime=containerd" \
--set "toolkit.env[0].name=CONTAINERD_CONFIG" \
--set "toolkit.env[0].value=/run/k0s/containerd-cri.toml" \
Comment on lines +216 to +217
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I wonder if these should point to the "real" containerd config instead of this k0s managed CRI merged one 🤔 I guess it boils down to what exactly the operator does with this.

This is what the docs say:

CONTAINERD_CONFIG
The path on the host to the top-level containerd config file. By default this will point to /etc/containerd/containerd.toml (the default location for containerd). It should be customized if your containerd installation is not in the default location.

Based on the above, I'd say this should point to /etc/k0s/containerd.toml which in turn imports the CRI config

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I just tried the config and it worked out-of-the box.

--set "toolkit.env[1].name=RUNTIME_DROP_IN_CONFIG" \
--set "toolkit.env[1].value=/etc/k0s/containerd.d/nvidia.toml" \
--set "toolkit.env[2].name=CONTAINERD_SOCKET" \
--set "toolkit.env[2].value=/run/k0s/containerd.sock " \
--set "toolkit.env[3].name=CONTAINERD_RUNTIME_CLASS" \
--set "toolkit.env[3].value=nvidia" \
nvidia/gpu-operator --version v25.10.0
```

With this Helm chart values, NVIDIA GPU operator will deploy both driver and toolkit to the GPU nodes and additionally will configure containerd with NVIDIA specific runtime.
Expand Down
Loading