Minimal nvidia-device-plugin container image.
Download this Image
The image is available on cgr.dev
:
docker pull cgr.dev/chainguard/nvidia-device-plugin:latest
Usage
Ensure your environment satisfies the prerequisites.
Jump to Quick Start to learn more.
To test:
$ docker run --rm cgr.dev/chainguard/nvidia-device-plugin --help
NAME:
NVIDIA Device Plugin - NVIDIA device plugin for Kubernetes
USAGE:
nvidia-device-plugin [global options] command [command options] [arguments...]
VERSION:
v0.14.0
commit: e6c111aff19eab995e8d0f4345169e8c310d2f9c
COMMANDS:
help, h Shows a list of commands or help for one command
GLOBAL OPTIONS:
--mig-strategy value the desired strategy for exposing MIG devices on GPUs that support it:
[none | single | mixed] (default: "none") [$MIG_STRATEGY]
--fail-on-init-error fail the plugin if an error is encountered during initialization, otherwise block indefinitely (default: true) [$FAIL_ON_INIT_ERROR]
--nvidia-driver-root value the root path for the NVIDIA driver installation (typical values are '/' or '/run/nvidia/driver') (default: "/") [$NVIDIA_DRIVER_ROOT]
--pass-device-specs pass the list of DeviceSpecs to the kubelet on Allocate() (default: false) [$PASS_DEVICE_SPECS]
--device-list-strategy value the desired strategy for passing the device list to the underlying runtime:
[envvar | volume-mounts | cdi-annotations] (default: "envvar") (accepts multiple inputs) [$DEVICE_LIST_STRATEGY]
--device-id-strategy value the desired strategy for passing device IDs to the underlying runtime:
[uuid | index] (default: "uuid") [$DEVICE_ID_STRATEGY]
--gds-enabled ensure that containers are started with NVIDIA_GDS=enabled (default: false) [$GDS_ENABLED]
--mofed-enabled ensure that containers are started with NVIDIA_MOFED=enabled (default: false) [$MOFED_ENABLED]
--config-file value the path to a config file as an alternative to command line options or environment variables [$CONFIG_FILE]
--cdi-annotation-prefix value the prefix to use for CDI container annotation keys (default: "cdi.k8s.io/") [$CDI_ANNOTATION_PREFIX]
--nvidia-ctk-path value the path to use for the nvidia-ctk in the generated CDI specification (default: "/usr/bin/nvidia-ctk") [$NVIDIA_CTK_PATH]
--container-driver-root value the path where the NVIDIA driver root is mounted in the container; used for generating CDI specifications (default: "/driver-root") [$CONTAINER_DRIVER_ROOT]
--help, -h show help (default: false)
--version, -v print the version (default: false)
To use the Chainguard Images variant, override the values below in your values.yaml
to use with Helm:
image:
repository: cgr.dev/chainguard/nvidia-device-plugin
tag: latest
When using this image with the nvidia/gpu-operator
helm chart bash
is required to run the mounted /bin/entrypoint.sh
script.
For this you should use the latest-dev
variant.
helm repo add nvidia https://helm.ngc.nvidia.com/nvidia
helm repo update
helm install --wait gpu -n gpu-operator --create-namespace nvidia/gpu-operator
Values:
driver:
enabled: false
devicePlugin:
repository: cgr.dev/chainguard
image: nvidia-device-plugin
tag: latest-dev