- Promote v1.17.0-rc.2 to v1.17.0
- Fix bug when using just-in-time CDI spec generation
- Check for valid paths in create-symlinks hook
- Fix bug in locating libcuda.so from ldcache
- Fix bug in sorting of symlink chain
- Remove unsupported print-ldcache command
- Remove csv-filename support from create-symlinks
- Fallback to
crio-status
ifcrio status
does not work when configuring the crio runtime
- Allow IMEX channels to be requested as volume mounts
- Fix typo in error message
- Add disable-imex-channel-creation feature flag
- Add -z,lazy to LDFLAGS
- Add imex channels to management CDI spec
- Add support to fetch current container runtime config from the command line.
- Add creation of select driver symlinks to CDI spec generation.
- Remove support for config overrides when configuring runtimes.
- Skip explicit creation of libnvidia-allocator.so.1 symlink
- Add vdpau as as a driver library search path.
- Add support for using libnvsandboxutils to generate CDI specifications.
- Allow opt-in features to be selected when deploying the toolkit-container.
- Bump CUDA base image version to 12.6.2
- Remove support for config overrides when configuring runtimes.
- Add no-create-imex-channels command line option.
- Exclude libnvidia-allocator from graphics mounts. This fixes a bug that leaks mounts when a container is started with bi-directional mount propagation.
- Use empty string for default runtime-config-override. This removes a redundant warning for runtimes (e.g. Docker) where this is not applicable.
- Bump CUDA base image version to 12.6.0
- Add no-gsp-firmware command line option
- Add no-fabricmanager command line option
- Add no-persistenced command line option
- Skip directories and symlinks when mounting libraries.
- Fix bug with processing errors during CDI spec generation for MIG devices
- Promote v1.16.0-rc.2 to v1.16.0
- Bump CUDA base image version to 12.5.1
- Use relative path to locate driver libraries
- Add RelativeToRoot function to Driver
- Inject additional libraries for full X11 functionality
- Extract options from default runtime if runc does not exist
- Avoid using map pointers as maps are always passed by reference
- Reduce logging for the NVIDIA Container runtime
- Fix bug in argument parsing for logger creation
-
Support vulkan ICD files directly in a driver root. This allows for the discovery of vulkan files in GKE driver installations.
-
Increase priority of ld.so.conf.d config file injected into container. This ensures that injected libraries are preferred over libraries present in the container.
-
Set default CDI spec permissions to 644. This fixes permission issues when using the
nvidia-ctk cdi transform
functions. -
Add
dev-root
option tonvidia-ctk system create-device-nodes
command. -
Fix location of
libnvidia-ml.so.1
when a non-standard driver root is used. This enabled CDI spec generation when using the driver container on a host. -
Recalculate minimum required CDI spec version on save.
-
Move
nvidia-ctk hook
commands to a separatenvidia-cdi-hook
binary. The same subcommands are supported. -
Use
:
as annvidia-ctk config --set
list separator. This fixes a bug when trying to set config options that are lists. -
[toolkit-container] Bump CUDA base image version to 12.5.0
-
[toolkit-container] Allow the path to
toolkit.pid
to be specified directly. -
[toolkit-container] Remove provenance information from image manifests.
-
[toolkit-container] Add
dev-root
option when configuring the toolkit. This adds support for GKE driver installations.
-
Remove
nvidia-container-runtime
andnvidia-docker2
packages. -
Use
XDG_DATA_DIRS
environment variable when locating config files such as graphics config files. -
Add support for v0.7.0 Container Device Interface (CDI) specification.
-
Add
--config-search-path
option tonvidia-ctk cdi generate
command. These paths are used when locating driver files such as graphics config files. -
Use D3DKMTEnumAdapters3 to enumerate adpaters on WSL2 if available.
-
Add support for v1.2.0 OCI Runtime specification.
-
Explicitly set
NVIDIA_VISIBLE_DEVICES=void
in generated CDI specifications. This prevents the NVIDIA Container Runtime from making additional modifications. -
[libnvidia-container] Use D3DKMTEnumAdapters3 to enumerate adpaters on WSL2 if available.
-
[toolkit-container] Bump CUDA base image version to 12.4.1
-
Add a
--spec-dir
option to thenvidia-ctk cdi generate
command. This allows specs outside of/etc/cdi
and/var/run/cdi
to be processed. -
Add support for extracting device major number from
/proc/devices
ifnvidia
is used as a device name overnvidia-frontend
. -
Allow multiple device naming strategies for
nvidia-ctk cdi generate
command. This allows a single CDI spec to be generated that includes GPUs by index and UUID. -
Set the default
--device-name-strategy
for thenvidia-ctk cdi generate
command to[index, uuid]
. -
Remove
libnvidia-container0
jetpack dependency included for legacy Tegra-based systems. -
Add
NVIDIA_VISIBLE_DEVICES=void
to generated CDI specifications. -
[toolkit-container] Remove centos7 image. The ubi8 image can be used on all RPM-based platforms.
-
[toolkit-container] Bump CUDA base image version to 12.3.2
- Fix bug in
nvidia-ctk hook update-ldcache
where default--ldconfig-path
value was not applied.
-
Extend the
runtime.nvidia.com/gpu
CDI kind to support full-GPUs and MIG devices specified by index or UUID. -
Fix bug when specifying
--dev-root
for Tegra-based systems. -
Log explicitly requested runtime mode.
-
Remove package dependency on libseccomp.
-
Added detection of libnvdxgdmal.so.1 on WSL2
-
Use devRoot to resolve MIG device nodes.
-
Fix bug in determining default nvidia-container-runtime.user config value on SUSE-based systems.
-
Add
crun
to the list of configured low-level runtimes. -
Added support for
--ldconfig-path
tonvidia-ctk cdi generate
command. -
Fix
nvidia-ctk runtime configure --cdi.enabled
for Docker. -
Add discovery of the GDRCopy device (
gdrdrv
) if theNVIDIA_GDRCOPY
environment variable of the container is set toenabled
-
[toolkit-container] Bump CUDA base image version to 12.3.1.
-
Skip update of ldcache in containers without ldconfig. The .so.SONAME symlinks are still created.
-
Normalize ldconfig path on use. This automatically adjust the ldconfig setting applied to ldconfig.real on systems where this exists.
-
Include
nvidia/nvoptix.bin
in list of graphics mounts. -
Include
vulkan/icd.d/nvidia_layers.json
in list of graphics mounts. -
Add support for
--library-search-paths
tonvidia-ctk cdi generate
command. -
Add support for injecting /dev/nvidia-nvswitch* devices if the NVIDIA_NVSWITCH=enabled envvar is specified.
-
Added support for
nvidia-ctk runtime configure --enable-cdi
for thedocker
runtime. Note that this requires Docker >= 25. -
Fixed bug in
nvidia-ctk config
command when using--set
. The types of applied config options are now applied correctly. -
Add
--relative-to
option tonvidia-ctk transform root
command. This controls whether the root transformation is applied to host or container paths. -
Added automatic CDI spec generation when the
runtime.nvidia.com/gpu=all
device is requested by a container. -
[libnvidia-container] Fix device permission check when using cgroupv2 (fixes #227)
- [toolkit-container] Bump CUDA base image version to 12.2.2.
- Fix bug on Tegra-based systems where symlinks were not created in containers.
- Add --csv.ignore-pattern command line option to nvidia-ctk cdi generate command.
-
Fixed bug where contents of
/etc/nvidia-container-runtime/config.toml
is ignored by the NVIDIA Container Runtime Hook. -
[libnvidia-container] Use libelf.so on RPM-based systems due to removed mageia repositories hosting pmake and bmake.
- Promote v1.14.0-rc.3 to v1.14.0
-
Added support for generating OCI hook JSON file to
nvidia-ctk runtime configure
command. -
Remove installation of OCI hook JSON from RPM package.
-
Refactored config for
nvidia-container-runtime-hook
. -
Added a
nvidia-ctk config
command which supports setting config options using a--set
flag. -
Added
--library-search-path
option tonvidia-ctk cdi generate
command incsv
mode. This allows folders where libraries are located to be specified explicitly. -
Updated go-nvlib to support devices which are not present in the PCI device database. This allows the creation of dev/char symlinks on systems with such devices installed.
-
Added
UsesNVGPUModule
info function for more robust platform detection. This is required on Tegra-based systems where libnvidia-ml.so is also supported. -
[toolkit-container] Set
NVIDIA_VISIBLE_DEVICES=void
to prevent injection of NVIDIA devices and drivers into the NVIDIA Container Toolkit container.
-
Fix bug causing incorrect nvidia-smi symlink to be created on WSL2 systems with multiple driver roots.
-
Remove dependency on coreutils when installing package on RPM-based systems.
-
Create output folders if required when running
nvidia-ctk runtime configure
-
Generate default config as post-install step.
-
Added support for detecting GSP firmware at custom paths when generating CDI specifications.
-
Added logic to skip the extraction of image requirements if
NVIDIA_DISABLE_REQUIRES
is set totrue
. -
[libnvidia-container] Include Shared Compiler Library (libnvidia-gpucomp.so) in the list of compute libaries.
-
[toolkit-container] Ensure that common envvars have higher priority when configuring the container engines.
-
[toolkit-container] Bump CUDA base image version to 12.2.0.
-
[toolkit-container] Remove installation of nvidia-experimental runtime. This is superceded by the NVIDIA Container Runtime in CDI mode.
-
Add support for updating containerd configs to the
nvidia-ctk runtime configure
command. -
Create file in
etc/ld.so.conf.d
with permissions644
to support non-root containers. -
Generate CDI specification files with
644
permissions to allow rootless applications (e.g. podman) -
Add
nvidia-ctk cdi list
command to show the known CDI devices. -
Add support for generating merged devices (e.g.
all
device) to the nvcdi API. -
Use . pattern to locate libcuda.so when generating a CDI specification to support platforms where a patch version is not specified.
-
Update go-nvlib to skip devices that are not MIG capable when generating CDI specifications.
-
Add
nvidia-container-runtime-hook.path
config option to specify NVIDIA Container Runtime Hook path explicitly. -
Fix bug in creation of
/dev/char
symlinks by failing operation if kernel modules are not loaded. -
Add option to load kernel modules when creating device nodes
-
Add option to create device nodes when creating
/dev/char
symlinks -
[libnvidia-container] Support OpenSSL 3 with the Encrypt/Decrypt library
-
[toolkit-container] Allow same envars for all runtime configs
- Update
update-ldcache
hook to only update ldcache if it exists. - Update
update-ldcache
hook to create/etc/ld.so.conf.d
folder if it doesn't exist. - Fix failure when libcuda cannot be located during XOrg library discovery.
- Fix CDI spec generation on systems that use
/etc/alternatives
(e.g. Debian)
- Promote 1.13.0-rc.3 to 1.13.0
-
Only initialize NVML for modes that require it when runing
nvidia-ctk cdi generate
. -
Prefer /run over /var/run when locating nvidia-persistenced and nvidia-fabricmanager sockets.
-
Fix the generation of CDI specifications for management containers when the driver libraries are not in the LDCache.
-
Add transformers to deduplicate and simplify CDI specifications.
-
Generate a simplified CDI specification by default. This means that entities in the common edits in a spec are not included in device definitions.
-
Also return an error from the nvcdi.New constructor instead of panicing.
-
Detect XOrg libraries for injection and CDI spec generation.
-
Add
nvidia-ctk system create-device-nodes
command to create control devices. -
Add
nvidia-ctk cdi transform
command to apply transforms to CDI specifications. -
Add
--vendor
and--class
options tonvidia-ctk cdi generate
-
[libnvidia-container] Fix segmentation fault when RPC initialization fails.
-
[libnvidia-container] Build centos variants of the NVIDIA Container Library with static libtirpc v1.3.2.
-
[libnvidia-container] Remove make targets for fedora35 as the centos8 packages are compatible.
-
[toolkit-container] Add
nvidia-container-runtime.modes.cdi.annotation-prefixes
config option that allows the CDI annotation prefixes that are read to be overridden. -
[toolkit-container] Create device nodes when generating CDI specification for management containers.
-
[toolkit-container] Add
nvidia-container-runtime.runtimes
config option to set the low-level runtime for the NVIDIA Container Runtime
- Don't fail chmod hook if paths are not injected
- Only create
by-path
symlinks if CDI devices are actually requested. - Fix possible blank
nvidia-ctk
path in generated CDI specifications - Fix error in postun scriplet on RPM-based systems
- Only check
NVIDIA_VISIBLE_DEVICES
for environment variables if no annotations are specified. - Add
cdi.default-kind
config option for constructing fully-qualified CDI device names in CDI mode - Add support for
accept-nvidia-visible-devices-envvar-unprivileged
config setting in CDI mode - Add
nvidia-container-runtime-hook.skip-mode-detection
config option to bypass mode detection. This allowslegacy
andcdi
mode, for example, to be used at the same time. - Add support for generating CDI specifications for GDS and MOFED devices
- Ensure CDI specification is validated on save when generating a spec
- Rename
--discovery-mode
argument to--mode
fornvidia-ctk cdi generate
- [libnvidia-container] Fix segfault on WSL2 systems
- [toolkit-container] Add
--cdi-enabled
flag to toolkit config - [toolkit-container] Install
nvidia-ctk
from toolkit container - [toolkit-container] Use installed
nvidia-ctk
path in NVIDIA Container Toolkit config - [toolkit-container] Bump CUDA base images to 12.1.0
- [toolkit-container] Set
nvidia-ctk
path in the - [toolkit-container] Add
cdi.k8s.io/*
to set of allowed annotations in containerd config - [toolkit-container] Generate CDI specification for use in management containers
- [toolkit-container] Install experimental runtime as
nvidia-container-runtime.experimental
instead ofnvidia-container-runtime-experimental
- [toolkit-container] Install and configure mode-specific runtimes for
cdi
andlegacy
modes
- Include MIG-enabled devices as GPUs when generating CDI specification
- Fix missing NVML symbols when running
nvidia-ctk
on some platforms [#49] - Add CDI spec generation for WSL2-based systems to
nvidia-ctk cdi generate
command - Add
auto
mode tonvidia-ctk cdi generate
command to automatically detect a WSL2-based system over a standard NVML-based system. - Add mode-specific (
.cdi
and.legacy
) NVIDIA Container Runtime binaries for use in the GPU Operator - Discover all
gsb*.bin
GSP firmware files when generating CDI specification. - Align
.deb
and.rpm
release candidate package versions - Remove
fedora35
packaging targets - [libnvidia-container] Include all
gsp*.bin
firmware files if present - [libnvidia-container] Align
.deb
and.rpm
release candidate package versions - [libnvidia-container] Remove
fedora35
packaging targets - [toolkit-container] Install
nvidia-container-toolkit-operator-extensions
package for mode-specific executables. - [toolkit-container] Allow
nvidia-container-runtime.mode
to be set when configuring the NVIDIA Container Toolkit
- Promote
v1.12.0-rc.5
tov1.12.0
- Rename
nvidia cdi generate
--root
flag to--driver-root
to better indicate intent - [libnvidia-container] Add nvcubins.bin to DriverStore components under WSL2
- [toolkit-container] Bump CUDA base images to 12.0.1
- Fix bug here the
nvidia-ctk
path was not properly resolved. This causes failures to run containers when the runtime is configured incsv
mode or if theNVIDIA_DRIVER_CAPABILITIES
includesgraphics
ordisplay
(e.g.all
).
- Generate a minimum CDI spec version for improved compatibility.
- Add
--device-name-strategy
options to thenvidia-ctk cdi generate
command that can be used to control how device names are constructed. - Set default for CDI device name generation to
index
to generate device names such asnvidia.com/gpu=0
ornvidia.com/gpu=1:0
by default.
- Don't fail if by-path symlinks for DRM devices do not exist
- Replace the --json flag with a --format [json|yaml] flag for the nvidia-ctk cdi generate command
- Ensure that the CDI output folder is created if required
- When generating a CDI specification use a blank host path for devices to ensure compatibility with the v0.4.0 CDI specification
- Add injection of Wayland JSON files
- Add GSP firmware paths to generated CDI specification
- Add --root flag to nvidia-ctk cdi generate command
- Inject Direct Rendering Manager (DRM) devices into a container using the NVIDIA Container Runtime
- Improve logging of errors from the NVIDIA Container Runtime
- Improve CDI specification generation to support rootless podman
- Use
nvidia-ctk cdi generate
to generate CDI specifications instead ofnvidia-ctk info generate-cdi
- [libnvidia-container] Skip creation of existing files when these are already mounted
- Add support for multiple Docker Swarm resources
- Improve injection of Vulkan configurations and libraries
- Add
nvidia-ctk info generate-cdi
command to generated CDI specification for available devices - [libnvidia-container] Include NVVM compiler library in compute libs
- Promote v1.11.0-rc.3 to v1.11.0
- Build fedora35 packages
- Introduce an
nvidia-container-toolkit-base
package for better dependency management - Fix removal of
nvidia-container-runtime-hook
on RPM-based systems - Inject platform files into container on Tegra-based systems
- [toolkit container] Update CUDA base images to 11.7.1
- [libnvidia-container] Preload libgcc_s.so.1 on arm64 systems
- Allow
accept-nvidia-visible-devices-*
config options to be set by toolkit container - [libnvidia-container] Fix bug where LDCache was not updated when the
--no-pivot-root
option was specified
- Add discovery of GPUDirect Storage (
nvidia-fs*
) devices if theNVIDIA_GDS
environment variable of the container is set toenabled
- Add discovery of MOFED Infiniband devices if the
NVIDIA_MOFED
environment variable of the container is set toenabled
- Fix bug in CSV mode where libraries listed as
sym
entries in mount specification are not added to the LDCache. - Rename
nvidia-container-toolkit
executable tonvidia-container-runtime-hook
and createnvidia-container-toolkit
as a symlink tonvidia-container-runtime-hook
instead. - Add
nvidia-ctk runtime configure
command to configure the Docker config file (e.g./etc/docker/daemon.json
) for use with the NVIDIA Container Runtime.
- Promote v1.10.0-rc.3 to v1.10.0
- Use default config instead of raising an error if config file cannot be found
- Ignore NVIDIA_REQUIRE_JETPACK* environment variables for requirement checks
- Fix bug in detection of Tegra systems where
/sys/devices/soc0/family
is ignored - Fix bug where links to devices were detected as devices
- [libnvida-container] Fix bug introduced when adding libcudadebugger.so to list of libraries
- Add support for NVIDIA_REQUIRE_* checks for cuda version and arch to csv mode
- Switch to debug logging to reduce log verbosity
- Support logging to logs requested in command line
- Fix bug when launching containers with relative root path (e.g. using containerd)
- Allow low-level runtime path to be set explicitly as nvidia-container-runtime.runtimes option
- Fix failure to locate low-level runtime if PATH envvar is unset
- Replace experimental option for NVIDIA Container Runtime with nvidia-container-runtime.mode = csv option
- Use csv as default mode on Tegra systems without NVML
- Add --version flag to all CLIs
- [libnvidia-container] Bump libtirpc to 1.3.2
- [libnvidia-container] Fix bug when running host ldconfig using glibc compiled with a non-standard prefix
- [libnvidia-container] Add libcudadebugger.so to list of compute libraries
- Include nvidia-ctk CLI in installed binaries
- Add experimental option to NVIDIA Container Runtime
- [libnvidia-container] Add additional check for Tegra in /sys/.../family file in CLI
- [libnvidia-container] Update jetpack-specific CLI option to only load Base CSV files by default
- [libnvidia-container] Fix bug (from 1.8.0) when mounting GSP firmware into containers without /lib to /usr/lib symlinks
- [libnvidia-container] Update nvml.h to CUDA 11.6.1 nvML_DEV 11.6.55
- [libnvidia-container] Update switch statement to include new brands from latest nvml.h
- [libnvidia-container] Process all --require flags on Jetson platforms
- [libnvidia-container] Fix long-standing issue with running ldconfig on Debian systems
- [libnvidia-container] Fix bug in determining cgroup root when running in nested containers
- [libnvidia-container] Fix permission issue when determining cgroup version
- Promote 1.8.0-rc.2-1 to 1.8.0
- Remove support for building amazonlinux1 packages
- [libnvidia-container] Add support for cgroupv2
- Release toolkit-container images from nvidia-container-toolkit repository
- Promote 1.7.0-rc.1-1 to 1.7.0
- Bump Golang version to 1.16.4
- Specify containerd runtime type as string in config tools to remove dependency on containerd package
- Add supported-driver-capabilities config option to allow for a subset of all driver capabilities to be specified
- Promote 1.6.0-rc.3-1 to 1.6.0
- Fix unnecessary logging to stderr instead of configured nvidia-container-runtime log file
- Add supported-driver-capabilities config option to the nvidia-container-toolkit
- Move OCI and command line checks for runtime to internal oci package
- Use relative path to OCI specification file (config.json) if bundle path is not specified as an argument to the nvidia-container-runtime
- Add AARCH64 package for Amazon Linux 2
- Include nvidia-container-runtime into nvidia-container-toolkit package
- Fix bug where Docker Swarm device selection is ignored if NVIDIA_VISIBLE_DEVICES is also set
- Improve unit testing by using require package and adding coverage reports
- Remove unneeded go dependencies by running go mod tidy
- Move contents of pkg directory to cmd for CLI tools
- Ensure make binary target explicitly sets GOOS
- Add dependence on libnvidia-container-tools >= 1.4.0
- Add golang check targets to Makefile
- Add Jenkinsfile definition for build targets
- Move docker.mk to docker folder
- Add dependence on libnvidia-container-tools >= 1.3.3
- Ignore NVIDIA_VISIBLE_DEVICES for containers with insufficent privileges
- Add dependence on libnvidia-container-tools >= 1.3.2
- Add 'compute' capability to list of defaults
- Add dependence on libnvidia-container-tools >= 1.3.1
- Promote 1.3.0-rc.2-1 to 1.3.0
- Add dependence on libnvidia-container-tools >= 1.3.0
- 2c180947 Add more tests for new semantics with device list from volume mounts
- 7c003857 Refactor accepting device lists from volume mounts as a boolean
- b50d86c1 Update build system to accept a TAG variable for things like rc.x
- fe65573b Add common CI tests for things like golint, gofmt, unit tests, etc.
- da6fbb34 Revert "Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*"
- a7fb3330 Flip build-all targets to run automatically on merge requests
- 8b248b66 Rename github.com/NVIDIA/container-toolkit to nvidia-container-toolkit
- da36874e Add new config options to pull device list from mounted files instead of ENVVAR
- 4e6e0ed4 Add 'ngx' to list ofall driver capabilities
- 2f4af743 List config.toml as a config file in the RPM SPEC
- 8e0aab46 Fix repo listed in changelog for debian distributions
- 320bb6e4 Update dependence on libnvidia-container to 1.2.0
- 6cfc8097 Update package license to match source license
- e7dc3cbb Fix debian copyright file
- d3aee3e0 Add the 'ngx' driver capability
- c32237f3 Add support for parsing Linux Capabilities for older OCI specs
- d202aded Update dependence to libnvidia-container 1.1.1
- 4e4de762 Update build system to support multi-arch builds
- fcc1d116 Add support for MIG (Multi-Instance GPUs)
- d4ff0416 Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*
- 60f165ad Add no-pivot option to toolkit
- Initial release. Replaces older package nvidia-container-runtime-hook. (Closes: #XXXXXX)