Signed-off-by: Evan Lezar <elezar@nvidia.com>
25 KiB
NVIDIA Container Toolkit Changelog
v1.17.0
- Promote v1.17.0-rc.2 to v1.17.0
- Fix bug when using just-in-time CDI spec generation
- Check for valid paths in create-symlinks hook
v1.17.0-rc.2
- Fix bug in locating libcuda.so from ldcache
- Fix bug in sorting of symlink chain
- Remove unsupported print-ldcache command
- Remove csv-filename support from create-symlinks
Changes in the Toolkit Container
- Fallback to
crio-status
ifcrio status
does not work when configuring the crio runtime
v1.17.0-rc.1
- Allow IMEX channels to be requested as volume mounts
- Fix typo in error message
- Add disable-imex-channel-creation feature flag
- Add -z,lazy to LDFLAGS
- Add imex channels to management CDI spec
- Add support to fetch current container runtime config from the command line.
- Add creation of select driver symlinks to CDI spec generation.
- Remove support for config overrides when configuring runtimes.
- Skip explicit creation of libnvidia-allocator.so.1 symlink
- Add vdpau as as a driver library search path.
- Add support for using libnvsandboxutils to generate CDI specifications.
Changes in the Toolkit Container
- Allow opt-in features to be selected when deploying the toolkit-container.
- Bump CUDA base image version to 12.6.2
- Remove support for config overrides when configuring runtimes.
Changes in libnvidia-container
- Add no-create-imex-channels command line option.
v1.16.2
- Exclude libnvidia-allocator from graphics mounts. This fixes a bug that leaks mounts when a container is started with bi-directional mount propagation.
- Use empty string for default runtime-config-override. This removes a redundant warning for runtimes (e.g. Docker) where this is not applicable.
Changes in the Toolkit Container
- Bump CUDA base image version to 12.6.0
Changes in libnvidia-container
- Add no-gsp-firmware command line option
- Add no-fabricmanager command line option
- Add no-persistenced command line option
- Skip directories and symlinks when mounting libraries.
v1.16.1
- Fix bug with processing errors during CDI spec generation for MIG devices
v1.16.0
- Promote v1.16.0-rc.2 to v1.16.0
Changes in the Toolkit Container
- Bump CUDA base image version to 12.5.1
v1.16.0-rc.2
- Use relative path to locate driver libraries
- Add RelativeToRoot function to Driver
- Inject additional libraries for full X11 functionality
- Extract options from default runtime if runc does not exist
- Avoid using map pointers as maps are always passed by reference
- Reduce logging for the NVIDIA Container runtime
- Fix bug in argument parsing for logger creation
v1.16.0-rc.1
-
Support vulkan ICD files directly in a driver root. This allows for the discovery of vulkan files in GKE driver installations.
-
Increase priority of ld.so.conf.d config file injected into container. This ensures that injected libraries are preferred over libraries present in the container.
-
Set default CDI spec permissions to 644. This fixes permission issues when using the
nvidia-ctk cdi transform
functions. -
Add
dev-root
option tonvidia-ctk system create-device-nodes
command. -
Fix location of
libnvidia-ml.so.1
when a non-standard driver root is used. This enabled CDI spec generation when using the driver container on a host. -
Recalculate minimum required CDI spec version on save.
-
Move
nvidia-ctk hook
commands to a separatenvidia-cdi-hook
binary. The same subcommands are supported. -
Use
:
as annvidia-ctk config --set
list separator. This fixes a bug when trying to set config options that are lists. -
[toolkit-container] Bump CUDA base image version to 12.5.0
-
[toolkit-container] Allow the path to
toolkit.pid
to be specified directly. -
[toolkit-container] Remove provenance information from image manifests.
-
[toolkit-container] Add
dev-root
option when configuring the toolkit. This adds support for GKE driver installations.
v1.15.0
-
Remove
nvidia-container-runtime
andnvidia-docker2
packages. -
Use
XDG_DATA_DIRS
environment variable when locating config files such as graphics config files. -
Add support for v0.7.0 Container Device Interface (CDI) specification.
-
Add
--config-search-path
option tonvidia-ctk cdi generate
command. These paths are used when locating driver files such as graphics config files. -
Use D3DKMTEnumAdapters3 to enumerate adpaters on WSL2 if available.
-
Add support for v1.2.0 OCI Runtime specification.
-
Explicitly set
NVIDIA_VISIBLE_DEVICES=void
in generated CDI specifications. This prevents the NVIDIA Container Runtime from making additional modifications. -
[libnvidia-container] Use D3DKMTEnumAdapters3 to enumerate adpaters on WSL2 if available.
-
[toolkit-container] Bump CUDA base image version to 12.4.1
v1.15.0-rc.4
-
Add a
--spec-dir
option to thenvidia-ctk cdi generate
command. This allows specs outside of/etc/cdi
and/var/run/cdi
to be processed. -
Add support for extracting device major number from
/proc/devices
ifnvidia
is used as a device name overnvidia-frontend
. -
Allow multiple device naming strategies for
nvidia-ctk cdi generate
command. This allows a single CDI spec to be generated that includes GPUs by index and UUID. -
Set the default
--device-name-strategy
for thenvidia-ctk cdi generate
command to[index, uuid]
. -
Remove
libnvidia-container0
jetpack dependency included for legacy Tegra-based systems. -
Add
NVIDIA_VISIBLE_DEVICES=void
to generated CDI specifications. -
[toolkit-container] Remove centos7 image. The ubi8 image can be used on all RPM-based platforms.
-
[toolkit-container] Bump CUDA base image version to 12.3.2
v1.15.0-rc.3
- Fix bug in
nvidia-ctk hook update-ldcache
where default--ldconfig-path
value was not applied.
v1.15.0-rc.2
-
Extend the
runtime.nvidia.com/gpu
CDI kind to support full-GPUs and MIG devices specified by index or UUID. -
Fix bug when specifying
--dev-root
for Tegra-based systems. -
Log explicitly requested runtime mode.
-
Remove package dependency on libseccomp.
-
Added detection of libnvdxgdmal.so.1 on WSL2
-
Use devRoot to resolve MIG device nodes.
-
Fix bug in determining default nvidia-container-runtime.user config value on SUSE-based systems.
-
Add
crun
to the list of configured low-level runtimes. -
Added support for
--ldconfig-path
tonvidia-ctk cdi generate
command. -
Fix
nvidia-ctk runtime configure --cdi.enabled
for Docker. -
Add discovery of the GDRCopy device (
gdrdrv
) if theNVIDIA_GDRCOPY
environment variable of the container is set toenabled
-
[toolkit-container] Bump CUDA base image version to 12.3.1.
v1.15.0-rc.1
-
Skip update of ldcache in containers without ldconfig. The .so.SONAME symlinks are still created.
-
Normalize ldconfig path on use. This automatically adjust the ldconfig setting applied to ldconfig.real on systems where this exists.
-
Include
nvidia/nvoptix.bin
in list of graphics mounts. -
Include
vulkan/icd.d/nvidia_layers.json
in list of graphics mounts. -
Add support for
--library-search-paths
tonvidia-ctk cdi generate
command. -
Add support for injecting /dev/nvidia-nvswitch* devices if the NVIDIA_NVSWITCH=enabled envvar is specified.
-
Added support for
nvidia-ctk runtime configure --enable-cdi
for thedocker
runtime. Note that this requires Docker >= 25. -
Fixed bug in
nvidia-ctk config
command when using--set
. The types of applied config options are now applied correctly. -
Add
--relative-to
option tonvidia-ctk transform root
command. This controls whether the root transformation is applied to host or container paths. -
Added automatic CDI spec generation when the
runtime.nvidia.com/gpu=all
device is requested by a container. -
[libnvidia-container] Fix device permission check when using cgroupv2 (fixes #227)
v1.14.3
- [toolkit-container] Bump CUDA base image version to 12.2.2.
v1.14.2
- Fix bug on Tegra-based systems where symlinks were not created in containers.
- Add --csv.ignore-pattern command line option to nvidia-ctk cdi generate command.
v1.14.1
-
Fixed bug where contents of
/etc/nvidia-container-runtime/config.toml
is ignored by the NVIDIA Container Runtime Hook. -
[libnvidia-container] Use libelf.so on RPM-based systems due to removed mageia repositories hosting pmake and bmake.
v1.14.0
- Promote v1.14.0-rc.3 to v1.14.0
v1.14.0-rc.3
-
Added support for generating OCI hook JSON file to
nvidia-ctk runtime configure
command. -
Remove installation of OCI hook JSON from RPM package.
-
Refactored config for
nvidia-container-runtime-hook
. -
Added a
nvidia-ctk config
command which supports setting config options using a--set
flag. -
Added
--library-search-path
option tonvidia-ctk cdi generate
command incsv
mode. This allows folders where libraries are located to be specified explicitly. -
Updated go-nvlib to support devices which are not present in the PCI device database. This allows the creation of dev/char symlinks on systems with such devices installed.
-
Added
UsesNVGPUModule
info function for more robust platform detection. This is required on Tegra-based systems where libnvidia-ml.so is also supported. -
[toolkit-container] Set
NVIDIA_VISIBLE_DEVICES=void
to prevent injection of NVIDIA devices and drivers into the NVIDIA Container Toolkit container.
v1.14.0-rc.2
-
Fix bug causing incorrect nvidia-smi symlink to be created on WSL2 systems with multiple driver roots.
-
Remove dependency on coreutils when installing package on RPM-based systems.
-
Create output folders if required when running
nvidia-ctk runtime configure
-
Generate default config as post-install step.
-
Added support for detecting GSP firmware at custom paths when generating CDI specifications.
-
Added logic to skip the extraction of image requirements if
NVIDIA_DISABLE_REQUIRES
is set totrue
. -
[libnvidia-container] Include Shared Compiler Library (libnvidia-gpucomp.so) in the list of compute libaries.
-
[toolkit-container] Ensure that common envvars have higher priority when configuring the container engines.
-
[toolkit-container] Bump CUDA base image version to 12.2.0.
-
[toolkit-container] Remove installation of nvidia-experimental runtime. This is superceded by the NVIDIA Container Runtime in CDI mode.
v1.14.0-rc.1
-
Add support for updating containerd configs to the
nvidia-ctk runtime configure
command. -
Create file in
etc/ld.so.conf.d
with permissions644
to support non-root containers. -
Generate CDI specification files with
644
permissions to allow rootless applications (e.g. podman) -
Add
nvidia-ctk cdi list
command to show the known CDI devices. -
Add support for generating merged devices (e.g.
all
device) to the nvcdi API. -
Use . pattern to locate libcuda.so when generating a CDI specification to support platforms where a patch version is not specified.
-
Update go-nvlib to skip devices that are not MIG capable when generating CDI specifications.
-
Add
nvidia-container-runtime-hook.path
config option to specify NVIDIA Container Runtime Hook path explicitly. -
Fix bug in creation of
/dev/char
symlinks by failing operation if kernel modules are not loaded. -
Add option to load kernel modules when creating device nodes
-
Add option to create device nodes when creating
/dev/char
symlinks -
[libnvidia-container] Support OpenSSL 3 with the Encrypt/Decrypt library
-
[toolkit-container] Allow same envars for all runtime configs
v1.13.1
- Update
update-ldcache
hook to only update ldcache if it exists. - Update
update-ldcache
hook to create/etc/ld.so.conf.d
folder if it doesn't exist. - Fix failure when libcuda cannot be located during XOrg library discovery.
- Fix CDI spec generation on systems that use
/etc/alternatives
(e.g. Debian)
v1.13.0
- Promote 1.13.0-rc.3 to 1.13.0
v1.13.0-rc.3
-
Only initialize NVML for modes that require it when runing
nvidia-ctk cdi generate
. -
Prefer /run over /var/run when locating nvidia-persistenced and nvidia-fabricmanager sockets.
-
Fix the generation of CDI specifications for management containers when the driver libraries are not in the LDCache.
-
Add transformers to deduplicate and simplify CDI specifications.
-
Generate a simplified CDI specification by default. This means that entities in the common edits in a spec are not included in device definitions.
-
Also return an error from the nvcdi.New constructor instead of panicing.
-
Detect XOrg libraries for injection and CDI spec generation.
-
Add
nvidia-ctk system create-device-nodes
command to create control devices. -
Add
nvidia-ctk cdi transform
command to apply transforms to CDI specifications. -
Add
--vendor
and--class
options tonvidia-ctk cdi generate
-
[libnvidia-container] Fix segmentation fault when RPC initialization fails.
-
[libnvidia-container] Build centos variants of the NVIDIA Container Library with static libtirpc v1.3.2.
-
[libnvidia-container] Remove make targets for fedora35 as the centos8 packages are compatible.
-
[toolkit-container] Add
nvidia-container-runtime.modes.cdi.annotation-prefixes
config option that allows the CDI annotation prefixes that are read to be overridden. -
[toolkit-container] Create device nodes when generating CDI specification for management containers.
-
[toolkit-container] Add
nvidia-container-runtime.runtimes
config option to set the low-level runtime for the NVIDIA Container Runtime
v1.13.0-rc.2
- Don't fail chmod hook if paths are not injected
- Only create
by-path
symlinks if CDI devices are actually requested. - Fix possible blank
nvidia-ctk
path in generated CDI specifications - Fix error in postun scriplet on RPM-based systems
- Only check
NVIDIA_VISIBLE_DEVICES
for environment variables if no annotations are specified. - Add
cdi.default-kind
config option for constructing fully-qualified CDI device names in CDI mode - Add support for
accept-nvidia-visible-devices-envvar-unprivileged
config setting in CDI mode - Add
nvidia-container-runtime-hook.skip-mode-detection
config option to bypass mode detection. This allowslegacy
andcdi
mode, for example, to be used at the same time. - Add support for generating CDI specifications for GDS and MOFED devices
- Ensure CDI specification is validated on save when generating a spec
- Rename
--discovery-mode
argument to--mode
fornvidia-ctk cdi generate
- [libnvidia-container] Fix segfault on WSL2 systems
- [toolkit-container] Add
--cdi-enabled
flag to toolkit config - [toolkit-container] Install
nvidia-ctk
from toolkit container - [toolkit-container] Use installed
nvidia-ctk
path in NVIDIA Container Toolkit config - [toolkit-container] Bump CUDA base images to 12.1.0
- [toolkit-container] Set
nvidia-ctk
path in the - [toolkit-container] Add
cdi.k8s.io/*
to set of allowed annotations in containerd config - [toolkit-container] Generate CDI specification for use in management containers
- [toolkit-container] Install experimental runtime as
nvidia-container-runtime.experimental
instead ofnvidia-container-runtime-experimental
- [toolkit-container] Install and configure mode-specific runtimes for
cdi
andlegacy
modes
v1.13.0-rc.1
- Include MIG-enabled devices as GPUs when generating CDI specification
- Fix missing NVML symbols when running
nvidia-ctk
on some platforms [#49] - Add CDI spec generation for WSL2-based systems to
nvidia-ctk cdi generate
command - Add
auto
mode tonvidia-ctk cdi generate
command to automatically detect a WSL2-based system over a standard NVML-based system. - Add mode-specific (
.cdi
and.legacy
) NVIDIA Container Runtime binaries for use in the GPU Operator - Discover all
gsb*.bin
GSP firmware files when generating CDI specification. - Align
.deb
and.rpm
release candidate package versions - Remove
fedora35
packaging targets - [libnvidia-container] Include all
gsp*.bin
firmware files if present - [libnvidia-container] Align
.deb
and.rpm
release candidate package versions - [libnvidia-container] Remove
fedora35
packaging targets - [toolkit-container] Install
nvidia-container-toolkit-operator-extensions
package for mode-specific executables. - [toolkit-container] Allow
nvidia-container-runtime.mode
to be set when configuring the NVIDIA Container Toolkit
v1.12.0
- Promote
v1.12.0-rc.5
tov1.12.0
- Rename
nvidia cdi generate
--root
flag to--driver-root
to better indicate intent - [libnvidia-container] Add nvcubins.bin to DriverStore components under WSL2
- [toolkit-container] Bump CUDA base images to 12.0.1
v1.12.0-rc.5
- Fix bug here the
nvidia-ctk
path was not properly resolved. This causes failures to run containers when the runtime is configured incsv
mode or if theNVIDIA_DRIVER_CAPABILITIES
includesgraphics
ordisplay
(e.g.all
).
v1.12.0-rc.4
- Generate a minimum CDI spec version for improved compatibility.
- Add
--device-name-strategy
options to thenvidia-ctk cdi generate
command that can be used to control how device names are constructed. - Set default for CDI device name generation to
index
to generate device names such asnvidia.com/gpu=0
ornvidia.com/gpu=1:0
by default.
v1.12.0-rc.3
- Don't fail if by-path symlinks for DRM devices do not exist
- Replace the --json flag with a --format [json|yaml] flag for the nvidia-ctk cdi generate command
- Ensure that the CDI output folder is created if required
- When generating a CDI specification use a blank host path for devices to ensure compatibility with the v0.4.0 CDI specification
- Add injection of Wayland JSON files
- Add GSP firmware paths to generated CDI specification
- Add --root flag to nvidia-ctk cdi generate command
v1.12.0-rc.2
- Inject Direct Rendering Manager (DRM) devices into a container using the NVIDIA Container Runtime
- Improve logging of errors from the NVIDIA Container Runtime
- Improve CDI specification generation to support rootless podman
- Use
nvidia-ctk cdi generate
to generate CDI specifications instead ofnvidia-ctk info generate-cdi
- [libnvidia-container] Skip creation of existing files when these are already mounted
v1.12.0-rc.1
- Add support for multiple Docker Swarm resources
- Improve injection of Vulkan configurations and libraries
- Add
nvidia-ctk info generate-cdi
command to generated CDI specification for available devices - [libnvidia-container] Include NVVM compiler library in compute libs
v1.11.0
- Promote v1.11.0-rc.3 to v1.11.0
v1.11.0-rc.3
- Build fedora35 packages
- Introduce an
nvidia-container-toolkit-base
package for better dependency management - Fix removal of
nvidia-container-runtime-hook
on RPM-based systems - Inject platform files into container on Tegra-based systems
- [toolkit container] Update CUDA base images to 11.7.1
- [libnvidia-container] Preload libgcc_s.so.1 on arm64 systems
v1.11.0-rc.2
- Allow
accept-nvidia-visible-devices-*
config options to be set by toolkit container - [libnvidia-container] Fix bug where LDCache was not updated when the
--no-pivot-root
option was specified
v1.11.0-rc.1
- Add discovery of GPUDirect Storage (
nvidia-fs*
) devices if theNVIDIA_GDS
environment variable of the container is set toenabled
- Add discovery of MOFED Infiniband devices if the
NVIDIA_MOFED
environment variable of the container is set toenabled
- Fix bug in CSV mode where libraries listed as
sym
entries in mount specification are not added to the LDCache. - Rename
nvidia-container-toolkit
executable tonvidia-container-runtime-hook
and createnvidia-container-toolkit
as a symlink tonvidia-container-runtime-hook
instead. - Add
nvidia-ctk runtime configure
command to configure the Docker config file (e.g./etc/docker/daemon.json
) for use with the NVIDIA Container Runtime.
v1.10.0
- Promote v1.10.0-rc.3 to v1.10.0
v1.10.0-rc.3
- Use default config instead of raising an error if config file cannot be found
- Ignore NVIDIA_REQUIRE_JETPACK* environment variables for requirement checks
- Fix bug in detection of Tegra systems where
/sys/devices/soc0/family
is ignored - Fix bug where links to devices were detected as devices
- [libnvida-container] Fix bug introduced when adding libcudadebugger.so to list of libraries
v1.10.0-rc.2
- Add support for NVIDIA_REQUIRE_* checks for cuda version and arch to csv mode
- Switch to debug logging to reduce log verbosity
- Support logging to logs requested in command line
- Fix bug when launching containers with relative root path (e.g. using containerd)
- Allow low-level runtime path to be set explicitly as nvidia-container-runtime.runtimes option
- Fix failure to locate low-level runtime if PATH envvar is unset
- Replace experimental option for NVIDIA Container Runtime with nvidia-container-runtime.mode = csv option
- Use csv as default mode on Tegra systems without NVML
- Add --version flag to all CLIs
- [libnvidia-container] Bump libtirpc to 1.3.2
- [libnvidia-container] Fix bug when running host ldconfig using glibc compiled with a non-standard prefix
- [libnvidia-container] Add libcudadebugger.so to list of compute libraries
v1.10.0-rc.1
- Include nvidia-ctk CLI in installed binaries
- Add experimental option to NVIDIA Container Runtime
v1.9.0
- [libnvidia-container] Add additional check for Tegra in /sys/.../family file in CLI
- [libnvidia-container] Update jetpack-specific CLI option to only load Base CSV files by default
- [libnvidia-container] Fix bug (from 1.8.0) when mounting GSP firmware into containers without /lib to /usr/lib symlinks
- [libnvidia-container] Update nvml.h to CUDA 11.6.1 nvML_DEV 11.6.55
- [libnvidia-container] Update switch statement to include new brands from latest nvml.h
- [libnvidia-container] Process all --require flags on Jetson platforms
- [libnvidia-container] Fix long-standing issue with running ldconfig on Debian systems
v1.8.1
- [libnvidia-container] Fix bug in determining cgroup root when running in nested containers
- [libnvidia-container] Fix permission issue when determining cgroup version
v1.8.0
- Promote 1.8.0-rc.2-1 to 1.8.0
v1.8.0-rc.2
- Remove support for building amazonlinux1 packages
v1.8.0-rc.1
- [libnvidia-container] Add support for cgroupv2
- Release toolkit-container images from nvidia-container-toolkit repository
v1.7.0
- Promote 1.7.0-rc.1-1 to 1.7.0
- Bump Golang version to 1.16.4
v1.7.0-rc.1
- Specify containerd runtime type as string in config tools to remove dependency on containerd package
- Add supported-driver-capabilities config option to allow for a subset of all driver capabilities to be specified
v1.6.0
- Promote 1.6.0-rc.3-1 to 1.6.0
- Fix unnecessary logging to stderr instead of configured nvidia-container-runtime log file
v1.6.0-rc.3
- Add supported-driver-capabilities config option to the nvidia-container-toolkit
- Move OCI and command line checks for runtime to internal oci package
v1.6.0-rc.2
- Use relative path to OCI specification file (config.json) if bundle path is not specified as an argument to the nvidia-container-runtime
v1.6.0-rc.1
- Add AARCH64 package for Amazon Linux 2
- Include nvidia-container-runtime into nvidia-container-toolkit package
v1.5.1
- Fix bug where Docker Swarm device selection is ignored if NVIDIA_VISIBLE_DEVICES is also set
- Improve unit testing by using require package and adding coverage reports
- Remove unneeded go dependencies by running go mod tidy
- Move contents of pkg directory to cmd for CLI tools
- Ensure make binary target explicitly sets GOOS
v1.5.0
- Add dependence on libnvidia-container-tools >= 1.4.0
- Add golang check targets to Makefile
- Add Jenkinsfile definition for build targets
- Move docker.mk to docker folder
v1.4.2
- Add dependence on libnvidia-container-tools >= 1.3.3
v1.4.1
- Ignore NVIDIA_VISIBLE_DEVICES for containers with insufficent privileges
- Add dependence on libnvidia-container-tools >= 1.3.2
v1.4.0
- Add 'compute' capability to list of defaults
- Add dependence on libnvidia-container-tools >= 1.3.1
v1.3.0
- Promote 1.3.0-rc.2-1 to 1.3.0
- Add dependence on libnvidia-container-tools >= 1.3.0
v1.3.0-rc.2
2c180947
Add more tests for new semantics with device list from volume mounts7c003857
Refactor accepting device lists from volume mounts as a boolean
v1.3.0-rc.1
b50d86c1
Update build system to accept a TAG variable for things like rc.xfe65573b
Add common CI tests for things like golint, gofmt, unit tests, etc.da6fbb34
Revert "Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*"a7fb3330
Flip build-all targets to run automatically on merge requests8b248b66
Rename github.com/NVIDIA/container-toolkit to nvidia-container-toolkitda36874e
Add new config options to pull device list from mounted files instead of ENVVAR
v1.2.1
4e6e0ed4
Add 'ngx' to list ofall driver capabilities2f4af743
List config.toml as a config file in the RPM SPEC
v1.2.0
8e0aab46
Fix repo listed in changelog for debian distributions320bb6e4
Update dependence on libnvidia-container to 1.2.06cfc8097
Update package license to match source licensee7dc3cbb
Fix debian copyright filed3aee3e0
Add the 'ngx' driver capability
v1.1.2
c32237f3
Add support for parsing Linux Capabilities for older OCI specs
v1.1.1
d202aded
Update dependence to libnvidia-container 1.1.1
v1.1.0
4e4de762
Update build system to support multi-arch buildsfcc1d116
Add support for MIG (Multi-Instance GPUs)d4ff0416
Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*60f165ad
Add no-pivot option to toolkit
v1.0.5
- Initial release. Replaces older package nvidia-container-runtime-hook. (Closes: #XXXXXX)