This commit is contained in:
Shay Molcho 2025-03-21 09:12:13 +08:00 committed by GitHub
commit 62760c55d0
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -17,7 +17,7 @@
## v1.17.3
- Only allow host-relative LDConfig paths by default.
### Changes in libnvidia-container
- Create virtual copy of host ldconfig binary before calling fexecve()
- Create virtual copy of host ldconfig binary before calling fexecve().
## v1.17.2
- Fixed a bug where legacy images would set imex channels as `all`.
@ -32,24 +32,24 @@
## v1.17.0
- Promote v1.17.0-rc.2 to v1.17.0
- Fix bug when using just-in-time CDI spec generation
- Check for valid paths in create-symlinks hook
- Fix bug when using just-in-time CDI spec generation.
- Check for valid paths in create-symlinks hook.
## v1.17.0-rc.2
- Fix bug in locating libcuda.so from ldcache
- Fix bug in sorting of symlink chain
- Remove unsupported print-ldcache command
- Remove csv-filename support from create-symlinks
- Fix bug in locating libcuda.so from ldcache.
- Fix bug in sorting of symlink chain.
- Remove unsupported print-ldcache command.
- Remove csv-filename support from create-symlinks.
### Changes in the Toolkit Container
- Fallback to `crio-status` if `crio status` does not work when configuring the crio runtime
- Fallback to `crio-status` if `crio status` does not work when configuring the crio runtime.
## v1.17.0-rc.1
- Allow IMEX channels to be requested as volume mounts
- Fix typo in error message
- Add disable-imex-channel-creation feature flag
- Add -z,lazy to LDFLAGS
- Add imex channels to management CDI spec
- Allow IMEX channels to be requested as volume mounts.
- Fix typo in error message.
- Add disable-imex-channel-creation feature flag.
- Add -z,lazy to LDFLAGS.
- Add imex channels to management CDI spec.
- Add support to fetch current container runtime config from the command line.
- Add creation of select driver symlinks to CDI spec generation.
- Remove support for config overrides when configuring runtimes.
@ -75,13 +75,13 @@
- Bump CUDA base image version to 12.6.0
### Changes in libnvidia-container
- Add no-gsp-firmware command line option
- Add no-fabricmanager command line option
- Add no-persistenced command line option
- Add no-gsp-firmware command line option.
- Add no-fabricmanager command line option.
- Add no-persistenced command line option.
- Skip directories and symlinks when mounting libraries.
## v1.16.1
- Fix bug with processing errors during CDI spec generation for MIG devices
- Fix bug with processing errors during CDI spec generation for MIG devices.
## v1.16.0
- Promote v1.16.0-rc.2 to v1.16.0
@ -90,13 +90,13 @@
- Bump CUDA base image version to 12.5.1
## v1.16.0-rc.2
- Use relative path to locate driver libraries
- Add RelativeToRoot function to Driver
- Inject additional libraries for full X11 functionality
- Extract options from default runtime if runc does not exist
- Avoid using map pointers as maps are always passed by reference
- Reduce logging for the NVIDIA Container runtime
- Fix bug in argument parsing for logger creation
- Use relative path to locate driver libraries.
- Add RelativeToRoot function to Driver.
- Inject additional libraries for full X11 functionality.
- Extract options from default runtime if runc does not exist.
- Avoid using map pointers as maps are always passed by reference.
- Reduce logging for the NVIDIA Container runtime.
- Fix bug in argument parsing for logger creation.
## v1.16.0-rc.1
@ -148,13 +148,13 @@
* Fix bug when specifying `--dev-root` for Tegra-based systems.
* Log explicitly requested runtime mode.
* Remove package dependency on libseccomp.
* Added detection of libnvdxgdmal.so.1 on WSL2
* Added detection of libnvdxgdmal.so.1 on WSL2.
* Use devRoot to resolve MIG device nodes.
* Fix bug in determining default nvidia-container-runtime.user config value on SUSE-based systems.
* Add `crun` to the list of configured low-level runtimes.
* Added support for `--ldconfig-path` to `nvidia-ctk cdi generate` command.
* Fix `nvidia-ctk runtime configure --cdi.enabled` for Docker.
* Add discovery of the GDRCopy device (`gdrdrv`) if the `NVIDIA_GDRCOPY` environment variable of the container is set to `enabled`
* Add discovery of the GDRCopy device (`gdrdrv`) if the `NVIDIA_GDRCOPY` environment variable of the container is set to `enabled`.
* [toolkit-container] Bump CUDA base image version to 12.3.1.
@ -170,7 +170,7 @@
* Add `--relative-to` option to `nvidia-ctk transform root` command. This controls whether the root transformation is applied to host or container paths.
* Added automatic CDI spec generation when the `runtime.nvidia.com/gpu=all` device is requested by a container.
* [libnvidia-container] Fix device permission check when using cgroupv2 (fixes #227)
* [libnvidia-container] Fix device permission check when using cgroupv2 (fixes #227).
## v1.14.3
* [toolkit-container] Bump CUDA base image version to 12.2.2.
@ -202,7 +202,7 @@
## v1.14.0-rc.2
* Fix bug causing incorrect nvidia-smi symlink to be created on WSL2 systems with multiple driver roots.
* Remove dependency on coreutils when installing package on RPM-based systems.
* Create output folders if required when running `nvidia-ctk runtime configure`
* Create output folders if required when running `nvidia-ctk runtime configure`.
* Generate default config as post-install step.
* Added support for detecting GSP firmware at custom paths when generating CDI specifications.
* Added logic to skip the extraction of image requirements if `NVIDIA_DISABLE_REQUIRES` is set to `true`.
@ -224,19 +224,19 @@
* Update go-nvlib to skip devices that are not MIG capable when generating CDI specifications.
* Add `nvidia-container-runtime-hook.path` config option to specify NVIDIA Container Runtime Hook path explicitly.
* Fix bug in creation of `/dev/char` symlinks by failing operation if kernel modules are not loaded.
* Add option to load kernel modules when creating device nodes
* Add option to create device nodes when creating `/dev/char` symlinks
* Add option to load kernel modules when creating device nodes.
* Add option to create device nodes when creating `/dev/char` symlinks.
* [libnvidia-container] Support OpenSSL 3 with the Encrypt/Decrypt library
* [libnvidia-container] Support OpenSSL 3 with the Encrypt/Decrypt library.
* [toolkit-container] Allow same envars for all runtime configs
* [toolkit-container] Allow same envars for all runtime configs.
## v1.13.1
* Update `update-ldcache` hook to only update ldcache if it exists.
* Update `update-ldcache` hook to create `/etc/ld.so.conf.d` folder if it doesn't exist.
* Fix failure when libcuda cannot be located during XOrg library discovery.
* Fix CDI spec generation on systems that use `/etc/alternatives` (e.g. Debian)
* Fix CDI spec generation on systems that use `/etc/alternatives` (e.g. Debian).
## v1.13.0
@ -253,7 +253,7 @@
* Detect XOrg libraries for injection and CDI spec generation.
* Add `nvidia-ctk system create-device-nodes` command to create control devices.
* Add `nvidia-ctk cdi transform` command to apply transforms to CDI specifications.
* Add `--vendor` and `--class` options to `nvidia-ctk cdi generate`
* Add `--vendor` and `--class` options to `nvidia-ctk cdi generate`.
* [libnvidia-container] Fix segmentation fault when RPC initialization fails.
* [libnvidia-container] Build centos variants of the NVIDIA Container Library with static libtirpc v1.3.2.
@ -261,53 +261,53 @@
* [toolkit-container] Add `nvidia-container-runtime.modes.cdi.annotation-prefixes` config option that allows the CDI annotation prefixes that are read to be overridden.
* [toolkit-container] Create device nodes when generating CDI specification for management containers.
* [toolkit-container] Add `nvidia-container-runtime.runtimes` config option to set the low-level runtime for the NVIDIA Container Runtime
* [toolkit-container] Add `nvidia-container-runtime.runtimes` config option to set the low-level runtime for the NVIDIA Container Runtime.
## v1.13.0-rc.2
* Don't fail chmod hook if paths are not injected
* Don't fail chmod hook if paths are not injected.
* Only create `by-path` symlinks if CDI devices are actually requested.
* Fix possible blank `nvidia-ctk` path in generated CDI specifications
* Fix error in postun scriplet on RPM-based systems
* Fix possible blank `nvidia-ctk` path in generated CDI specifications.
* Fix error in postun scriplet on RPM-based systems.
* Only check `NVIDIA_VISIBLE_DEVICES` for environment variables if no annotations are specified.
* Add `cdi.default-kind` config option for constructing fully-qualified CDI device names in CDI mode
* Add support for `accept-nvidia-visible-devices-envvar-unprivileged` config setting in CDI mode
* Add `cdi.default-kind` config option for constructing fully-qualified CDI device names in CDI mode.
* Add support for `accept-nvidia-visible-devices-envvar-unprivileged` config setting in CDI mode.
* Add `nvidia-container-runtime-hook.skip-mode-detection` config option to bypass mode detection. This allows `legacy` and `cdi` mode, for example, to be used at the same time.
* Add support for generating CDI specifications for GDS and MOFED devices
* Ensure CDI specification is validated on save when generating a spec
* Add support for generating CDI specifications for GDS and MOFED devices.
* Ensure CDI specification is validated on save when generating a spec.
* Rename `--discovery-mode` argument to `--mode` for `nvidia-ctk cdi generate`
* [libnvidia-container] Fix segfault on WSL2 systems
* [toolkit-container] Add `--cdi-enabled` flag to toolkit config
* [toolkit-container] Install `nvidia-ctk` from toolkit container
* [toolkit-container] Use installed `nvidia-ctk` path in NVIDIA Container Toolkit config
* [libnvidia-container] Fix segfault on WSL2 systems.
* [toolkit-container] Add `--cdi-enabled` flag to toolkit config.
* [toolkit-container] Install `nvidia-ctk` from toolkit container.
* [toolkit-container] Use installed `nvidia-ctk` path in NVIDIA Container Toolkit config.
* [toolkit-container] Bump CUDA base images to 12.1.0
* [toolkit-container] Set `nvidia-ctk` path in the
* [toolkit-container] Add `cdi.k8s.io/*` to set of allowed annotations in containerd config
* [toolkit-container] Generate CDI specification for use in management containers
* [toolkit-container] Install experimental runtime as `nvidia-container-runtime.experimental` instead of `nvidia-container-runtime-experimental`
* [toolkit-container] Install and configure mode-specific runtimes for `cdi` and `legacy` modes
* [toolkit-container] Set `nvidia-ctk` path in the.
* [toolkit-container] Add `cdi.k8s.io/*` to set of allowed annotations in containerd config.
* [toolkit-container] Generate CDI specification for use in management containers.
* [toolkit-container] Install experimental runtime as `nvidia-container-runtime.experimental` instead of `nvidia-container-runtime-experimental`.
* [toolkit-container] Install and configure mode-specific runtimes for `cdi` and `legacy` modes.
## v1.13.0-rc.1
* Include MIG-enabled devices as GPUs when generating CDI specification
* Fix missing NVML symbols when running `nvidia-ctk` on some platforms [#49]
* Add CDI spec generation for WSL2-based systems to `nvidia-ctk cdi generate` command
* Include MIG-enabled devices as GPUs when generating CDI specification.
* Fix missing NVML symbols when running `nvidia-ctk` on some platforms [#49].
* Add CDI spec generation for WSL2-based systems to `nvidia-ctk cdi generate` command.
* Add `auto` mode to `nvidia-ctk cdi generate` command to automatically detect a WSL2-based system over a standard NVML-based system.
* Add mode-specific (`.cdi` and `.legacy`) NVIDIA Container Runtime binaries for use in the GPU Operator
* Add mode-specific (`.cdi` and `.legacy`) NVIDIA Container Runtime binaries for use in the GPU Operator.
* Discover all `gsb*.bin` GSP firmware files when generating CDI specification.
* Align `.deb` and `.rpm` release candidate package versions
* Remove `fedora35` packaging targets
* [libnvidia-container] Include all `gsp*.bin` firmware files if present
* [libnvidia-container] Align `.deb` and `.rpm` release candidate package versions
* [libnvidia-container] Remove `fedora35` packaging targets
* Align `.deb` and `.rpm` release candidate package versions.
* Remove `fedora35` packaging targets.
* [libnvidia-container] Include all `gsp*.bin` firmware files if present.
* [libnvidia-container] Align `.deb` and `.rpm` release candidate package versions.
* [libnvidia-container] Remove `fedora35` packaging targets.
* [toolkit-container] Install `nvidia-container-toolkit-operator-extensions` package for mode-specific executables.
* [toolkit-container] Allow `nvidia-container-runtime.mode` to be set when configuring the NVIDIA Container Toolkit
* [toolkit-container] Allow `nvidia-container-runtime.mode` to be set when configuring the NVIDIA Container Toolkit.
## v1.12.0
* Promote `v1.12.0-rc.5` to `v1.12.0`
* Rename `nvidia cdi generate` `--root` flag to `--driver-root` to better indicate intent
* [libnvidia-container] Add nvcubins.bin to DriverStore components under WSL2
* Promote `v1.12.0-rc.5` to `v1.12.0`.
* Rename `nvidia cdi generate` `--root` flag to `--driver-root` to better indicate intent.
* [libnvidia-container] Add nvcubins.bin to DriverStore components under WSL2.
* [toolkit-container] Bump CUDA base images to 12.0.1
## v1.12.0-rc.5
@ -322,28 +322,28 @@
## v1.12.0-rc.3
* Don't fail if by-path symlinks for DRM devices do not exist
* Replace the --json flag with a --format [json|yaml] flag for the nvidia-ctk cdi generate command
* Ensure that the CDI output folder is created if required
* When generating a CDI specification use a blank host path for devices to ensure compatibility with the v0.4.0 CDI specification
* Add injection of Wayland JSON files
* Add GSP firmware paths to generated CDI specification
* Add --root flag to nvidia-ctk cdi generate command
* Don't fail if by-path symlinks for DRM devices do not exist.
* Replace the --json flag with a --format [json|yaml] flag for the nvidia-ctk cdi generate command.
* Ensure that the CDI output folder is created if required.
* When generating a CDI specification use a blank host path for devices to ensure compatibility with the v0.4.0 CDI specification.
* Add injection of Wayland JSON files.
* Add GSP firmware paths to generated CDI specification.
* Add --root flag to nvidia-ctk cdi generate command.
## v1.12.0-rc.2
* Inject Direct Rendering Manager (DRM) devices into a container using the NVIDIA Container Runtime
* Improve logging of errors from the NVIDIA Container Runtime
* Improve CDI specification generation to support rootless podman
* Use `nvidia-ctk cdi generate` to generate CDI specifications instead of `nvidia-ctk info generate-cdi`
* [libnvidia-container] Skip creation of existing files when these are already mounted
* Inject Direct Rendering Manager (DRM) devices into a container using the NVIDIA Container Runtime.
* Improve logging of errors from the NVIDIA Container Runtime.
* Improve CDI specification generation to support rootless podman.
* Use `nvidia-ctk cdi generate` to generate CDI specifications instead of `nvidia-ctk info generate-cdi`.
* [libnvidia-container] Skip creation of existing files when these are already mounted.
## v1.12.0-rc.1
* Add support for multiple Docker Swarm resources
* Improve injection of Vulkan configurations and libraries
* Add `nvidia-ctk info generate-cdi` command to generated CDI specification for available devices
* [libnvidia-container] Include NVVM compiler library in compute libs
* Add support for multiple Docker Swarm resources.
* Improve injection of Vulkan configurations and libraries.
* Add `nvidia-ctk info generate-cdi` command to generated CDI specification for available devices.
* [libnvidia-container] Include NVVM compiler library in compute libs.
## v1.11.0
@ -351,22 +351,22 @@
## v1.11.0-rc.3
* Build fedora35 packages
* Introduce an `nvidia-container-toolkit-base` package for better dependency management
* Fix removal of `nvidia-container-runtime-hook` on RPM-based systems
* Inject platform files into container on Tegra-based systems
* Build fedora35 packages.
* Introduce an `nvidia-container-toolkit-base` package for better dependency management.
* Fix removal of `nvidia-container-runtime-hook` on RPM-based systems.
* Inject platform files into container on Tegra-based systems.
* [toolkit container] Update CUDA base images to 11.7.1
* [libnvidia-container] Preload libgcc_s.so.1 on arm64 systems
* [libnvidia-container] Preload libgcc_s.so.1 on arm64 systems.
## v1.11.0-rc.2
* Allow `accept-nvidia-visible-devices-*` config options to be set by toolkit container
* Allow `accept-nvidia-visible-devices-*` config options to be set by toolkit container.
* [libnvidia-container] Fix bug where LDCache was not updated when the `--no-pivot-root` option was specified
## v1.11.0-rc.1
* Add discovery of GPUDirect Storage (`nvidia-fs*`) devices if the `NVIDIA_GDS` environment variable of the container is set to `enabled`
* Add discovery of MOFED Infiniband devices if the `NVIDIA_MOFED` environment variable of the container is set to `enabled`
* Add discovery of GPUDirect Storage (`nvidia-fs*`) devices if the `NVIDIA_GDS` environment variable of the container is set to `enabled`.
* Add discovery of MOFED Infiniband devices if the `NVIDIA_MOFED` environment variable of the container is set to `enabled`.
* Fix bug in CSV mode where libraries listed as `sym` entries in mount specification are not added to the LDCache.
* Rename `nvidia-container-toolkit` executable to `nvidia-container-runtime-hook` and create `nvidia-container-toolkit` as a symlink to `nvidia-container-runtime-hook` instead.
* Add `nvidia-ctk runtime configure` command to configure the Docker config file (e.g. `/etc/docker/daemon.json`) for use with the NVIDIA Container Runtime.
@ -377,46 +377,46 @@
## v1.10.0-rc.3
* Use default config instead of raising an error if config file cannot be found
* Ignore NVIDIA_REQUIRE_JETPACK* environment variables for requirement checks
* Fix bug in detection of Tegra systems where `/sys/devices/soc0/family` is ignored
* Fix bug where links to devices were detected as devices
* [libnvida-container] Fix bug introduced when adding libcudadebugger.so to list of libraries
* Use default config instead of raising an error if config file cannot be found.
* Ignore NVIDIA_REQUIRE_JETPACK* environment variables for requirement checks.
* Fix bug in detection of Tegra systems where `/sys/devices/soc0/family` is ignored.
* Fix bug where links to devices were detected as devices.
* [libnvida-container] Fix bug introduced when adding libcudadebugger.so to list of libraries.
## v1.10.0-rc.2
* Add support for NVIDIA_REQUIRE_* checks for cuda version and arch to csv mode
* Switch to debug logging to reduce log verbosity
* Support logging to logs requested in command line
* Fix bug when launching containers with relative root path (e.g. using containerd)
* Allow low-level runtime path to be set explicitly as nvidia-container-runtime.runtimes option
* Fix failure to locate low-level runtime if PATH envvar is unset
* Replace experimental option for NVIDIA Container Runtime with nvidia-container-runtime.mode = csv option
* Use csv as default mode on Tegra systems without NVML
* Add --version flag to all CLIs
* Add support for NVIDIA_REQUIRE_* checks for cuda version and arch to csv mode.
* Switch to debug logging to reduce log verbosity.
* Support logging to logs requested in command line.
* Fix bug when launching containers with relative root path (e.g. using containerd).
* Allow low-level runtime path to be set explicitly as nvidia-container-runtime.runtimes option.
* Fix failure to locate low-level runtime if PATH envvar is unset.
* Replace experimental option for NVIDIA Container Runtime with nvidia-container-runtime.mode = csv option.
* Use csv as default mode on Tegra systems without NVML.
* Add --version flag to all CLIs.
* [libnvidia-container] Bump libtirpc to 1.3.2
* [libnvidia-container] Fix bug when running host ldconfig using glibc compiled with a non-standard prefix
* [libnvidia-container] Add libcudadebugger.so to list of compute libraries
* [libnvidia-container] Fix bug when running host ldconfig using glibc compiled with a non-standard prefix.
* [libnvidia-container] Add libcudadebugger.so to list of compute libraries.
## v1.10.0-rc.1
* Include nvidia-ctk CLI in installed binaries
* Add experimental option to NVIDIA Container Runtime
* Include nvidia-ctk CLI in installed binaries.
* Add experimental option to NVIDIA Container Runtime.
## v1.9.0
* [libnvidia-container] Add additional check for Tegra in /sys/.../family file in CLI
* [libnvidia-container] Update jetpack-specific CLI option to only load Base CSV files by default
* [libnvidia-container] Fix bug (from 1.8.0) when mounting GSP firmware into containers without /lib to /usr/lib symlinks
* [libnvidia-container] Add additional check for Tegra in /sys/.../family file in CLI.
* [libnvidia-container] Update jetpack-specific CLI option to only load Base CSV files by default.
* [libnvidia-container] Fix bug (from 1.8.0) when mounting GSP firmware into containers without /lib to /usr/lib symlinks.
* [libnvidia-container] Update nvml.h to CUDA 11.6.1 nvML_DEV 11.6.55
* [libnvidia-container] Update switch statement to include new brands from latest nvml.h
* [libnvidia-container] Process all --require flags on Jetson platforms
* [libnvidia-container] Fix long-standing issue with running ldconfig on Debian systems
* [libnvidia-container] Process all --require flags on Jetson platforms.
* [libnvidia-container] Fix long-standing issue with running ldconfig on Debian systems.
## v1.8.1
* [libnvidia-container] Fix bug in determining cgroup root when running in nested containers
* [libnvidia-container] Fix permission issue when determining cgroup version
* [libnvidia-container] Fix bug in determining cgroup root when running in nested containers.
* [libnvidia-container] Fix permission issue when determining cgroup version.
## v1.8.0
@ -424,12 +424,12 @@
## v1.8.0-rc.2
* Remove support for building amazonlinux1 packages
* Remove support for building amazonlinux1 packages.
## v1.8.0-rc.1
* [libnvidia-container] Add support for cgroupv2
* Release toolkit-container images from nvidia-container-toolkit repository
* [libnvidia-container] Add support for cgroupv2.
* Release toolkit-container images from nvidia-container-toolkit repository.
## v1.7.0
@ -438,42 +438,42 @@
## v1.7.0-rc.1
* Specify containerd runtime type as string in config tools to remove dependency on containerd package
* Add supported-driver-capabilities config option to allow for a subset of all driver capabilities to be specified
* Specify containerd runtime type as string in config tools to remove dependency on containerd package.
* Add supported-driver-capabilities config option to allow for a subset of all driver capabilities to be specified.
## v1.6.0
* Promote 1.6.0-rc.3-1 to 1.6.0
* Fix unnecessary logging to stderr instead of configured nvidia-container-runtime log file
* Fix unnecessary logging to stderr instead of configured nvidia-container-runtime log file.
## v1.6.0-rc.3
* Add supported-driver-capabilities config option to the nvidia-container-toolkit
* Move OCI and command line checks for runtime to internal oci package
* Add supported-driver-capabilities config option to the nvidia-container-toolkit.
* Move OCI and command line checks for runtime to internal oci package.
## v1.6.0-rc.2
* Use relative path to OCI specification file (config.json) if bundle path is not specified as an argument to the nvidia-container-runtime
* Use relative path to OCI specification file (config.json) if bundle path is not specified as an argument to the nvidia-container-runtime.
## v1.6.0-rc.1
* Add AARCH64 package for Amazon Linux 2
* Include nvidia-container-runtime into nvidia-container-toolkit package
* Add AARCH64 package for Amazon Linux 2.
* Include nvidia-container-runtime into nvidia-container-toolkit package.
## v1.5.1
* Fix bug where Docker Swarm device selection is ignored if NVIDIA_VISIBLE_DEVICES is also set
* Improve unit testing by using require package and adding coverage reports
* Remove unneeded go dependencies by running go mod tidy
* Move contents of pkg directory to cmd for CLI tools
* Ensure make binary target explicitly sets GOOS
* Fix bug where Docker Swarm device selection is ignored if NVIDIA_VISIBLE_DEVICES is also set.
* Improve unit testing by using require package and adding coverage reports.
* Remove unneeded go dependencies by running go mod tidy.
* Move contents of pkg directory to cmd for CLI tools.
* Ensure make binary target explicitly sets GOOS.
## v1.5.0
* Add dependence on libnvidia-container-tools >= 1.4.0
* Add golang check targets to Makefile
* Add Jenkinsfile definition for build targets
* Move docker.mk to docker folder
* Add golang check targets to Makefile.
* Add Jenkinsfile definition for build targets.
* Move docker.mk to docker folder.
## v1.4.2
@ -481,12 +481,12 @@
## v1.4.1
* Ignore NVIDIA_VISIBLE_DEVICES for containers with insufficent privileges
* Ignore NVIDIA_VISIBLE_DEVICES for containers with insufficent privileges.
* Add dependence on libnvidia-container-tools >= 1.3.2
## v1.4.0
* Add 'compute' capability to list of defaults
* Add 'compute' capability to list of defaults.
* Add dependence on libnvidia-container-tools >= 1.3.1
## v1.3.0
@ -496,34 +496,34 @@
## v1.3.0-rc.2
* 2c180947 Add more tests for new semantics with device list from volume mounts
* 7c003857 Refactor accepting device lists from volume mounts as a boolean
* 2c180947 Add more tests for new semantics with device list from volume mounts.
* 7c003857 Refactor accepting device lists from volume mounts as a boolean.
## v1.3.0-rc.1
* b50d86c1 Update build system to accept a TAG variable for things like rc.x
* fe65573b Add common CI tests for things like golint, gofmt, unit tests, etc.
* da6fbb34 Revert "Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*"
* a7fb3330 Flip build-all targets to run automatically on merge requests
* 8b248b66 Rename github.com/NVIDIA/container-toolkit to nvidia-container-toolkit
* da36874e Add new config options to pull device list from mounted files instead of ENVVAR
* da6fbb34 Revert "Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*".
* a7fb3330 Flip build-all targets to run automatically on merge requests.
* 8b248b66 Rename github.com/NVIDIA/container-toolkit to nvidia-container-toolkit.
* da36874e Add new config options to pull device list from mounted files instead of ENVVAR.
## v1.2.1
* 4e6e0ed4 Add 'ngx' to list of*all* driver capabilities
* 2f4af743 List config.toml as a config file in the RPM SPEC
* 4e6e0ed4 Add 'ngx' to list of*all* driver capabilities.
* 2f4af743 List config.toml as a config file in the RPM SPEC.
## v1.2.0
* 8e0aab46 Fix repo listed in changelog for debian distributions
* 8e0aab46 Fix repo listed in changelog for debian distributions.
* 320bb6e4 Update dependence on libnvidia-container to 1.2.0
* 6cfc8097 Update package license to match source license
* e7dc3cbb Fix debian copyright file
* d3aee3e0 Add the 'ngx' driver capability
* 6cfc8097 Update package license to match source license.
* e7dc3cbb Fix debian copyright file.
* d3aee3e0 Add the 'ngx' driver capability.
## v1.1.2
* c32237f3 Add support for parsing Linux Capabilities for older OCI specs
* c32237f3 Add support for parsing Linux Capabilities for older OCI specs.
## v1.1.1
@ -531,10 +531,10 @@
## v1.1.0
* 4e4de762 Update build system to support multi-arch builds
* fcc1d116 Add support for MIG (Multi-Instance GPUs)
* 4e4de762 Update build system to support multi-arch builds.
* fcc1d116 Add support for MIG (Multi-Instance GPUs).
* d4ff0416 Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*
* 60f165ad Add no-pivot option to toolkit
* 60f165ad Add no-pivot option to toolkit.
## v1.0.5