diff --git a/CHANGELOG.md b/CHANGELOG.md
index dd79e7de..c1075d2c 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -17,7 +17,7 @@
 ## v1.17.3
 - Only allow host-relative LDConfig paths by default.
 ### Changes in libnvidia-container
-- Create virtual copy of host ldconfig binary before calling fexecve()
+- Create virtual copy of host ldconfig binary before calling fexecve().
 
 ## v1.17.2
 - Fixed a bug where legacy images would set imex channels as `all`.
@@ -32,24 +32,24 @@
 
 ## v1.17.0
 - Promote v1.17.0-rc.2 to v1.17.0
-- Fix bug when using just-in-time CDI spec generation
-- Check for valid paths in create-symlinks hook
+- Fix bug when using just-in-time CDI spec generation.
+- Check for valid paths in create-symlinks hook.
 
 ## v1.17.0-rc.2
-- Fix bug in locating libcuda.so from ldcache
-- Fix bug in sorting of symlink chain
-- Remove unsupported print-ldcache command
-- Remove csv-filename support from create-symlinks
+- Fix bug in locating libcuda.so from ldcache.
+- Fix bug in sorting of symlink chain.
+- Remove unsupported print-ldcache command.
+- Remove csv-filename support from create-symlinks.
 
 ### Changes in the Toolkit Container
-- Fallback to `crio-status` if `crio status` does not work when configuring the crio runtime
+- Fallback to `crio-status` if `crio status` does not work when configuring the crio runtime.
 
 ## v1.17.0-rc.1
-- Allow IMEX channels to be requested as volume mounts
-- Fix typo in error message
-- Add disable-imex-channel-creation feature flag
-- Add -z,lazy to LDFLAGS
-- Add imex channels to management CDI spec
+- Allow IMEX channels to be requested as volume mounts.
+- Fix typo in error message.
+- Add disable-imex-channel-creation feature flag.
+- Add -z,lazy to LDFLAGS.
+- Add imex channels to management CDI spec.
 - Add support to fetch current container runtime config from the command line.
 - Add creation of select driver symlinks to CDI spec generation.
 - Remove support for config overrides when configuring runtimes.
@@ -75,13 +75,13 @@
 - Bump CUDA base image version to 12.6.0
 
 ### Changes in libnvidia-container
-- Add no-gsp-firmware command line option
-- Add no-fabricmanager command line option
-- Add no-persistenced command line option
+- Add no-gsp-firmware command line option.
+- Add no-fabricmanager command line option.
+- Add no-persistenced command line option.
 - Skip directories and symlinks when mounting libraries.
 
 ## v1.16.1
-- Fix bug with processing errors during CDI spec generation for MIG devices
+- Fix bug with processing errors during CDI spec generation for MIG devices.
 
 ## v1.16.0
 - Promote v1.16.0-rc.2 to v1.16.0
@@ -90,13 +90,13 @@
 - Bump CUDA base image version to 12.5.1
 
 ## v1.16.0-rc.2
-- Use relative path to locate driver libraries
-- Add RelativeToRoot function to Driver
-- Inject additional libraries for full X11 functionality
-- Extract options from default runtime if runc does not exist
-- Avoid using map pointers as maps are always passed by reference
-- Reduce logging for the NVIDIA Container runtime
-- Fix bug in argument parsing for logger creation
+- Use relative path to locate driver libraries.
+- Add RelativeToRoot function to Driver.
+- Inject additional libraries for full X11 functionality.
+- Extract options from default runtime if runc does not exist.
+- Avoid using map pointers as maps are always passed by reference.
+- Reduce logging for the NVIDIA Container runtime.
+- Fix bug in argument parsing for logger creation.
 
 ## v1.16.0-rc.1
 
@@ -148,13 +148,13 @@
 * Fix bug when specifying `--dev-root` for Tegra-based systems.
 * Log explicitly requested runtime mode.
 * Remove package dependency on libseccomp.
-* Added detection of libnvdxgdmal.so.1 on WSL2
+* Added detection of libnvdxgdmal.so.1 on WSL2.
 * Use devRoot to resolve MIG device nodes.
 * Fix bug in determining default nvidia-container-runtime.user config value on SUSE-based systems.
 * Add `crun` to the list of configured low-level runtimes.
 * Added support for `--ldconfig-path` to `nvidia-ctk cdi generate` command.
 * Fix `nvidia-ctk runtime configure --cdi.enabled` for Docker.
-* Add discovery of the GDRCopy device (`gdrdrv`) if the `NVIDIA_GDRCOPY` environment variable of the container is set to `enabled`
+* Add discovery of the GDRCopy device (`gdrdrv`) if the `NVIDIA_GDRCOPY` environment variable of the container is set to `enabled`.
 
 * [toolkit-container] Bump CUDA base image version to 12.3.1.
 
@@ -170,7 +170,7 @@
 * Add `--relative-to` option to `nvidia-ctk transform root` command. This controls whether the root transformation is applied to host or container paths.
 * Added automatic CDI spec generation when the `runtime.nvidia.com/gpu=all` device is requested by a container.
 
-* [libnvidia-container] Fix device permission check when using cgroupv2 (fixes #227)
+* [libnvidia-container] Fix device permission check when using cgroupv2 (fixes #227).
 
 ## v1.14.3
 * [toolkit-container] Bump CUDA base image version to 12.2.2.
@@ -202,7 +202,7 @@
 ## v1.14.0-rc.2
 * Fix bug causing incorrect nvidia-smi symlink to be created on WSL2 systems with multiple driver roots.
 * Remove dependency on coreutils when installing package on RPM-based systems.
-* Create output folders if required when running `nvidia-ctk runtime configure`
+* Create output folders if required when running `nvidia-ctk runtime configure`.
 * Generate default config as post-install step.
 * Added support for detecting GSP firmware at custom paths when generating CDI specifications.
 * Added logic to skip the extraction of image requirements if `NVIDIA_DISABLE_REQUIRES` is set to `true`.
@@ -224,19 +224,19 @@
 * Update go-nvlib to skip devices that are not MIG capable when generating CDI specifications.
 * Add `nvidia-container-runtime-hook.path` config option to specify NVIDIA Container Runtime Hook path explicitly.
 * Fix bug in creation of `/dev/char` symlinks by failing operation if kernel modules are not loaded.
-* Add option to load kernel modules when creating device nodes
-* Add option to create device nodes when creating `/dev/char` symlinks
+* Add option to load kernel modules when creating device nodes.
+* Add option to create device nodes when creating `/dev/char` symlinks.
 
-* [libnvidia-container] Support OpenSSL 3 with the Encrypt/Decrypt library
+* [libnvidia-container] Support OpenSSL 3 with the Encrypt/Decrypt library.
 
-* [toolkit-container] Allow same envars for all runtime configs
+* [toolkit-container] Allow same envars for all runtime configs.
 
 ## v1.13.1
 
 * Update `update-ldcache` hook to only update ldcache if it exists.
 * Update `update-ldcache` hook to create `/etc/ld.so.conf.d` folder if it doesn't exist.
 * Fix failure when libcuda cannot be located during XOrg library discovery.
-* Fix CDI spec generation on systems that use `/etc/alternatives` (e.g. Debian)
+* Fix CDI spec generation on systems that use `/etc/alternatives` (e.g. Debian).
 
 ## v1.13.0
 
@@ -253,7 +253,7 @@
 * Detect XOrg libraries for injection and CDI spec generation.
 * Add `nvidia-ctk system create-device-nodes` command to create control devices.
 * Add `nvidia-ctk cdi transform` command to apply transforms to CDI specifications.
-* Add `--vendor` and `--class` options to `nvidia-ctk cdi generate`
+* Add `--vendor` and `--class` options to `nvidia-ctk cdi generate`.
 
 * [libnvidia-container] Fix segmentation fault when RPC initialization fails.
 * [libnvidia-container] Build centos variants of the NVIDIA Container Library with static libtirpc v1.3.2.
@@ -261,53 +261,53 @@
 
 * [toolkit-container] Add `nvidia-container-runtime.modes.cdi.annotation-prefixes` config option that allows the CDI annotation prefixes that are read to be overridden.
 * [toolkit-container] Create device nodes when generating CDI specification for management containers.
-* [toolkit-container] Add `nvidia-container-runtime.runtimes` config option to set the low-level runtime for the NVIDIA Container Runtime
+* [toolkit-container] Add `nvidia-container-runtime.runtimes` config option to set the low-level runtime for the NVIDIA Container Runtime.
 
 ## v1.13.0-rc.2
 
-* Don't fail chmod hook if paths are not injected
+* Don't fail chmod hook if paths are not injected.
 * Only create `by-path` symlinks if CDI devices are actually requested.
-* Fix possible blank `nvidia-ctk` path in generated CDI specifications
-* Fix error in postun scriplet on RPM-based systems
+* Fix possible blank `nvidia-ctk` path in generated CDI specifications.
+* Fix error in postun scriplet on RPM-based systems.
 * Only check `NVIDIA_VISIBLE_DEVICES` for environment variables if no annotations are specified.
-* Add `cdi.default-kind` config option for constructing fully-qualified CDI device names in CDI mode
-* Add support for `accept-nvidia-visible-devices-envvar-unprivileged` config setting in CDI mode
+* Add `cdi.default-kind` config option for constructing fully-qualified CDI device names in CDI mode.
+* Add support for `accept-nvidia-visible-devices-envvar-unprivileged` config setting in CDI mode.
 * Add `nvidia-container-runtime-hook.skip-mode-detection` config option to bypass mode detection. This allows `legacy` and `cdi` mode, for example, to be used at the same time.
-* Add support for generating CDI specifications for GDS and MOFED devices
-* Ensure CDI specification is validated on save when generating a spec
+* Add support for generating CDI specifications for GDS and MOFED devices.
+* Ensure CDI specification is validated on save when generating a spec.
 * Rename `--discovery-mode` argument to `--mode` for `nvidia-ctk cdi generate`
-* [libnvidia-container] Fix segfault on WSL2 systems
-* [toolkit-container] Add `--cdi-enabled` flag to toolkit config
-* [toolkit-container] Install `nvidia-ctk` from toolkit container
-* [toolkit-container] Use installed `nvidia-ctk` path in NVIDIA Container Toolkit config
+* [libnvidia-container] Fix segfault on WSL2 systems.
+* [toolkit-container] Add `--cdi-enabled` flag to toolkit config.
+* [toolkit-container] Install `nvidia-ctk` from toolkit container.
+* [toolkit-container] Use installed `nvidia-ctk` path in NVIDIA Container Toolkit config.
 * [toolkit-container] Bump CUDA base images to 12.1.0
-* [toolkit-container] Set `nvidia-ctk` path in the
-* [toolkit-container] Add `cdi.k8s.io/*` to set of allowed annotations in containerd config
-* [toolkit-container] Generate CDI specification for use in management containers
-* [toolkit-container] Install experimental runtime as `nvidia-container-runtime.experimental` instead of `nvidia-container-runtime-experimental`
-* [toolkit-container] Install and configure mode-specific runtimes for `cdi` and `legacy` modes
+* [toolkit-container] Set `nvidia-ctk` path in the.
+* [toolkit-container] Add `cdi.k8s.io/*` to set of allowed annotations in containerd config.
+* [toolkit-container] Generate CDI specification for use in management containers.
+* [toolkit-container] Install experimental runtime as `nvidia-container-runtime.experimental` instead of `nvidia-container-runtime-experimental`.
+* [toolkit-container] Install and configure mode-specific runtimes for `cdi` and `legacy` modes.
 
 ## v1.13.0-rc.1
 
-* Include MIG-enabled devices as GPUs when generating CDI specification
-* Fix missing NVML symbols when running `nvidia-ctk` on some platforms [#49]
-* Add CDI spec generation for WSL2-based systems to `nvidia-ctk cdi generate` command
+* Include MIG-enabled devices as GPUs when generating CDI specification.
+* Fix missing NVML symbols when running `nvidia-ctk` on some platforms [#49].
+* Add CDI spec generation for WSL2-based systems to `nvidia-ctk cdi generate` command.
 * Add `auto` mode to `nvidia-ctk cdi generate` command to automatically detect a WSL2-based system over a standard NVML-based system.
-* Add mode-specific (`.cdi` and `.legacy`) NVIDIA Container Runtime binaries for use in the GPU Operator
+* Add mode-specific (`.cdi` and `.legacy`) NVIDIA Container Runtime binaries for use in the GPU Operator.
 * Discover all `gsb*.bin` GSP firmware files when generating CDI specification.
-* Align `.deb` and `.rpm` release candidate package versions
-* Remove `fedora35` packaging targets
-* [libnvidia-container] Include all `gsp*.bin` firmware files if present
-* [libnvidia-container] Align `.deb` and `.rpm` release candidate package versions
-* [libnvidia-container] Remove `fedora35` packaging targets
+* Align `.deb` and `.rpm` release candidate package versions.
+* Remove `fedora35` packaging targets.
+* [libnvidia-container] Include all `gsp*.bin` firmware files if present.
+* [libnvidia-container] Align `.deb` and `.rpm` release candidate package versions.
+* [libnvidia-container] Remove `fedora35` packaging targets.
 * [toolkit-container] Install `nvidia-container-toolkit-operator-extensions` package for mode-specific executables.
-* [toolkit-container] Allow `nvidia-container-runtime.mode` to be set when configuring the NVIDIA Container Toolkit
+* [toolkit-container] Allow `nvidia-container-runtime.mode` to be set when configuring the NVIDIA Container Toolkit.
 
 ## v1.12.0
 
-* Promote `v1.12.0-rc.5` to `v1.12.0`
-* Rename `nvidia cdi generate` `--root` flag to `--driver-root` to better indicate intent
-* [libnvidia-container] Add nvcubins.bin to DriverStore components under WSL2
+* Promote `v1.12.0-rc.5` to `v1.12.0`.
+* Rename `nvidia cdi generate` `--root` flag to `--driver-root` to better indicate intent.
+* [libnvidia-container] Add nvcubins.bin to DriverStore components under WSL2.
 * [toolkit-container] Bump CUDA base images to 12.0.1
 
 ## v1.12.0-rc.5
@@ -322,28 +322,28 @@
 
 ## v1.12.0-rc.3
 
-* Don't fail if by-path symlinks for DRM devices do not exist
-* Replace the --json flag with a --format [json|yaml] flag for the nvidia-ctk cdi generate command
-* Ensure that the CDI output folder is created if required
-* When generating a CDI specification use a blank host path for devices to ensure compatibility with the v0.4.0 CDI specification
-* Add injection of Wayland JSON files
-* Add GSP firmware paths to generated CDI specification
-* Add --root flag to nvidia-ctk cdi generate command
+* Don't fail if by-path symlinks for DRM devices do not exist.
+* Replace the --json flag with a --format [json|yaml] flag for the nvidia-ctk cdi generate command.
+* Ensure that the CDI output folder is created if required.
+* When generating a CDI specification use a blank host path for devices to ensure compatibility with the v0.4.0 CDI specification.
+* Add injection of Wayland JSON files.
+* Add GSP firmware paths to generated CDI specification.
+* Add --root flag to nvidia-ctk cdi generate command.
 
 ## v1.12.0-rc.2
 
-* Inject Direct Rendering Manager (DRM) devices into a container using the NVIDIA Container Runtime
-* Improve logging of errors from the NVIDIA Container Runtime
-* Improve CDI specification generation to support rootless podman
-* Use `nvidia-ctk cdi generate` to generate CDI specifications instead of `nvidia-ctk info generate-cdi`
-* [libnvidia-container] Skip creation of existing files when these are already mounted
+* Inject Direct Rendering Manager (DRM) devices into a container using the NVIDIA Container Runtime.
+* Improve logging of errors from the NVIDIA Container Runtime.
+* Improve CDI specification generation to support rootless podman.
+* Use `nvidia-ctk cdi generate` to generate CDI specifications instead of `nvidia-ctk info generate-cdi`.
+* [libnvidia-container] Skip creation of existing files when these are already mounted.
 
 ## v1.12.0-rc.1
 
-* Add support for multiple Docker Swarm resources
-* Improve injection of Vulkan configurations and libraries
-* Add `nvidia-ctk info generate-cdi` command to generated CDI specification for available devices
-* [libnvidia-container] Include NVVM compiler library in compute libs
+* Add support for multiple Docker Swarm resources.
+* Improve injection of Vulkan configurations and libraries.
+* Add `nvidia-ctk info generate-cdi` command to generated CDI specification for available devices.
+* [libnvidia-container] Include NVVM compiler library in compute libs.
 
 ## v1.11.0
 
@@ -351,22 +351,22 @@
 
 ## v1.11.0-rc.3
 
-* Build fedora35 packages
-* Introduce an `nvidia-container-toolkit-base` package for better dependency management
-* Fix removal of `nvidia-container-runtime-hook` on RPM-based systems
-* Inject platform files into container on Tegra-based systems
+* Build fedora35 packages.
+* Introduce an `nvidia-container-toolkit-base` package for better dependency management.
+* Fix removal of `nvidia-container-runtime-hook` on RPM-based systems.
+* Inject platform files into container on Tegra-based systems.
 * [toolkit container] Update CUDA base images to 11.7.1
-* [libnvidia-container] Preload libgcc_s.so.1 on arm64 systems
+* [libnvidia-container] Preload libgcc_s.so.1 on arm64 systems.
 
 ## v1.11.0-rc.2
 
-* Allow `accept-nvidia-visible-devices-*` config options to be set by toolkit container
+* Allow `accept-nvidia-visible-devices-*` config options to be set by toolkit container.
 * [libnvidia-container] Fix bug where LDCache was not updated when the `--no-pivot-root` option was specified
 
 ## v1.11.0-rc.1
 
-* Add discovery of GPUDirect Storage (`nvidia-fs*`) devices if the `NVIDIA_GDS` environment variable of the container is set to `enabled`
-* Add discovery of MOFED Infiniband devices if the `NVIDIA_MOFED` environment variable of the container is set to `enabled`
+* Add discovery of GPUDirect Storage (`nvidia-fs*`) devices if the `NVIDIA_GDS` environment variable of the container is set to `enabled`.
+* Add discovery of MOFED Infiniband devices if the `NVIDIA_MOFED` environment variable of the container is set to `enabled`.
 * Fix bug in CSV mode where libraries listed as `sym` entries in mount specification are not added to the LDCache.
 * Rename `nvidia-container-toolkit` executable to `nvidia-container-runtime-hook` and create `nvidia-container-toolkit` as a symlink to `nvidia-container-runtime-hook` instead.
 * Add `nvidia-ctk runtime configure` command to configure the Docker config file (e.g. `/etc/docker/daemon.json`) for use with the NVIDIA Container Runtime.
@@ -377,46 +377,46 @@
 
 ## v1.10.0-rc.3
 
-* Use default config instead of raising an error if config file cannot be found
-* Ignore NVIDIA_REQUIRE_JETPACK* environment variables for requirement checks
-* Fix bug in detection of Tegra systems where `/sys/devices/soc0/family` is ignored
-* Fix bug where links to devices were detected as devices
-* [libnvida-container] Fix bug introduced when adding libcudadebugger.so to list of libraries
+* Use default config instead of raising an error if config file cannot be found.
+* Ignore NVIDIA_REQUIRE_JETPACK* environment variables for requirement checks.
+* Fix bug in detection of Tegra systems where `/sys/devices/soc0/family` is ignored.
+* Fix bug where links to devices were detected as devices.
+* [libnvida-container] Fix bug introduced when adding libcudadebugger.so to list of libraries.
 
 ## v1.10.0-rc.2
 
-* Add support for NVIDIA_REQUIRE_* checks for cuda version and arch to csv mode
-* Switch to debug logging to reduce log verbosity
-* Support logging to logs requested in command line
-* Fix bug when launching containers with relative root path (e.g. using containerd)
-* Allow low-level runtime path to be set explicitly as nvidia-container-runtime.runtimes option
-* Fix failure to locate low-level runtime if PATH envvar is unset
-* Replace experimental option for NVIDIA Container Runtime with nvidia-container-runtime.mode = csv option
-* Use csv as default mode on Tegra systems without NVML
-* Add --version flag to all CLIs
+* Add support for NVIDIA_REQUIRE_* checks for cuda version and arch to csv mode.
+* Switch to debug logging to reduce log verbosity.
+* Support logging to logs requested in command line.
+* Fix bug when launching containers with relative root path (e.g. using containerd).
+* Allow low-level runtime path to be set explicitly as nvidia-container-runtime.runtimes option.
+* Fix failure to locate low-level runtime if PATH envvar is unset.
+* Replace experimental option for NVIDIA Container Runtime with nvidia-container-runtime.mode = csv option.
+* Use csv as default mode on Tegra systems without NVML.
+* Add --version flag to all CLIs.
 * [libnvidia-container] Bump libtirpc to 1.3.2
-* [libnvidia-container] Fix bug when running host ldconfig using glibc compiled with a non-standard prefix
-* [libnvidia-container] Add libcudadebugger.so to list of compute libraries
+* [libnvidia-container] Fix bug when running host ldconfig using glibc compiled with a non-standard prefix.
+* [libnvidia-container] Add libcudadebugger.so to list of compute libraries.
 
 ## v1.10.0-rc.1
 
-* Include nvidia-ctk CLI in installed binaries
-* Add experimental option to NVIDIA Container Runtime
+* Include nvidia-ctk CLI in installed binaries.
+* Add experimental option to NVIDIA Container Runtime.
 
 ## v1.9.0
 
-* [libnvidia-container] Add additional check for Tegra in /sys/.../family file in CLI
-* [libnvidia-container] Update jetpack-specific CLI option to only load Base CSV files by default
-* [libnvidia-container] Fix bug (from 1.8.0) when mounting GSP firmware into containers without /lib to /usr/lib symlinks
+* [libnvidia-container] Add additional check for Tegra in /sys/.../family file in CLI.
+* [libnvidia-container] Update jetpack-specific CLI option to only load Base CSV files by default.
+* [libnvidia-container] Fix bug (from 1.8.0) when mounting GSP firmware into containers without /lib to /usr/lib symlinks.
 * [libnvidia-container] Update nvml.h to CUDA 11.6.1 nvML_DEV 11.6.55
 * [libnvidia-container] Update switch statement to include new brands from latest nvml.h
-* [libnvidia-container] Process all --require flags on Jetson platforms
-* [libnvidia-container] Fix long-standing issue with running ldconfig on Debian systems
+* [libnvidia-container] Process all --require flags on Jetson platforms.
+* [libnvidia-container] Fix long-standing issue with running ldconfig on Debian systems.
 
 ## v1.8.1
 
-* [libnvidia-container] Fix bug in determining cgroup root when running in nested containers
-* [libnvidia-container] Fix permission issue when determining cgroup version
+* [libnvidia-container] Fix bug in determining cgroup root when running in nested containers.
+* [libnvidia-container] Fix permission issue when determining cgroup version.
 
 ## v1.8.0
 
@@ -424,12 +424,12 @@
 
 ## v1.8.0-rc.2
 
-* Remove support for building amazonlinux1 packages
+* Remove support for building amazonlinux1 packages.
 
 ## v1.8.0-rc.1
 
-* [libnvidia-container] Add support for cgroupv2
-* Release toolkit-container images from nvidia-container-toolkit repository
+* [libnvidia-container] Add support for cgroupv2.
+* Release toolkit-container images from nvidia-container-toolkit repository.
 
 ## v1.7.0
 
@@ -438,42 +438,42 @@
 
 ## v1.7.0-rc.1
 
-* Specify containerd runtime type as string in config tools to remove dependency on containerd package
-* Add supported-driver-capabilities config option to allow for a subset of all driver capabilities to be specified
+* Specify containerd runtime type as string in config tools to remove dependency on containerd package.
+* Add supported-driver-capabilities config option to allow for a subset of all driver capabilities to be specified.
 
 ## v1.6.0
 
 * Promote 1.6.0-rc.3-1 to 1.6.0
- * Fix unnecessary logging to stderr instead of configured nvidia-container-runtime log file
+ * Fix unnecessary logging to stderr instead of configured nvidia-container-runtime log file.
 
 ## v1.6.0-rc.3
 
-* Add supported-driver-capabilities config option to the nvidia-container-toolkit
-* Move OCI and command line checks for runtime to internal oci package
+* Add supported-driver-capabilities config option to the nvidia-container-toolkit.
+* Move OCI and command line checks for runtime to internal oci package.
 
 ## v1.6.0-rc.2
 
-* Use relative path to OCI specification file (config.json) if bundle path is not specified as an argument to the nvidia-container-runtime
+* Use relative path to OCI specification file (config.json) if bundle path is not specified as an argument to the nvidia-container-runtime.
 
 ## v1.6.0-rc.1
 
-* Add AARCH64 package for Amazon Linux 2
-* Include nvidia-container-runtime into nvidia-container-toolkit package
+* Add AARCH64 package for Amazon Linux 2.
+* Include nvidia-container-runtime into nvidia-container-toolkit package.
 
 ## v1.5.1
 
-* Fix bug where Docker Swarm device selection is ignored if NVIDIA_VISIBLE_DEVICES is also set
-* Improve unit testing by using require package and adding coverage reports
-* Remove unneeded go dependencies by running go mod tidy
-* Move contents of pkg directory to cmd for CLI tools
-* Ensure make binary target explicitly sets GOOS
+* Fix bug where Docker Swarm device selection is ignored if NVIDIA_VISIBLE_DEVICES is also set.
+* Improve unit testing by using require package and adding coverage reports.
+* Remove unneeded go dependencies by running go mod tidy.
+* Move contents of pkg directory to cmd for CLI tools.
+* Ensure make binary target explicitly sets GOOS.
 
 ## v1.5.0
 
 * Add dependence on libnvidia-container-tools >= 1.4.0
-* Add golang check targets to Makefile
-* Add Jenkinsfile definition for build targets
-* Move docker.mk to docker folder
+* Add golang check targets to Makefile.
+* Add Jenkinsfile definition for build targets.
+* Move docker.mk to docker folder.
 
 ## v1.4.2
 
@@ -481,12 +481,12 @@
 
 ## v1.4.1
 
-* Ignore NVIDIA_VISIBLE_DEVICES for containers with insufficent privileges
+* Ignore NVIDIA_VISIBLE_DEVICES for containers with insufficent privileges.
 * Add dependence on libnvidia-container-tools >= 1.3.2
 
 ## v1.4.0
 
-* Add 'compute' capability to list of defaults
+* Add 'compute' capability to list of defaults.
 * Add dependence on libnvidia-container-tools >= 1.3.1
 
 ## v1.3.0
@@ -496,34 +496,34 @@
 
 ## v1.3.0-rc.2
 
-* 2c180947 Add more tests for new semantics with device list from volume mounts
-* 7c003857 Refactor accepting device lists from volume mounts as a boolean
+* 2c180947 Add more tests for new semantics with device list from volume mounts.
+* 7c003857 Refactor accepting device lists from volume mounts as a boolean.
 
 ## v1.3.0-rc.1
 
 * b50d86c1 Update build system to accept a TAG variable for things like rc.x
 * fe65573b Add common CI tests for things like golint, gofmt, unit tests, etc.
-* da6fbb34 Revert "Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*"
-* a7fb3330 Flip build-all targets to run automatically on merge requests
-* 8b248b66 Rename github.com/NVIDIA/container-toolkit to nvidia-container-toolkit
-* da36874e Add new config options to pull device list from mounted files instead of ENVVAR
+* da6fbb34 Revert "Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*".
+* a7fb3330 Flip build-all targets to run automatically on merge requests.
+* 8b248b66 Rename github.com/NVIDIA/container-toolkit to nvidia-container-toolkit.
+* da36874e Add new config options to pull device list from mounted files instead of ENVVAR.
 
 ## v1.2.1
 
-* 4e6e0ed4 Add 'ngx' to list of*all* driver capabilities
-* 2f4af743 List config.toml as a config file in the RPM SPEC
+* 4e6e0ed4 Add 'ngx' to list of*all* driver capabilities.
+* 2f4af743 List config.toml as a config file in the RPM SPEC.
 
 ## v1.2.0
 
-*  8e0aab46 Fix repo listed in changelog for debian distributions
+*  8e0aab46 Fix repo listed in changelog for debian distributions.
 *  320bb6e4 Update dependence on libnvidia-container to 1.2.0
-*  6cfc8097 Update package license to match source license
-*  e7dc3cbb Fix debian copyright file
-*  d3aee3e0 Add the 'ngx' driver capability
+*  6cfc8097 Update package license to match source license.
+*  e7dc3cbb Fix debian copyright file.
+*  d3aee3e0 Add the 'ngx' driver capability.
 
 ## v1.1.2
 
-* c32237f3 Add support for parsing Linux Capabilities for older OCI specs
+* c32237f3 Add support for parsing Linux Capabilities for older OCI specs.
 
 ## v1.1.1
 
@@ -531,10 +531,10 @@
 
 ## v1.1.0
 
-* 4e4de762 Update build system to support multi-arch builds
-* fcc1d116 Add support for MIG (Multi-Instance GPUs)
+* 4e4de762 Update build system to support multi-arch builds.
+* fcc1d116 Add support for MIG (Multi-Instance GPUs).
 * d4ff0416 Add ability to merge envars of the form NVIDIA_VISIBLE_DEVICES_*
-* 60f165ad Add no-pivot option to toolkit
+* 60f165ad Add no-pivot option to toolkit.
 
 ## v1.0.5