nixpkgs/pkgs/development/cuda-modules
2024-02-16 19:04:55 +00:00
..
cuda cudaPackages.cuda_nvcc: back-end cc already exposes the c++ stdlib 2024-01-20 17:22:44 +00:00
cuda-library-samples cudaPackages.cuda{,-library}-samples: move to cuda-modules/ 2024-01-12 20:24:50 +00:00
cuda-samples cudaPackages.cuda{,-library}-samples: move to cuda-modules/ 2024-01-12 20:24:50 +00:00
cudatoolkit cudaPackages.cudatoolkit: add libglvnd to inputs 2024-02-01 12:14:18 +01:00
cudnn tree-wide: cudaPackages should not break default eval 2024-01-10 01:15:01 +00:00
cutensor tree-wide: cudaPackages should not break default eval 2024-01-10 01:15:01 +00:00
generic-builders cudaPackage: don't rely on non-portable builtins 2024-01-31 21:32:25 +01:00
modules Apply markdown lints to CUDA READMEs 2023-12-12 10:39:18 +01:00
nccl cudaPackages.nccl: 2.19.3-1 -> 2.20.3-1 2024-02-16 19:04:55 +00:00
nccl-tests cudaPackages.nccl-tests: 2.13.8 -> 2.13.9 2024-02-07 15:19:31 +00:00
saxpy cudaPackages_11_4.saxpy: clean(er) eval error 2024-01-12 20:24:50 +00:00
setup-hooks cudaPackages.autoAddCudaCompatRunpathHook: Correctly inherit platforms from cuda_compat 2024-01-17 15:34:40 +11:00
tensorrt tree-wide: cudaPackages should not break default eval 2024-01-10 01:15:01 +00:00
backend-stdenv.nix cudaPackages.backendStdenv: fix eval error blocking the channel 2024-01-21 12:33:18 +00:00
flags.nix tree-wide: cudaPackages should not break default eval 2024-01-10 01:15:01 +00:00
gpus.nix cudaPackages: move derivations to cuda-modules & support aarch64 2023-12-07 16:45:54 +00:00
nvcc-compatibilities.nix cudaPackages_12_3: init at 12.3.0 2023-12-07 16:45:56 +00:00
README.md Add basic documentation on cuda_compat 2023-12-12 16:02:55 +01:00

Cuda modules

Note

This document is meant to help CUDA maintainers understand the structure of the CUDA packages in Nixpkgs. It is not meant to be a user-facing document. For a user-facing document, see the CUDA section of the manual.

The files in this directory are added (in some way) to the cudaPackages package set by cuda-packages.nix.

Top-level files

Top-level nix files are included in the initial creation of the cudaPackages scope. These are typically required for the creation of the finalized cudaPackages scope:

  • backend-stdenv.nix: Standard environment for CUDA packages.
  • flags.nix: Flags set, or consumed by, NVCC in order to build packages.
  • gpus.nix: A list of supported NVIDIA GPUs.
  • nvcc-compatibilities.nix: NVCC releases and the version range of GCC/Clang they support.

Top-level directories

  • cuda: CUDA redistributables! Provides extension to cudaPackages scope.
  • cudatoolkit: monolothic CUDA Toolkit run-file installer. Provides extension to cudaPackages scope.
  • cudnn: NVIDIA cuDNN library.
  • cutensor: NVIDIA cuTENSOR library.
  • generic-builders:
    • Contains a builder manifest.nix which operates on the Manifest type defined in modules/generic/manifests. Most packages are built using this builder.
    • Contains a builder multiplex.nix which leverages the Manifest builder. In short, the Multiplex builder adds multiple versions of a single package to single instance of the CUDA Packages package set. It is used primarily for packages like cudnn and cutensor.
  • modules: Nixpkgs modules to check the shape and content of CUDA redistributable and feature manifests. These modules additionally use shims provided by some CUDA packages to allow them to re-use the genericManifestBuilder, even if they don't have manifest files of their own. cudnn and tensorrt are examples of packages which provide such shims. These modules are further described in the Modules documentation.
  • nccl: NVIDIA NCCL library.
  • nccl-tests: NVIDIA NCCL tests.
  • saxpy: Example CMake project that uses CUDA.
  • setup-hooks: Nixpkgs setup hooks for CUDA.
  • tensorrt: NVIDIA TensorRT library.

Distinguished packages

Cuda compatibility

Cuda Compatibility, available as cudaPackages.cuda_compat, is a component which makes it possible to run applications built against a newer CUDA toolkit (for example CUDA 12) on a machine with an older CUDA driver (for example CUDA 11), which isn't possible out of the box. At the time of writing, Cuda Compatibility is only available on the Nvidia Jetson architecture, but Nvidia might release support for more architectures in the future.

As Cuda Compatibility strictly increases the range of supported applications, we try our best to enable it by default on supported platforms.

Functioning

cuda_compat simply provides a new libcuda.so (and associated variants) that needs to be used in place of the default CUDA driver's libcuda.so. However, the other shared libraries of the default driver must still be accessible: cuda_compat isn't a complete drop-in replacement for the driver (and that's the point, otherwise, it would just be a newer driver).

Nvidia's recommendation is to set LD_LIBRARY_PATH to points to cuda_compat's driver. This is fine for a manual, one-shot usage, but in general setting LD_LIBRARY_PATH is a red flag. This is global state which short-circuits most of other dynamic libraries resolution mechanisms and can break things in non-obvious ways, especially with other Nix-built software.

Cuda compat with Nix

Since cuda_compat is a known derivation, the easy way to do this in Nix would be to add cuda_compat as a dependency of CUDA libraries and applications and let Nix does its magic by filling the DT_RUNPATH fields. However, cuda_compat itself depends on libnvrm_mem and libnvrm_gpu which are loaded dynamically at runtime from /run/opengl-driver. This doesn't please the Nix sandbox when building, which can't find those (a second minor issue is that addOpenGLRunpathHook prepends the /run/opengl-driver path, so that would still take precedence).

The current solution is to do something similar to addOpenGLRunpathHook: the addCudaCompatRunpathHook prepends to the path to cuda_compat's libcuda.so to the DT_RUNPATH of whichever package includes the hook as a dependency, and we include the hook by default for packages in cudaPackages (by adding it as a inputs in genericManifestBuilder). We also make sure it's included after addOpenGLRunpathHook, so that it appears before in the DT_RUNPATH and takes precedence.