2023-11-07 14:35:11 +00:00
|
|
|
# Notes:
|
|
|
|
#
|
|
|
|
# Silvan (Tweag) covered some things on recursive attribute sets in the Nix Hour:
|
|
|
|
# https://www.youtube.com/watch?v=BgnUFtd1Ivs
|
|
|
|
#
|
|
|
|
# I (@connorbaker) highly recommend watching it.
|
|
|
|
#
|
|
|
|
# Most helpful comment regarding recursive attribute sets:
|
|
|
|
#
|
|
|
|
# https://github.com/NixOS/nixpkgs/pull/256324#issuecomment-1749935979
|
|
|
|
#
|
|
|
|
# To summarize:
|
|
|
|
#
|
|
|
|
# - `prev` should only be used to access attributes which are going to be overriden.
|
|
|
|
# - `final` should only be used to access `callPackage` to build new packages.
|
|
|
|
# - Attribute names should be computable without relying on `final`.
|
|
|
|
# - Extensions should take arguments to build attribute names before relying on `final`.
|
|
|
|
#
|
2024-07-04 21:59:11 +00:00
|
|
|
# Silvan's recommendation then is to explicitly use `callPackage` to provide everything our
|
|
|
|
# extensions need to compute the attribute names, without relying on `final`.
|
2023-11-07 14:35:11 +00:00
|
|
|
#
|
|
|
|
# I've (@connorbaker) attempted to do that, though I'm unsure of how this will interact with overrides.
|
|
|
|
{
|
|
|
|
callPackage,
|
|
|
|
cudaVersion,
|
|
|
|
lib,
|
|
|
|
newScope,
|
|
|
|
pkgs,
|
2024-03-19 16:35:08 +00:00
|
|
|
config,
|
2022-04-03 09:19:04 +00:00
|
|
|
}:
|
|
|
|
let
|
2023-11-07 14:35:11 +00:00
|
|
|
inherit (lib)
|
|
|
|
attrsets
|
|
|
|
customisation
|
|
|
|
fixedPoints
|
2024-09-16 23:18:03 +00:00
|
|
|
lists
|
2023-11-07 14:35:11 +00:00
|
|
|
strings
|
2024-09-16 23:18:03 +00:00
|
|
|
trivial
|
2023-11-07 14:35:11 +00:00
|
|
|
versions
|
|
|
|
;
|
|
|
|
# Backbone
|
|
|
|
gpus = builtins.import ../development/cuda-modules/gpus.nix;
|
|
|
|
nvccCompatibilities = builtins.import ../development/cuda-modules/nvcc-compatibilities.nix;
|
2024-03-27 20:16:29 +00:00
|
|
|
flags = callPackage ../development/cuda-modules/flags.nix { inherit cudaVersion gpus; };
|
|
|
|
passthruFunction = final: ({
|
|
|
|
inherit cudaVersion lib pkgs;
|
|
|
|
inherit gpus nvccCompatibilities flags;
|
|
|
|
cudaMajorVersion = versions.major cudaVersion;
|
|
|
|
cudaMajorMinorVersion = versions.majorMinor cudaVersion;
|
|
|
|
cudaOlder = strings.versionOlder cudaVersion;
|
|
|
|
cudaAtLeast = strings.versionAtLeast cudaVersion;
|
2023-11-07 14:35:11 +00:00
|
|
|
|
2024-03-27 20:16:29 +00:00
|
|
|
# Maintain a reference to the final cudaPackages.
|
2024-07-04 21:59:11 +00:00
|
|
|
# Without this, if we use `final.callPackage` and a package accepts `cudaPackages` as an
|
|
|
|
# argument, it's provided with `cudaPackages` from the top-level scope, which is not what we
|
|
|
|
# want. We want to provide the `cudaPackages` from the final scope -- that is, the *current*
|
|
|
|
# scope. However, we also want to prevent `pkgs/top-level/release-attrpaths-superset.nix` from
|
|
|
|
# recursing more than one level here.
|
|
|
|
cudaPackages = final // {
|
|
|
|
__attrsFailEvaluation = true;
|
|
|
|
};
|
2023-11-07 14:35:11 +00:00
|
|
|
|
2024-03-27 20:16:29 +00:00
|
|
|
# TODO(@connorbaker): `cudaFlags` is an alias for `flags` which should be removed in the future.
|
|
|
|
cudaFlags = flags;
|
2023-11-07 14:35:11 +00:00
|
|
|
|
2024-03-27 20:16:29 +00:00
|
|
|
# Exposed as cudaPackages.backendStdenv.
|
|
|
|
# This is what nvcc uses as a backend,
|
|
|
|
# and it has to be an officially supported one (e.g. gcc11 for cuda11).
|
|
|
|
#
|
|
|
|
# It, however, propagates current stdenv's libstdc++ to avoid "GLIBCXX_* not found errors"
|
|
|
|
# when linked with other C++ libraries.
|
|
|
|
# E.g. for cudaPackages_11_8 we use gcc11 with gcc12's libstdc++
|
|
|
|
# Cf. https://github.com/NixOS/nixpkgs/pull/218265 for context
|
|
|
|
backendStdenv = final.callPackage ../development/cuda-modules/backend-stdenv.nix { };
|
2023-11-07 14:35:11 +00:00
|
|
|
|
2024-03-27 20:16:29 +00:00
|
|
|
# Loose packages
|
2024-04-03 09:17:17 +00:00
|
|
|
|
|
|
|
# TODO: Move to aliases.nix once all Nixpkgs has migrated to the splayed CUDA packages
|
|
|
|
cudatoolkit = final.callPackage ../development/cuda-modules/cudatoolkit/redist-wrapper.nix { };
|
|
|
|
cudatoolkit-legacy-runfile = final.callPackage ../development/cuda-modules/cudatoolkit { };
|
|
|
|
|
2024-03-27 20:16:29 +00:00
|
|
|
saxpy = final.callPackage ../development/cuda-modules/saxpy { };
|
|
|
|
nccl = final.callPackage ../development/cuda-modules/nccl { };
|
|
|
|
nccl-tests = final.callPackage ../development/cuda-modules/nccl-tests { };
|
2024-06-26 00:29:42 +00:00
|
|
|
|
2024-09-16 23:18:03 +00:00
|
|
|
tests =
|
|
|
|
let
|
|
|
|
bools = [
|
|
|
|
true
|
|
|
|
false
|
|
|
|
];
|
|
|
|
configs = {
|
|
|
|
openCVFirst = bools;
|
|
|
|
useOpenCVDefaultCuda = bools;
|
|
|
|
useTorchDefaultCuda = bools;
|
|
|
|
};
|
|
|
|
builder =
|
|
|
|
{
|
|
|
|
openCVFirst,
|
|
|
|
useOpenCVDefaultCuda,
|
|
|
|
useTorchDefaultCuda,
|
|
|
|
}@config:
|
|
|
|
{
|
|
|
|
name = strings.concatStringsSep "-" (
|
|
|
|
[
|
|
|
|
"test"
|
|
|
|
(if openCVFirst then "opencv" else "torch")
|
|
|
|
]
|
|
|
|
++ lists.optionals (if openCVFirst then useOpenCVDefaultCuda else useTorchDefaultCuda) [
|
|
|
|
"with-default-cuda"
|
|
|
|
]
|
|
|
|
++ [
|
|
|
|
"then"
|
|
|
|
(if openCVFirst then "torch" else "opencv")
|
|
|
|
]
|
|
|
|
++ lists.optionals (if openCVFirst then useTorchDefaultCuda else useOpenCVDefaultCuda) [
|
|
|
|
"with-default-cuda"
|
|
|
|
]
|
|
|
|
);
|
|
|
|
value = final.callPackage ../development/cuda-modules/tests/opencv-and-torch config;
|
|
|
|
};
|
|
|
|
in
|
|
|
|
attrsets.listToAttrs (attrsets.mapCartesianProduct builder configs);
|
|
|
|
|
2024-07-20 12:03:36 +00:00
|
|
|
writeGpuTestPython = final.callPackage ../development/cuda-modules/write-gpu-test-python.nix { };
|
2024-03-27 20:16:29 +00:00
|
|
|
});
|
2023-11-07 14:35:11 +00:00
|
|
|
|
|
|
|
mkVersionedPackageName =
|
|
|
|
name: version:
|
|
|
|
strings.concatStringsSep "_" [
|
|
|
|
name
|
2024-03-27 20:16:29 +00:00
|
|
|
(strings.replaceStrings [ "." ] [ "_" ] (versions.majorMinor version))
|
2023-11-07 14:35:11 +00:00
|
|
|
];
|
|
|
|
|
2024-03-27 20:16:29 +00:00
|
|
|
composedExtension = fixedPoints.composeManyExtensions (
|
|
|
|
[
|
|
|
|
(import ../development/cuda-modules/setup-hooks/extension.nix)
|
|
|
|
(callPackage ../development/cuda-modules/cuda/extension.nix { inherit cudaVersion; })
|
2024-04-04 02:03:20 +00:00
|
|
|
(import ../development/cuda-modules/cuda/overrides.nix)
|
2024-03-27 20:16:29 +00:00
|
|
|
(callPackage ../development/cuda-modules/generic-builders/multiplex.nix {
|
|
|
|
inherit cudaVersion flags mkVersionedPackageName;
|
|
|
|
pname = "cudnn";
|
|
|
|
releasesModule = ../development/cuda-modules/cudnn/releases.nix;
|
|
|
|
shimsFn = ../development/cuda-modules/cudnn/shims.nix;
|
|
|
|
fixupFn = ../development/cuda-modules/cudnn/fixup.nix;
|
|
|
|
})
|
|
|
|
(callPackage ../development/cuda-modules/cutensor/extension.nix {
|
|
|
|
inherit cudaVersion flags mkVersionedPackageName;
|
|
|
|
})
|
|
|
|
(callPackage ../development/cuda-modules/generic-builders/multiplex.nix {
|
|
|
|
inherit cudaVersion flags mkVersionedPackageName;
|
|
|
|
pname = "tensorrt";
|
|
|
|
releasesModule = ../development/cuda-modules/tensorrt/releases.nix;
|
|
|
|
shimsFn = ../development/cuda-modules/tensorrt/shims.nix;
|
|
|
|
fixupFn = ../development/cuda-modules/tensorrt/fixup.nix;
|
|
|
|
})
|
|
|
|
(callPackage ../development/cuda-modules/cuda-samples/extension.nix { inherit cudaVersion; })
|
|
|
|
(callPackage ../development/cuda-modules/cuda-library-samples/extension.nix { })
|
|
|
|
]
|
|
|
|
++ lib.optionals config.allowAliases [ (import ../development/cuda-modules/aliases.nix) ]
|
|
|
|
);
|
2023-11-07 14:35:11 +00:00
|
|
|
|
|
|
|
cudaPackages = customisation.makeScope newScope (
|
|
|
|
fixedPoints.extends composedExtension passthruFunction
|
|
|
|
);
|
|
|
|
in
|
cudaPackages_{10*,11*}: warn about upcoming removal
We currently package all CUDA versions from 10.0 onwards. In
some cases, CUDA is the only thing preventing us from removing old
versions of GCC. Since we currently don’t deprecate or remove CUDA
versions, this will be an increasing drag on compiler maintenance in
Nixpkgs going forward unless we establish a sensible policy. After
discussing this with @SomeoneSerge in the context of old versions
of GCC, I learned that there was already a desire to remove at least
versions prior to 11.3, as those versions were only packaged in the
old “runfile” format, but that it was blocked on someone doing
the work to warn about the upcoming deprecation for a release cycle.
This change adds a release note and warnings indicating that CUDA 10.x
and 11.x will be removed in Nixpkgs 25.05, about 8 months from now.
I chose this version cut‐off because these versions of CUDA require
GCC < 12. GCC releases a major version every year, and seems to
support about four releases at a time, releasing the last update to
the oldest version and marking it as unsupported on their site around
the time of the release of the next major version. Therefore, by the
time of the 25.05 release, we should expect GCC 15 to be released
and GCC 11 to become unsupported. Adding a warning and communicating
the policy of only shipping CUDA versions that work with supported
compilers in the release notes means that we should be able to
clean up old versions as required without any issue or extensive
deprecation period in future, without obligating us to do so if there
is a strongly compelling reason to be more lenient. That should help
solve both shipping an indefinitely‐growing list of CUDA versions
and an indefinitely‐growing list of GCC and LLVM versions.
As I’m not a user of CUDA myself, I can’t be sure of how sensible
this version support policy is, but I think it’s fair to say that
it’s reasonable for Nixpkgs to choose not to maintain compiler
versions that are unsupported upstream just for the sake of versions
of CUDA that are also unmaintained. CUDA 11.x has not received an
update for two years already, and would only become unsupported in
Nixpkgs in over half a year’s time.
CUDA 10.x is currently unused in‐tree except for the unmaintained
Caffe and NVIDIA DCGM, which depends on multiple CUDA versions solely
so that it can provide plugins for those versions. The latest DCGM
version has already removed support for CUDA 10.x and is just awaiting
an update in Nixpkgs. They maintain a list of supported versions to
build plugins for in their CMake build system, so it should be simple
enough for us to only build support for the versions of CUDA that we
support in Nixpkgs.
From what I can tell, CUDA 11.x is currently used by the following
packages other than DCGM:
* `catboost`, because of
<https://github.com/catboost/catboost/issues/2540>. It looks like
upstream has since redesigned this part of their build system, so
perhaps the problem is no longer present, or would be easier to fix.
* `magma_2_6_2`, an old version from before upstream added CUDA
12 support. This seems okay to break to me; that version is not
maintained and will never be updated for new CUDA versions, and
the CUDA support is optional.
* `paddlepaddle`, which, uh, also requires OpenSSL 1.1 of all
things. <https://github.com/PaddlePaddle/Paddle/issues/67571>
states that PaddlePaddle supports up to 12.3.
* `python3Packages.cupy`, which is listed as “possibly incompatible
with cutensor 2.0 that comes with `cudaPackages_12`”. I’m
not sure what the “possibly” means here, but according to
<https://github.com/cupy/cupy/tree/v13.3.0?tab=readme-ov-file#installation>
they ship binary wheels using CUDA 12.x so I think this should
be fine.
* `python3Packages.tensorrt`, which supports CUDA 12.x going by
<https://github.com/NVIDIA/TensorRT/blob/release/10.4/CMakeLists.txt#L111>.
* TensorFlow, which has a link to
<https://www.tensorflow.org/install/source#gpu> above the
`python3Packages.tensorflow-bin` definition, but that page lists
the versions we package as supporting CUDA 12.x.
Given the years since CUDA 11.x received any update upstream, and the
seemingly very limited set of packages that truly require it, I think
the policy of being able to drop versions that require unsupported
compilers starting from the next Nixpkgs release is a reasonable
one, but of course I’m open to feedback from the CUDA maintainers
about this.
2024-09-15 16:16:06 +00:00
|
|
|
# We want to warn users about the upcoming deprecation of old CUDA
|
|
|
|
# versions, without breaking Nixpkgs CI with evaluation warnings. This
|
|
|
|
# gross hack ensures that the warning only triggers if aliases are
|
|
|
|
# enabled, which is true by default, but not for ofborg.
|
|
|
|
lib.warnIf (cudaPackages.cudaOlder "12.0" && config.allowAliases)
|
|
|
|
"CUDA versions older than 12.0 will be removed in Nixpkgs 25.05; see the 24.11 release notes for more information"
|
|
|
|
cudaPackages
|