diff --git a/overrides/default.nix b/overrides/default.nix index aea9be8..f4f1797 100644 --- a/overrides/default.nix +++ b/overrides/default.nix @@ -2537,7 +2537,7 @@ lib.composeManyExtensions [ preFixup = '' addAutoPatchelfSearchPath "${self.torch}/${self.python.sitePackages}/torch/lib" ''; - + buildInputs = (old.buildInputs or [ ]) ++ [ self.torch ]; diff --git a/tests/default.nix b/tests/default.nix index 07f8f24..a56c798 100644 --- a/tests/default.nix +++ b/tests/default.nix @@ -105,7 +105,7 @@ builtins.removeAttrs rmfuse = skipOSX pkgs.rmfuse; ml-stack = callTest ./ml-stack { }; - + # Test building poetry inherit poetry; diff --git a/tests/ml-stack/default.nix b/tests/ml-stack/default.nix index 3b0d4c9..d23c2d1 100644 --- a/tests/ml-stack/default.nix +++ b/tests/ml-stack/default.nix @@ -7,10 +7,10 @@ let preferWheels = true; }; py = env.python; -in +in # Note: torch.cuda() will print False, even if you have a GPU, when this runs *during test.* -# But if you run the script below in your shell (rather than during build), it will print True. -# Presumably this is due to sandboxing. + # But if you run the script below in your shell (rather than during build), it will print True. + # Presumably this is due to sandboxing. runCommand "ml-stack-test" { } '' ${env}/bin/python -c 'import torch; import torchvision; print(torch.__version__); print(torchvision.__version__); print(torch.cuda.is_available()); x = torch.randn(3,3); x = x.cuda() if torch.cuda.is_available() else x; print(x**2)' > $out -'' \ No newline at end of file +''