mirror of
https://github.com/NixOS/nixpkgs.git
synced 2025-02-23 04:25:14 +00:00
python312Packages.llama-cpp-python: init at 0.3.1 (#349657)
This commit is contained in:
commit
aaa6191e18
103
pkgs/development/python-modules/llama-cpp-python/default.nix
Normal file
103
pkgs/development/python-modules/llama-cpp-python/default.nix
Normal file
@ -0,0 +1,103 @@
|
|||||||
|
{
|
||||||
|
lib,
|
||||||
|
stdenv,
|
||||||
|
buildPythonPackage,
|
||||||
|
cmake,
|
||||||
|
fetchFromGitHub,
|
||||||
|
gitUpdater,
|
||||||
|
ninja,
|
||||||
|
pathspec,
|
||||||
|
pyproject-metadata,
|
||||||
|
pytestCheckHook,
|
||||||
|
pythonOlder,
|
||||||
|
scikit-build-core,
|
||||||
|
llama-cpp-python,
|
||||||
|
|
||||||
|
config,
|
||||||
|
cudaSupport ? config.cudaSupport,
|
||||||
|
cudaPackages ? { },
|
||||||
|
|
||||||
|
diskcache,
|
||||||
|
jinja2,
|
||||||
|
numpy,
|
||||||
|
typing-extensions,
|
||||||
|
scipy,
|
||||||
|
huggingface-hub,
|
||||||
|
}:
|
||||||
|
let
|
||||||
|
version = "0.3.1";
|
||||||
|
in
|
||||||
|
buildPythonPackage {
|
||||||
|
pname = "llama-cpp-python";
|
||||||
|
inherit version;
|
||||||
|
pyproject = true;
|
||||||
|
|
||||||
|
disabled = pythonOlder "3.7";
|
||||||
|
|
||||||
|
stdenv = if cudaSupport then cudaPackages.backendStdenv else stdenv;
|
||||||
|
|
||||||
|
src = fetchFromGitHub {
|
||||||
|
owner = "abetlen";
|
||||||
|
repo = "llama-cpp-python";
|
||||||
|
rev = "refs/tags/v${version}";
|
||||||
|
hash = "sha256-eO1zvNJZBE5BCnbgbh00tFIRWBCWor1lIsrLXs/HFds=";
|
||||||
|
fetchSubmodules = true;
|
||||||
|
};
|
||||||
|
|
||||||
|
dontUseCmakeConfigure = true;
|
||||||
|
SKBUILD_CMAKE_ARGS = lib.strings.concatStringsSep ";" (
|
||||||
|
lib.optionals cudaSupport [
|
||||||
|
"-DGGML_CUDA=on"
|
||||||
|
"-DCUDAToolkit_ROOT=${lib.getDev cudaPackages.cuda_nvcc}"
|
||||||
|
"-DCMAKE_CUDA_COMPILER=${lib.getExe cudaPackages.cuda_nvcc}"
|
||||||
|
]
|
||||||
|
);
|
||||||
|
|
||||||
|
nativeBuildInputs = [
|
||||||
|
cmake
|
||||||
|
ninja
|
||||||
|
pathspec
|
||||||
|
pyproject-metadata
|
||||||
|
scikit-build-core
|
||||||
|
];
|
||||||
|
|
||||||
|
buildInputs = lib.optionals cudaSupport (
|
||||||
|
with cudaPackages;
|
||||||
|
[
|
||||||
|
cuda_cudart # cuda_runtime.h
|
||||||
|
cuda_cccl # <thrust/*>
|
||||||
|
libcublas # cublas_v2.h
|
||||||
|
]
|
||||||
|
);
|
||||||
|
|
||||||
|
propagatedBuildInputs = [
|
||||||
|
diskcache
|
||||||
|
jinja2
|
||||||
|
numpy
|
||||||
|
typing-extensions
|
||||||
|
];
|
||||||
|
|
||||||
|
nativeCheckInputs = [
|
||||||
|
pytestCheckHook
|
||||||
|
scipy
|
||||||
|
huggingface-hub
|
||||||
|
];
|
||||||
|
|
||||||
|
disabledTests = [
|
||||||
|
# tries to download model from huggingface-hub
|
||||||
|
"test_real_model"
|
||||||
|
"test_real_llama"
|
||||||
|
];
|
||||||
|
|
||||||
|
pythonImportsCheck = [ "llama_cpp" ];
|
||||||
|
|
||||||
|
passthru.updateScript = gitUpdater { rev-prefix = "v"; };
|
||||||
|
passthru.tests.llama-cpp-python = llama-cpp-python.override { cudaSupport = true; };
|
||||||
|
|
||||||
|
meta = {
|
||||||
|
description = "Python bindings for llama.cpp";
|
||||||
|
homepage = "https://github.com/abetlen/llama-cpp-python";
|
||||||
|
license = lib.licenses.mit;
|
||||||
|
maintainers = with lib.maintainers; [ kirillrdy ];
|
||||||
|
};
|
||||||
|
}
|
@ -7476,6 +7476,8 @@ self: super: with self; {
|
|||||||
|
|
||||||
lizard = callPackage ../development/python-modules/lizard { };
|
lizard = callPackage ../development/python-modules/lizard { };
|
||||||
|
|
||||||
|
llama-cpp-python = callPackage ../development/python-modules/llama-cpp-python { };
|
||||||
|
|
||||||
llama-cloud = callPackage ../development/python-modules/llama-cloud { };
|
llama-cloud = callPackage ../development/python-modules/llama-cloud { };
|
||||||
|
|
||||||
llama-index = callPackage ../development/python-modules/llama-index { };
|
llama-index = callPackage ../development/python-modules/llama-index { };
|
||||||
|
Loading…
Reference in New Issue
Block a user