Loading pkgs/by-name/lo/local-ai/package.nix +17 −23 Original line number Diff line number Diff line Loading @@ -97,21 +97,17 @@ let llama-cpp-rpc = (llama-cpp-grpc.overrideAttrs (prev: { name = "llama-cpp-rpc"; cmakeFlags = prev.cmakeFlags ++ [ (lib.cmakeBool "LLAMA_AVX" false) (lib.cmakeBool "LLAMA_AVX2" false) (lib.cmakeBool "LLAMA_AVX512" false) (lib.cmakeBool "LLAMA_FMA" false) (lib.cmakeBool "LLAMA_F16C" false) (lib.cmakeBool "LLAMA_RPC" true) (lib.cmakeBool "GGML_AVX" false) (lib.cmakeBool "GGML_AVX2" false) (lib.cmakeBool "GGML_AVX512" false) (lib.cmakeBool "GGML_FMA" false) (lib.cmakeBool "GGML_F16C" false) ]; postPatch = prev.postPatch + '' sed -i examples/rpc/CMakeLists.txt \ -e '$a\install(TARGETS rpc-server RUNTIME)' ''; })).override { cudaSupport = false; openclSupport = false; blasSupport = false; rpcSupport = true; }; llama-cpp-grpc = (llama-cpp.overrideAttrs (final: prev: { Loading @@ -119,8 +115,8 @@ let src = fetchFromGitHub { owner = "ggerganov"; repo = "llama.cpp"; rev = "37bef8943312d91183ff06d8f1214082a17344a5"; hash = "sha256-E3kCMDK5TXozBsprp4D581WHTVP9aljxB1KZUKug1pM="; rev = "cb5fad4c6c2cbef92e9b8b63449e1cb7664e4846"; hash = "sha256-cIJuDC+MFLd5hkA1kUxuaw2dZagHqn5fi5Q2XKvDEII="; fetchSubmodules = true; }; postPatch = prev.postPatch + '' Loading @@ -137,14 +133,12 @@ let ''; cmakeFlags = prev.cmakeFlags ++ [ (lib.cmakeBool "BUILD_SHARED_LIBS" false) (lib.cmakeBool "LLAMA_AVX" enable_avx) (lib.cmakeBool "LLAMA_AVX2" enable_avx2) (lib.cmakeBool "LLAMA_AVX512" enable_avx512) (lib.cmakeBool "LLAMA_FMA" enable_fma) (lib.cmakeBool "LLAMA_F16C" enable_f16c) (lib.cmakeBool "GGML_AVX" enable_avx) (lib.cmakeBool "GGML_AVX2" enable_avx2) (lib.cmakeBool "GGML_AVX512" enable_avx512) (lib.cmakeBool "GGML_FMA" enable_fma) (lib.cmakeBool "GGML_F16C" enable_f16c) ]; postInstall = null; buildInputs = prev.buildInputs ++ [ protobuf # provides also abseil_cpp as propagated build input grpc Loading Loading @@ -394,18 +388,18 @@ let stdenv; pname = "local-ai"; version = "2.17.1"; version = "2.18.1"; src = fetchFromGitHub { owner = "go-skynet"; repo = "LocalAI"; rev = "v${version}"; hash = "sha256-G9My4t3vJ1sWyD+vxUgON4ezXURVAAgu1nAtTjd3ZR8="; hash = "sha256-hRrbGUUawQV4fqxAn3eFBvn4/lZ+NrKhxnGHqpljrec="; }; self = buildGoModule.override { stdenv = effectiveStdenv; } { inherit pname version src; vendorHash = "sha256-Hu7aJFi40CKNWAxYOR47VBZI1A/9SlBIVQVcB8iqcxA="; vendorHash = "sha256-uvko1PQWW5P+6cgmwVKocKBm5GndszqCsSbxlXANqJs="; env.NIX_CFLAGS_COMPILE = lib.optionalString with_stablediffusion " -isystem ${opencv}/include/opencv4"; Loading Loading @@ -441,7 +435,7 @@ let cp ${llama-cpp-rpc}/bin/grpc-server backend-assets/grpc/llama-cpp-grpc mkdir -p backend-assets/util cp ${llama-cpp-rpc}/bin/rpc-server backend-assets/util/llama-cpp-rpc-server cp ${llama-cpp-rpc}/bin/llama-rpc-server backend-assets/util/llama-cpp-rpc-server ''; buildInputs = [ ] Loading Loading
pkgs/by-name/lo/local-ai/package.nix +17 −23 Original line number Diff line number Diff line Loading @@ -97,21 +97,17 @@ let llama-cpp-rpc = (llama-cpp-grpc.overrideAttrs (prev: { name = "llama-cpp-rpc"; cmakeFlags = prev.cmakeFlags ++ [ (lib.cmakeBool "LLAMA_AVX" false) (lib.cmakeBool "LLAMA_AVX2" false) (lib.cmakeBool "LLAMA_AVX512" false) (lib.cmakeBool "LLAMA_FMA" false) (lib.cmakeBool "LLAMA_F16C" false) (lib.cmakeBool "LLAMA_RPC" true) (lib.cmakeBool "GGML_AVX" false) (lib.cmakeBool "GGML_AVX2" false) (lib.cmakeBool "GGML_AVX512" false) (lib.cmakeBool "GGML_FMA" false) (lib.cmakeBool "GGML_F16C" false) ]; postPatch = prev.postPatch + '' sed -i examples/rpc/CMakeLists.txt \ -e '$a\install(TARGETS rpc-server RUNTIME)' ''; })).override { cudaSupport = false; openclSupport = false; blasSupport = false; rpcSupport = true; }; llama-cpp-grpc = (llama-cpp.overrideAttrs (final: prev: { Loading @@ -119,8 +115,8 @@ let src = fetchFromGitHub { owner = "ggerganov"; repo = "llama.cpp"; rev = "37bef8943312d91183ff06d8f1214082a17344a5"; hash = "sha256-E3kCMDK5TXozBsprp4D581WHTVP9aljxB1KZUKug1pM="; rev = "cb5fad4c6c2cbef92e9b8b63449e1cb7664e4846"; hash = "sha256-cIJuDC+MFLd5hkA1kUxuaw2dZagHqn5fi5Q2XKvDEII="; fetchSubmodules = true; }; postPatch = prev.postPatch + '' Loading @@ -137,14 +133,12 @@ let ''; cmakeFlags = prev.cmakeFlags ++ [ (lib.cmakeBool "BUILD_SHARED_LIBS" false) (lib.cmakeBool "LLAMA_AVX" enable_avx) (lib.cmakeBool "LLAMA_AVX2" enable_avx2) (lib.cmakeBool "LLAMA_AVX512" enable_avx512) (lib.cmakeBool "LLAMA_FMA" enable_fma) (lib.cmakeBool "LLAMA_F16C" enable_f16c) (lib.cmakeBool "GGML_AVX" enable_avx) (lib.cmakeBool "GGML_AVX2" enable_avx2) (lib.cmakeBool "GGML_AVX512" enable_avx512) (lib.cmakeBool "GGML_FMA" enable_fma) (lib.cmakeBool "GGML_F16C" enable_f16c) ]; postInstall = null; buildInputs = prev.buildInputs ++ [ protobuf # provides also abseil_cpp as propagated build input grpc Loading Loading @@ -394,18 +388,18 @@ let stdenv; pname = "local-ai"; version = "2.17.1"; version = "2.18.1"; src = fetchFromGitHub { owner = "go-skynet"; repo = "LocalAI"; rev = "v${version}"; hash = "sha256-G9My4t3vJ1sWyD+vxUgON4ezXURVAAgu1nAtTjd3ZR8="; hash = "sha256-hRrbGUUawQV4fqxAn3eFBvn4/lZ+NrKhxnGHqpljrec="; }; self = buildGoModule.override { stdenv = effectiveStdenv; } { inherit pname version src; vendorHash = "sha256-Hu7aJFi40CKNWAxYOR47VBZI1A/9SlBIVQVcB8iqcxA="; vendorHash = "sha256-uvko1PQWW5P+6cgmwVKocKBm5GndszqCsSbxlXANqJs="; env.NIX_CFLAGS_COMPILE = lib.optionalString with_stablediffusion " -isystem ${opencv}/include/opencv4"; Loading Loading @@ -441,7 +435,7 @@ let cp ${llama-cpp-rpc}/bin/grpc-server backend-assets/grpc/llama-cpp-grpc mkdir -p backend-assets/util cp ${llama-cpp-rpc}/bin/rpc-server backend-assets/util/llama-cpp-rpc-server cp ${llama-cpp-rpc}/bin/llama-rpc-server backend-assets/util/llama-cpp-rpc-server ''; buildInputs = [ ] Loading