Loading pkgs/by-name/lo/local-ai/package.nix +22 −18 Original line number Diff line number Diff line Loading @@ -20,6 +20,7 @@ ncurses, which, opencv, curl, enable_upx ? true, upx, Loading @@ -28,11 +29,11 @@ # https://github.com/NixOS/rfcs/pull/169 # CPU extensions enable_avx ? true, enable_avx2 ? true, enable_avx ? stdenv.hostPlatform.isx86_64, enable_avx2 ? stdenv.hostPlatform.isx86_64, enable_avx512 ? stdenv.hostPlatform.avx512Support, enable_f16c ? true, enable_fma ? true, enable_f16c ? stdenv.hostPlatform.isx86_64, enable_fma ? stdenv.hostPlatform.isx86_64, with_openblas ? false, openblas, Loading Loading @@ -105,10 +106,10 @@ let final: prev: { name = "llama-cpp-grpc"; src = fetchFromGitHub { owner = "ggml-org"; owner = "ggerganov"; repo = "llama.cpp"; rev = "300907b2110cc17b4337334dc397e05de2d8f5e0"; hash = "sha256-7jPgToZ8Xs+8DfXP5WFWZKhqFdYcZ9yFzWVKjvOttIA="; rev = "d6d2c2ab8c8865784ba9fef37f2b2de3f2134d33"; hash = "sha256-b9B5I3EbBFrkWc6RLXMWcCRKayyWjlGuQrogUcrISrc="; fetchSubmodules = true; }; postPatch = Loading @@ -116,7 +117,7 @@ let + '' cd examples cp -r --no-preserve=mode ${src}/backend/cpp/llama grpc-server cp llava/clip.* llava/llava.* grpc-server cp llava/clip* llava/llava.* grpc-server printf "\nadd_subdirectory(grpc-server)" >> CMakeLists.txt cp ${src}/backend/backend.proto grpc-server Loading @@ -137,6 +138,7 @@ let protobuf # provides also abseil_cpp as propagated build input grpc openssl curl ]; } )).override Loading Loading @@ -286,7 +288,7 @@ let src = fetchFromGitHub { owner = "PABannier"; repo = "bark.cpp"; rev = "v1.0.0"; tag = "v1.0.0"; hash = "sha256-wOcggRWe8lsUzEj/wqOAUlJVypgNFmit5ISs9fbwoCE="; fetchSubmodules = true; }; Loading @@ -303,10 +305,10 @@ let stable-diffusion = stdenv.mkDerivation { name = "stable-diffusion"; src = fetchFromGitHub { owner = "leejet"; owner = "richiejp"; repo = "stable-diffusion.cpp"; rev = "d46ed5e184b97c2018dc2e8105925bdb8775e02c"; hash = "sha256-5i2HjkdaQEmlUWeHucQyrS8zNS+xyB7Zj+1oA/xsv2k="; rev = "53e3b17eb3d0b5760ced06a1f98320b68b34aaae"; # branch cuda-fix hash = "sha256-z56jafOdibpX+XhRsrc7ieGbeug4bf737/UobqkpBV0="; fetchSubmodules = true; }; installPhase = '' Loading @@ -316,6 +318,9 @@ let | tar cf - --null --files-from - \ | tar xf - -C $out/build ''; cmakeFlags = [ (lib.cmakeFeature "GGML_BUILD_NUMBER" "1") ]; nativeBuildInputs = [ cmake ]; buildInputs = [ opencv ]; }; Loading @@ -331,12 +336,12 @@ let stdenv; pname = "local-ai"; version = "2.26.0"; version = "2.28.0"; src = fetchFromGitHub { owner = "go-skynet"; repo = "LocalAI"; rev = "v${version}"; hash = "sha256-eHylgEbPNQ8CVLlstkeQH6jqYOKfvel1uU5ro8DkLJs="; tag = "v${version}"; hash = "sha256-Hpz0dGkgasSY/FGO7mDzqsLjXut0LdQ9PUXGaURUOlY="; }; prepare-sources = Loading @@ -355,7 +360,7 @@ let self = buildGo123Module.override { stdenv = effectiveStdenv; } { inherit pname version src; vendorHash = "sha256-6loR8bvt5BlijufUBVDpxNS/cVCMmbaCwEhYpJKwGys="; vendorHash = "sha256-1OY/y1AeL0K+vOU4Jk/cj7rToVLC9EkkNhgifB+icDM="; env.NIX_CFLAGS_COMPILE = " -isystem ${opencv}/include/opencv4"; Loading @@ -364,7 +369,6 @@ let # TODO: add silero-vad sed -i Makefile \ -e '/mod download/ d' \ -e '/^ALL_GRPC_BACKENDS+=backend-assets\/grpc\/llama-cpp-fallback/ d' \ -e '/^ALL_GRPC_BACKENDS+=backend-assets\/grpc\/llama-cpp-avx/ d' \ -e '/^ALL_GRPC_BACKENDS+=backend-assets\/grpc\/llama-cpp-cuda/ d' \ -e '/^ALL_GRPC_BACKENDS+=backend-assets\/grpc\/silero-vad/ d' \ Loading @@ -383,7 +387,7 @@ let + '' shopt -s extglob mkdir -p backend-assets/grpc cp ${llama-cpp-grpc}/bin/grpc-server backend-assets/grpc/llama-cpp-avx2 cp ${llama-cpp-grpc}/bin/grpc-server backend-assets/grpc/llama-cpp-fallback cp ${llama-cpp-rpc}/bin/grpc-server backend-assets/grpc/llama-cpp-grpc mkdir -p backend/cpp/llama/llama.cpp Loading pkgs/by-name/lo/local-ai/tests.nix +1 −1 Original line number Diff line number Diff line Loading @@ -84,7 +84,7 @@ in nodes.machine = { imports = [ common-config ]; virtualisation.cores = 2; virtualisation.memorySize = 4096; virtualisation.memorySize = 3 * 4096; services.local-ai.models = models; }; passthru = { Loading Loading
pkgs/by-name/lo/local-ai/package.nix +22 −18 Original line number Diff line number Diff line Loading @@ -20,6 +20,7 @@ ncurses, which, opencv, curl, enable_upx ? true, upx, Loading @@ -28,11 +29,11 @@ # https://github.com/NixOS/rfcs/pull/169 # CPU extensions enable_avx ? true, enable_avx2 ? true, enable_avx ? stdenv.hostPlatform.isx86_64, enable_avx2 ? stdenv.hostPlatform.isx86_64, enable_avx512 ? stdenv.hostPlatform.avx512Support, enable_f16c ? true, enable_fma ? true, enable_f16c ? stdenv.hostPlatform.isx86_64, enable_fma ? stdenv.hostPlatform.isx86_64, with_openblas ? false, openblas, Loading Loading @@ -105,10 +106,10 @@ let final: prev: { name = "llama-cpp-grpc"; src = fetchFromGitHub { owner = "ggml-org"; owner = "ggerganov"; repo = "llama.cpp"; rev = "300907b2110cc17b4337334dc397e05de2d8f5e0"; hash = "sha256-7jPgToZ8Xs+8DfXP5WFWZKhqFdYcZ9yFzWVKjvOttIA="; rev = "d6d2c2ab8c8865784ba9fef37f2b2de3f2134d33"; hash = "sha256-b9B5I3EbBFrkWc6RLXMWcCRKayyWjlGuQrogUcrISrc="; fetchSubmodules = true; }; postPatch = Loading @@ -116,7 +117,7 @@ let + '' cd examples cp -r --no-preserve=mode ${src}/backend/cpp/llama grpc-server cp llava/clip.* llava/llava.* grpc-server cp llava/clip* llava/llava.* grpc-server printf "\nadd_subdirectory(grpc-server)" >> CMakeLists.txt cp ${src}/backend/backend.proto grpc-server Loading @@ -137,6 +138,7 @@ let protobuf # provides also abseil_cpp as propagated build input grpc openssl curl ]; } )).override Loading Loading @@ -286,7 +288,7 @@ let src = fetchFromGitHub { owner = "PABannier"; repo = "bark.cpp"; rev = "v1.0.0"; tag = "v1.0.0"; hash = "sha256-wOcggRWe8lsUzEj/wqOAUlJVypgNFmit5ISs9fbwoCE="; fetchSubmodules = true; }; Loading @@ -303,10 +305,10 @@ let stable-diffusion = stdenv.mkDerivation { name = "stable-diffusion"; src = fetchFromGitHub { owner = "leejet"; owner = "richiejp"; repo = "stable-diffusion.cpp"; rev = "d46ed5e184b97c2018dc2e8105925bdb8775e02c"; hash = "sha256-5i2HjkdaQEmlUWeHucQyrS8zNS+xyB7Zj+1oA/xsv2k="; rev = "53e3b17eb3d0b5760ced06a1f98320b68b34aaae"; # branch cuda-fix hash = "sha256-z56jafOdibpX+XhRsrc7ieGbeug4bf737/UobqkpBV0="; fetchSubmodules = true; }; installPhase = '' Loading @@ -316,6 +318,9 @@ let | tar cf - --null --files-from - \ | tar xf - -C $out/build ''; cmakeFlags = [ (lib.cmakeFeature "GGML_BUILD_NUMBER" "1") ]; nativeBuildInputs = [ cmake ]; buildInputs = [ opencv ]; }; Loading @@ -331,12 +336,12 @@ let stdenv; pname = "local-ai"; version = "2.26.0"; version = "2.28.0"; src = fetchFromGitHub { owner = "go-skynet"; repo = "LocalAI"; rev = "v${version}"; hash = "sha256-eHylgEbPNQ8CVLlstkeQH6jqYOKfvel1uU5ro8DkLJs="; tag = "v${version}"; hash = "sha256-Hpz0dGkgasSY/FGO7mDzqsLjXut0LdQ9PUXGaURUOlY="; }; prepare-sources = Loading @@ -355,7 +360,7 @@ let self = buildGo123Module.override { stdenv = effectiveStdenv; } { inherit pname version src; vendorHash = "sha256-6loR8bvt5BlijufUBVDpxNS/cVCMmbaCwEhYpJKwGys="; vendorHash = "sha256-1OY/y1AeL0K+vOU4Jk/cj7rToVLC9EkkNhgifB+icDM="; env.NIX_CFLAGS_COMPILE = " -isystem ${opencv}/include/opencv4"; Loading @@ -364,7 +369,6 @@ let # TODO: add silero-vad sed -i Makefile \ -e '/mod download/ d' \ -e '/^ALL_GRPC_BACKENDS+=backend-assets\/grpc\/llama-cpp-fallback/ d' \ -e '/^ALL_GRPC_BACKENDS+=backend-assets\/grpc\/llama-cpp-avx/ d' \ -e '/^ALL_GRPC_BACKENDS+=backend-assets\/grpc\/llama-cpp-cuda/ d' \ -e '/^ALL_GRPC_BACKENDS+=backend-assets\/grpc\/silero-vad/ d' \ Loading @@ -383,7 +387,7 @@ let + '' shopt -s extglob mkdir -p backend-assets/grpc cp ${llama-cpp-grpc}/bin/grpc-server backend-assets/grpc/llama-cpp-avx2 cp ${llama-cpp-grpc}/bin/grpc-server backend-assets/grpc/llama-cpp-fallback cp ${llama-cpp-rpc}/bin/grpc-server backend-assets/grpc/llama-cpp-grpc mkdir -p backend/cpp/llama/llama.cpp Loading
pkgs/by-name/lo/local-ai/tests.nix +1 −1 Original line number Diff line number Diff line Loading @@ -84,7 +84,7 @@ in nodes.machine = { imports = [ common-config ]; virtualisation.cores = 2; virtualisation.memorySize = 4096; virtualisation.memorySize = 3 * 4096; services.local-ai.models = models; }; passthru = { Loading