Loading pkgs/by-name/ll/llama-cpp/package.nix +4 −4 Original line number Diff line number Diff line Loading @@ -82,13 +82,13 @@ let in effectiveStdenv.mkDerivation (finalAttrs: { pname = "llama-cpp"; version = "4731"; version = "4783"; src = fetchFromGitHub { owner = "ggerganov"; owner = "ggml-org"; repo = "llama.cpp"; tag = "b${finalAttrs.version}"; hash = "sha256-XB7TI1/Ui9BGdsD5TVH5MMZalgEGquW9wYcAhxgAcP0="; hash = "sha256-C/yDquq6fsrUW50R6CBo7xx5rgOiOxY1Cexmr+XgMo4="; leaveDotGit = true; postFetch = '' git -C "$out" rev-parse --short HEAD > $out/COMMIT Loading Loading @@ -194,7 +194,7 @@ effectiveStdenv.mkDerivation (finalAttrs: { meta = with lib; { description = "Inference of Meta's LLaMA model (and others) in pure C/C++"; homepage = "https://github.com/ggerganov/llama.cpp/"; homepage = "https://github.com/ggml-org/llama.cpp"; license = licenses.mit; mainProgram = "llama"; maintainers = with maintainers; [ Loading pkgs/by-name/lo/local-ai/package.nix +1 −1 Original line number Diff line number Diff line Loading @@ -149,7 +149,7 @@ let final: prev: { name = "llama-cpp-grpc"; src = fetchFromGitHub { owner = "ggerganov"; owner = "ggml-org"; repo = "llama.cpp"; rev = "ba8a1f9c5b675459c55a83e3f97f10df3a66c788"; hash = "sha256-YunQh1760AcknBFwHc6uMZJ7V4OzEAKiwB9HBH1n4bc="; Loading pkgs/development/python-modules/llama-cpp-python/default.nix +1 −1 Original line number Diff line number Diff line Loading @@ -55,7 +55,7 @@ buildPythonPackage rec { patches = [ # fix segfault when running tests due to missing default Metal devices (fetchpatch2 { url = "https://github.com/ggerganov/llama.cpp/commit/acd38efee316f3a5ed2e6afcbc5814807c347053.patch?full_index=1"; url = "https://github.com/ggml-org/llama.cpp/commit/acd38efee316f3a5ed2e6afcbc5814807c347053.patch?full_index=1"; stripLen = 1; extraPrefix = "vendor/llama.cpp/"; hash = "sha256-71+Lpg9z5KPlaQTX9D85KS2LXFWLQNJJ18TJyyq3/pU="; Loading Loading
pkgs/by-name/ll/llama-cpp/package.nix +4 −4 Original line number Diff line number Diff line Loading @@ -82,13 +82,13 @@ let in effectiveStdenv.mkDerivation (finalAttrs: { pname = "llama-cpp"; version = "4731"; version = "4783"; src = fetchFromGitHub { owner = "ggerganov"; owner = "ggml-org"; repo = "llama.cpp"; tag = "b${finalAttrs.version}"; hash = "sha256-XB7TI1/Ui9BGdsD5TVH5MMZalgEGquW9wYcAhxgAcP0="; hash = "sha256-C/yDquq6fsrUW50R6CBo7xx5rgOiOxY1Cexmr+XgMo4="; leaveDotGit = true; postFetch = '' git -C "$out" rev-parse --short HEAD > $out/COMMIT Loading Loading @@ -194,7 +194,7 @@ effectiveStdenv.mkDerivation (finalAttrs: { meta = with lib; { description = "Inference of Meta's LLaMA model (and others) in pure C/C++"; homepage = "https://github.com/ggerganov/llama.cpp/"; homepage = "https://github.com/ggml-org/llama.cpp"; license = licenses.mit; mainProgram = "llama"; maintainers = with maintainers; [ Loading
pkgs/by-name/lo/local-ai/package.nix +1 −1 Original line number Diff line number Diff line Loading @@ -149,7 +149,7 @@ let final: prev: { name = "llama-cpp-grpc"; src = fetchFromGitHub { owner = "ggerganov"; owner = "ggml-org"; repo = "llama.cpp"; rev = "ba8a1f9c5b675459c55a83e3f97f10df3a66c788"; hash = "sha256-YunQh1760AcknBFwHc6uMZJ7V4OzEAKiwB9HBH1n4bc="; Loading
pkgs/development/python-modules/llama-cpp-python/default.nix +1 −1 Original line number Diff line number Diff line Loading @@ -55,7 +55,7 @@ buildPythonPackage rec { patches = [ # fix segfault when running tests due to missing default Metal devices (fetchpatch2 { url = "https://github.com/ggerganov/llama.cpp/commit/acd38efee316f3a5ed2e6afcbc5814807c347053.patch?full_index=1"; url = "https://github.com/ggml-org/llama.cpp/commit/acd38efee316f3a5ed2e6afcbc5814807c347053.patch?full_index=1"; stripLen = 1; extraPrefix = "vendor/llama.cpp/"; hash = "sha256-71+Lpg9z5KPlaQTX9D85KS2LXFWLQNJJ18TJyyq3/pU="; Loading