Loading pkgs/development/python-modules/llama-cloud-services/default.nix +2 −2 Original line number Diff line number Diff line Loading @@ -17,14 +17,14 @@ buildPythonPackage rec { pname = "llama-cloud-services"; version = "0.6.36"; version = "0.6.37"; pyproject = true; src = fetchFromGitHub { owner = "run-llama"; repo = "llama_cloud_services"; tag = "v${version}"; hash = "sha256-m3XC4CiDUJemy7enWMG5iYGX6s7LGSfc16vGmDWhAic="; hash = "sha256-Ag8HZZjKgJk6D9uipntydYwVS8TKqZH7pOWaJ9EPuE0="; }; pythonRelaxDeps = [ "llama-cloud" ]; Loading pkgs/development/python-modules/llama-cloud/default.nix +2 −2 Original line number Diff line number Diff line Loading @@ -10,7 +10,7 @@ buildPythonPackage rec { pname = "llama-cloud"; version = "0.1.23"; version = "0.1.29"; pyproject = true; disabled = pythonOlder "3.8"; Loading @@ -18,7 +18,7 @@ buildPythonPackage rec { src = fetchPypi { pname = "llama_cloud"; inherit version; hash = "sha256-PYSiSoYPBG05oQbAZ0LsDqOaV0rEK7+RcG/gJfROIz4="; hash = "sha256-aZWUPtTx0f5lT36l0BlwYRxdxwYC3q+TeGt/lUXnDx4="; }; build-system = [ poetry-core ]; Loading pkgs/development/python-modules/llama-index-cli/default.nix +4 −7 Original line number Diff line number Diff line Loading @@ -2,28 +2,25 @@ lib, buildPythonPackage, fetchPypi, hatchling, llama-index-core, llama-index-embeddings-openai, llama-index-llms-openai, llama-index-vector-stores-chroma, poetry-core, pythonOlder, }: buildPythonPackage rec { pname = "llama-index-cli"; version = "0.4.1"; version = "0.4.3"; pyproject = true; disabled = pythonOlder "3.8"; src = fetchPypi { pname = "llama_index_cli"; inherit version; hash = "sha256-P5fx+PX0Ad+1trxxcHF8F23NmBU4AXQwBz7xL/3L3fo="; hash = "sha256-2ugYOhBVG72JaGuU7SlKbPRGM8PdYoXE+ZHIUDG3pV8="; }; build-system = [ poetry-core ]; build-system = [ hatchling ]; dependencies = [ llama-index-core Loading pkgs/development/python-modules/llama-index-core/default.nix +12 −4 Original line number Diff line number Diff line Loading @@ -12,6 +12,7 @@ fsspec, hatchling, jsonpath-ng, llama-index-workflows, llamaindex-py-client, nest-asyncio, networkx, Loading @@ -38,16 +39,16 @@ buildPythonPackage rec { pname = "llama-index-core"; version = "0.12.37"; version = "0.12.44"; pyproject = true; disabled = pythonOlder "3.8"; disabled = pythonOlder "3.9"; src = fetchFromGitHub { owner = "run-llama"; repo = "llama_index"; tag = "v${version}"; hash = "sha256-M6DiCJZu9mtb8NxzEiBsbpLJmpStNScTtHdr70H7Dvk="; hash = "sha256-i/aH/PU2e03jy6dWYhrn2QhTrc4UMr7cRGqhkbMbqug="; }; sourceRoot = "${src.name}/${pname}"; Loading @@ -66,7 +67,10 @@ buildPythonPackage rec { cp -r ${nltk-data.punkt}/tokenizers/punkt/* llama_index/core/_static/nltk_cache/tokenizers/punkt/ ''; pythonRelaxDeps = [ "tenacity" ]; pythonRelaxDeps = [ "setuptools" "tenacity" ]; build-system = [ hatchling ]; Loading @@ -80,6 +84,7 @@ buildPythonPackage rec { filetype fsspec jsonpath-ng llama-index-workflows llamaindex-py-client nest-asyncio networkx Loading Loading @@ -129,6 +134,8 @@ buildPythonPackage rec { "tests/text_splitter/" "tests/token_predictor/" "tests/tools/" "tests/schema/" "tests/multi_modal_llms/" ]; disabledTests = [ Loading @@ -144,6 +151,7 @@ buildPythonPackage rec { "test_from_persist_dir" "test_mimetype_raw_data" "test_multiple_documents_context" "test_resource" # asyncio.exceptions.InvalidStateError: invalid state "test_workflow_context_to_dict_mid_run" "test_SimpleDirectoryReader" Loading pkgs/development/python-modules/llama-index-embeddings-huggingface/default.nix +2 −2 Original line number Diff line number Diff line Loading @@ -10,7 +10,7 @@ buildPythonPackage rec { pname = "llama-index-embeddings-huggingface"; version = "0.5.4"; version = "0.5.5"; pyproject = true; disabled = pythonOlder "3.8"; Loading @@ -18,7 +18,7 @@ buildPythonPackage rec { src = fetchPypi { pname = "llama_index_embeddings_huggingface"; inherit version; hash = "sha256-nFjFrPt38Heo5DZgsbA3RgEmqNGAwXVwP/MeidqxnT4="; hash = "sha256-f26aAx2RRvI131l8DM1igM3pa5tDf5kFLOebty5frF4="; }; build-system = [ hatchling ]; Loading Loading
pkgs/development/python-modules/llama-cloud-services/default.nix +2 −2 Original line number Diff line number Diff line Loading @@ -17,14 +17,14 @@ buildPythonPackage rec { pname = "llama-cloud-services"; version = "0.6.36"; version = "0.6.37"; pyproject = true; src = fetchFromGitHub { owner = "run-llama"; repo = "llama_cloud_services"; tag = "v${version}"; hash = "sha256-m3XC4CiDUJemy7enWMG5iYGX6s7LGSfc16vGmDWhAic="; hash = "sha256-Ag8HZZjKgJk6D9uipntydYwVS8TKqZH7pOWaJ9EPuE0="; }; pythonRelaxDeps = [ "llama-cloud" ]; Loading
pkgs/development/python-modules/llama-cloud/default.nix +2 −2 Original line number Diff line number Diff line Loading @@ -10,7 +10,7 @@ buildPythonPackage rec { pname = "llama-cloud"; version = "0.1.23"; version = "0.1.29"; pyproject = true; disabled = pythonOlder "3.8"; Loading @@ -18,7 +18,7 @@ buildPythonPackage rec { src = fetchPypi { pname = "llama_cloud"; inherit version; hash = "sha256-PYSiSoYPBG05oQbAZ0LsDqOaV0rEK7+RcG/gJfROIz4="; hash = "sha256-aZWUPtTx0f5lT36l0BlwYRxdxwYC3q+TeGt/lUXnDx4="; }; build-system = [ poetry-core ]; Loading
pkgs/development/python-modules/llama-index-cli/default.nix +4 −7 Original line number Diff line number Diff line Loading @@ -2,28 +2,25 @@ lib, buildPythonPackage, fetchPypi, hatchling, llama-index-core, llama-index-embeddings-openai, llama-index-llms-openai, llama-index-vector-stores-chroma, poetry-core, pythonOlder, }: buildPythonPackage rec { pname = "llama-index-cli"; version = "0.4.1"; version = "0.4.3"; pyproject = true; disabled = pythonOlder "3.8"; src = fetchPypi { pname = "llama_index_cli"; inherit version; hash = "sha256-P5fx+PX0Ad+1trxxcHF8F23NmBU4AXQwBz7xL/3L3fo="; hash = "sha256-2ugYOhBVG72JaGuU7SlKbPRGM8PdYoXE+ZHIUDG3pV8="; }; build-system = [ poetry-core ]; build-system = [ hatchling ]; dependencies = [ llama-index-core Loading
pkgs/development/python-modules/llama-index-core/default.nix +12 −4 Original line number Diff line number Diff line Loading @@ -12,6 +12,7 @@ fsspec, hatchling, jsonpath-ng, llama-index-workflows, llamaindex-py-client, nest-asyncio, networkx, Loading @@ -38,16 +39,16 @@ buildPythonPackage rec { pname = "llama-index-core"; version = "0.12.37"; version = "0.12.44"; pyproject = true; disabled = pythonOlder "3.8"; disabled = pythonOlder "3.9"; src = fetchFromGitHub { owner = "run-llama"; repo = "llama_index"; tag = "v${version}"; hash = "sha256-M6DiCJZu9mtb8NxzEiBsbpLJmpStNScTtHdr70H7Dvk="; hash = "sha256-i/aH/PU2e03jy6dWYhrn2QhTrc4UMr7cRGqhkbMbqug="; }; sourceRoot = "${src.name}/${pname}"; Loading @@ -66,7 +67,10 @@ buildPythonPackage rec { cp -r ${nltk-data.punkt}/tokenizers/punkt/* llama_index/core/_static/nltk_cache/tokenizers/punkt/ ''; pythonRelaxDeps = [ "tenacity" ]; pythonRelaxDeps = [ "setuptools" "tenacity" ]; build-system = [ hatchling ]; Loading @@ -80,6 +84,7 @@ buildPythonPackage rec { filetype fsspec jsonpath-ng llama-index-workflows llamaindex-py-client nest-asyncio networkx Loading Loading @@ -129,6 +134,8 @@ buildPythonPackage rec { "tests/text_splitter/" "tests/token_predictor/" "tests/tools/" "tests/schema/" "tests/multi_modal_llms/" ]; disabledTests = [ Loading @@ -144,6 +151,7 @@ buildPythonPackage rec { "test_from_persist_dir" "test_mimetype_raw_data" "test_multiple_documents_context" "test_resource" # asyncio.exceptions.InvalidStateError: invalid state "test_workflow_context_to_dict_mid_run" "test_SimpleDirectoryReader" Loading
pkgs/development/python-modules/llama-index-embeddings-huggingface/default.nix +2 −2 Original line number Diff line number Diff line Loading @@ -10,7 +10,7 @@ buildPythonPackage rec { pname = "llama-index-embeddings-huggingface"; version = "0.5.4"; version = "0.5.5"; pyproject = true; disabled = pythonOlder "3.8"; Loading @@ -18,7 +18,7 @@ buildPythonPackage rec { src = fetchPypi { pname = "llama_index_embeddings_huggingface"; inherit version; hash = "sha256-nFjFrPt38Heo5DZgsbA3RgEmqNGAwXVwP/MeidqxnT4="; hash = "sha256-f26aAx2RRvI131l8DM1igM3pa5tDf5kFLOebty5frF4="; }; build-system = [ hatchling ]; Loading