Loading pkgs/development/python-modules/openai-whisper/default.nix +19 −28 Original line number Diff line number Diff line Loading @@ -2,22 +2,23 @@ , fetchFromGitHub , buildPythonPackage , substituteAll , cudaSupport ? false # build-system , setuptools # runtime , ffmpeg-headless # propagates , numpy , torch , torchWithCuda , tqdm , more-itertools , transformers , numba , numpy , openai-triton , scipy , tiktoken , torch , tqdm , transformers # tests , pytestCheckHook Loading @@ -25,14 +26,14 @@ buildPythonPackage rec { pname = "whisper"; version = "20230918"; format = "setuptools"; version = "20231117"; pyproject = true; src = fetchFromGitHub { owner = "openai"; repo = pname; rev = "refs/tags/v${version}"; hash = "sha256-wBAanFVEIIzTcoX40P9eI26UdEu0SC/xuife/zi2Xho="; hash = "sha256-MJ1XjB/GuYUiECCuuHS0NWHvvs+ko0oTvLuDI7zLNiY="; }; patches = [ Loading @@ -42,32 +43,22 @@ buildPythonPackage rec { }) ]; nativeBuildInputs = [ setuptools ]; propagatedBuildInputs = [ numpy tqdm more-itertools transformers numba numpy openai-triton scipy tiktoken ] ++ lib.optionals (!cudaSupport) [ torch ] ++ lib.optionals (cudaSupport) [ openai-triton torchWithCuda tqdm transformers ]; postPatch = '' substituteInPlace requirements.txt \ --replace "tiktoken==0.3.3" "tiktoken>=0.3.3" '' # openai-triton is only needed for CUDA support. # triton needs CUDA to be build. # -> by making it optional, we can build whisper without unfree packages enabled + lib.optionalString (!cudaSupport) '' sed -i '/if sys.platform.startswith("linux") and platform.machine() == "x86_64":/{N;d}' setup.py ''; preCheck = '' export HOME=$TMPDIR ''; Loading @@ -85,7 +76,7 @@ buildPythonPackage rec { ]; meta = with lib; { changelog = "https://github.com/openai/whisper/blob/v$[version}/CHANGELOG.md"; changelog = "https://github.com/openai/whisper/blob/v${version}/CHANGELOG.md"; description = "General-purpose speech recognition model"; homepage = "https://github.com/openai/whisper"; license = licenses.mit; Loading pkgs/top-level/python-packages.nix +1 −4 Original line number Diff line number Diff line Loading @@ -8470,10 +8470,7 @@ self: super: with self; { openai-triton-bin = callPackage ../development/python-modules/openai-triton/bin.nix { }; openai-whisper = callPackage ../development/python-modules/openai-whisper { inherit (pkgs.config) cudaSupport; openai-triton = self.openai-triton-cuda; }; openai-whisper = callPackage ../development/python-modules/openai-whisper { }; openant = callPackage ../development/python-modules/openant { }; Loading Loading
pkgs/development/python-modules/openai-whisper/default.nix +19 −28 Original line number Diff line number Diff line Loading @@ -2,22 +2,23 @@ , fetchFromGitHub , buildPythonPackage , substituteAll , cudaSupport ? false # build-system , setuptools # runtime , ffmpeg-headless # propagates , numpy , torch , torchWithCuda , tqdm , more-itertools , transformers , numba , numpy , openai-triton , scipy , tiktoken , torch , tqdm , transformers # tests , pytestCheckHook Loading @@ -25,14 +26,14 @@ buildPythonPackage rec { pname = "whisper"; version = "20230918"; format = "setuptools"; version = "20231117"; pyproject = true; src = fetchFromGitHub { owner = "openai"; repo = pname; rev = "refs/tags/v${version}"; hash = "sha256-wBAanFVEIIzTcoX40P9eI26UdEu0SC/xuife/zi2Xho="; hash = "sha256-MJ1XjB/GuYUiECCuuHS0NWHvvs+ko0oTvLuDI7zLNiY="; }; patches = [ Loading @@ -42,32 +43,22 @@ buildPythonPackage rec { }) ]; nativeBuildInputs = [ setuptools ]; propagatedBuildInputs = [ numpy tqdm more-itertools transformers numba numpy openai-triton scipy tiktoken ] ++ lib.optionals (!cudaSupport) [ torch ] ++ lib.optionals (cudaSupport) [ openai-triton torchWithCuda tqdm transformers ]; postPatch = '' substituteInPlace requirements.txt \ --replace "tiktoken==0.3.3" "tiktoken>=0.3.3" '' # openai-triton is only needed for CUDA support. # triton needs CUDA to be build. # -> by making it optional, we can build whisper without unfree packages enabled + lib.optionalString (!cudaSupport) '' sed -i '/if sys.platform.startswith("linux") and platform.machine() == "x86_64":/{N;d}' setup.py ''; preCheck = '' export HOME=$TMPDIR ''; Loading @@ -85,7 +76,7 @@ buildPythonPackage rec { ]; meta = with lib; { changelog = "https://github.com/openai/whisper/blob/v$[version}/CHANGELOG.md"; changelog = "https://github.com/openai/whisper/blob/v${version}/CHANGELOG.md"; description = "General-purpose speech recognition model"; homepage = "https://github.com/openai/whisper"; license = licenses.mit; Loading
pkgs/top-level/python-packages.nix +1 −4 Original line number Diff line number Diff line Loading @@ -8470,10 +8470,7 @@ self: super: with self; { openai-triton-bin = callPackage ../development/python-modules/openai-triton/bin.nix { }; openai-whisper = callPackage ../development/python-modules/openai-whisper { inherit (pkgs.config) cudaSupport; openai-triton = self.openai-triton-cuda; }; openai-whisper = callPackage ../development/python-modules/openai-whisper { }; openant = callPackage ../development/python-modules/openant { }; Loading