Unverified Commit e7e5233f authored by Martin Weinelt's avatar Martin Weinelt Committed by GitHub
Browse files

Merge pull request #268177 from mweinelt/openai-whisper-20231117

openai-whisper: 20230918 -> 20231117
parents 8314afe1 0aeec365
Loading
Loading
Loading
Loading
+19 −28
Original line number Diff line number Diff line
@@ -2,22 +2,23 @@
, fetchFromGitHub
, buildPythonPackage
, substituteAll
, cudaSupport ? false

# build-system
, setuptools

# runtime
, ffmpeg-headless

# propagates
, numpy
, torch
, torchWithCuda
, tqdm
, more-itertools
, transformers
, numba
, numpy
, openai-triton
, scipy
, tiktoken
, torch
, tqdm
, transformers

# tests
, pytestCheckHook
@@ -25,14 +26,14 @@

buildPythonPackage rec {
  pname = "whisper";
  version = "20230918";
  format = "setuptools";
  version = "20231117";
  pyproject = true;

  src = fetchFromGitHub {
    owner = "openai";
    repo = pname;
    rev = "refs/tags/v${version}";
    hash = "sha256-wBAanFVEIIzTcoX40P9eI26UdEu0SC/xuife/zi2Xho=";
    hash = "sha256-MJ1XjB/GuYUiECCuuHS0NWHvvs+ko0oTvLuDI7zLNiY=";
  };

  patches = [
@@ -42,32 +43,22 @@ buildPythonPackage rec {
    })
  ];

  nativeBuildInputs = [
    setuptools
  ];

  propagatedBuildInputs = [
    numpy
    tqdm
    more-itertools
    transformers
    numba
    numpy
    openai-triton
    scipy
    tiktoken
  ] ++ lib.optionals (!cudaSupport) [
    torch
  ] ++ lib.optionals (cudaSupport) [
    openai-triton
    torchWithCuda
    tqdm
    transformers
  ];

  postPatch = ''
    substituteInPlace requirements.txt \
      --replace "tiktoken==0.3.3" "tiktoken>=0.3.3"
  ''
  # openai-triton is only needed for CUDA support.
  # triton needs CUDA to be build.
  # -> by making it optional, we can build whisper without unfree packages enabled
  + lib.optionalString (!cudaSupport) ''
    sed -i '/if sys.platform.startswith("linux") and platform.machine() == "x86_64":/{N;d}' setup.py
  '';

  preCheck = ''
    export HOME=$TMPDIR
  '';
@@ -85,7 +76,7 @@ buildPythonPackage rec {
  ];

  meta = with lib; {
    changelog = "https://github.com/openai/whisper/blob/v$[version}/CHANGELOG.md";
    changelog = "https://github.com/openai/whisper/blob/v${version}/CHANGELOG.md";
    description = "General-purpose speech recognition model";
    homepage = "https://github.com/openai/whisper";
    license = licenses.mit;
+1 −4
Original line number Diff line number Diff line
@@ -8470,10 +8470,7 @@ self: super: with self; {
  openai-triton-bin = callPackage ../development/python-modules/openai-triton/bin.nix { };
  openai-whisper = callPackage ../development/python-modules/openai-whisper {
    inherit (pkgs.config) cudaSupport;
    openai-triton = self.openai-triton-cuda;
  };
  openai-whisper = callPackage ../development/python-modules/openai-whisper { };
  openant = callPackage ../development/python-modules/openant { };