Fedir Zadniprovskyi 2024-09-08
chore: update default whisper model
@9539fc9bf15f137d9fc273947feac8378a7d0b71
Dockerfile.cpu
--- Dockerfile.cpu
+++ Dockerfile.cpu
@@ -17,9 +17,8 @@
 COPY ./src ./pyproject.toml ./uv.lock ./
 RUN --mount=type=cache,target=/root/.cache/uv \
     uv sync --frozen
-ENV WHISPER__MODEL=Systran/faster-whisper-medium.en
+ENV WHISPER__MODEL=Systran/faster-whisper-small
 ENV WHISPER__INFERENCE_DEVICE=cpu
-ENV WHISPER__COMPUTE_TYPE=int8
 ENV UVICORN_HOST=0.0.0.0
 ENV UVICORN_PORT=8000
 CMD ["uv", "run", "uvicorn", "faster_whisper_server.main:app"]
Dockerfile.cuda
--- Dockerfile.cuda
+++ Dockerfile.cuda
@@ -17,7 +17,7 @@
 COPY ./src ./pyproject.toml ./uv.lock ./
 RUN --mount=type=cache,target=/root/.cache/uv \
     uv sync --frozen
-ENV WHISPER__MODEL=Systran/faster-distil-whisper-large-v3
+ENV WHISPER__MODEL=Systran/faster-whisper-large-v3
 ENV WHISPER__INFERENCE_DEVICE=auto
 ENV UVICORN_HOST=0.0.0.0
 ENV UVICORN_PORT=8000
src/faster_whisper_server/config.py
--- src/faster_whisper_server/config.py
+++ src/faster_whisper_server/config.py
@@ -150,7 +150,7 @@
 class WhisperConfig(BaseModel):
     """See https://github.com/SYSTRAN/faster-whisper/blob/master/faster_whisper/transcribe.py#L599."""
 
-    model: str = Field(default="Systran/faster-whisper-medium.en")
+    model: str = Field(default="Systran/faster-whisper-small")
     """
     Default Huggingface model to use for transcription. Note, the model must support being ran using CTranslate2.
     This model will be used if no model is specified in the request.
@@ -205,7 +205,7 @@
     preload_models: list[str] = Field(
         default_factory=list,
         examples=[
-            ["Systran/faster-whisper-medium.en"],
+            ["Systran/faster-whisper-small"],
             ["Systran/faster-whisper-medium.en", "Systran/faster-whisper-small.en"],
         ],
     )
Add a comment
List