

chore: update docker tag to latest
@f89c7471e8d4ffc6639d1c128964da0bba41ce04
--- README.md
+++ README.md
... | ... | @@ -20,9 +20,9 @@ |
20 | 20 |
## Quick Start |
21 | 21 |
Using Docker |
22 | 22 |
```bash |
23 |
-docker run --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface fedirz/faster-whisper-server:0.1-cuda |
|
23 |
+docker run --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface fedirz/faster-whisper-server:latest-cuda |
|
24 | 24 |
# or |
25 |
-docker run --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface fedirz/faster-whisper-server:0.1-cpu |
|
25 |
+docker run --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface fedirz/faster-whisper-server:latest-cpu |
|
26 | 26 |
``` |
27 | 27 |
Using Docker Compose |
28 | 28 |
```bash |
--- compose.yaml
+++ compose.yaml
... | ... | @@ -1,7 +1,7 @@ |
1 | 1 |
# NOTE: arm images haven't been tested |
2 | 2 |
services: |
3 | 3 |
faster-whisper-server-cuda: |
4 |
- image: fedirz/faster-whisper-server:0.1-cuda |
|
4 |
+ image: fedirz/faster-whisper-server:latest-cuda |
|
5 | 5 |
build: |
6 | 6 |
dockerfile: Dockerfile.cuda |
7 | 7 |
context: . |
... | ... | @@ -23,7 +23,7 @@ |
23 | 23 |
devices: |
24 | 24 |
- capabilities: ["gpu"] |
25 | 25 |
faster-whisper-server-cpu: |
26 |
- image: fedirz/faster-whisper-server:0.1-cpu |
|
26 |
+ image: fedirz/faster-whisper-server:latest-cpu |
|
27 | 27 |
build: |
28 | 28 |
dockerfile: Dockerfile.cpu |
29 | 29 |
context: . |
--- examples/live-audio/script.sh
+++ examples/live-audio/script.sh
... | ... | @@ -10,9 +10,9 @@ |
10 | 10 |
export WHISPER_MODEL=Systran/faster-distil-whisper-large-v3 # or Systran/faster-whisper-tiny.en if you are running on a CPU for a faster inference. |
11 | 11 |
|
12 | 12 |
# Ensure you have `faster-whisper-server` running. If this is your first time running it expect to wait up-to a minute for the model to be downloaded and loaded into memory. You can run `curl localhost:8000/health` to check if the server is ready or watch the logs with `docker logs -f <container_id>`. |
13 |
-docker run --detach --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:0.1-cuda |
|
13 |
+docker run --detach --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:latest-cuda |
|
14 | 14 |
# or you can run it on a CPU |
15 |
-# docker run --detach --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:0.1-cpu |
|
15 |
+# docker run --detach --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:latest-cpu |
|
16 | 16 |
|
17 | 17 |
# `pv` is used to limit the rate at which the audio is streamed to the server. Audio is being streamed at a rate of 32kb/s(16000 sample rate * 16-bit sample / 8 bits per byte = 32000 bytes per second). This emulutes live audio input from a microphone: `ffmpeg -loglevel quiet -f alsa -i default -ac 1 -ar 16000 -f s16le` |
18 | 18 |
# shellcheck disable=SC2002 |
--- examples/youtube/script.sh
+++ examples/youtube/script.sh
... | ... | @@ -6,9 +6,9 @@ |
6 | 6 |
export WHISPER_MODEL=Systran/faster-distil-whisper-large-v3 # or Systran/faster-whisper-tiny.en if you are running on a CPU for a faster inference. |
7 | 7 |
|
8 | 8 |
# Ensure you have `faster-whisper-server` running. If this is your first time running it expect to wait up-to a minute for the model to be downloaded and loaded into memory. You can run `curl localhost:8000/health` to check if the server is ready or watch the logs with `docker logs -f <container_id>`. |
9 |
-docker run --detach --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:0.1-cuda |
|
9 |
+docker run --detach --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:latest-cuda |
|
10 | 10 |
# or you can run it on a CPU |
11 |
-# docker run --detach --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:0.1-cpu |
|
11 |
+# docker run --detach --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:latest-cpu |
|
12 | 12 |
|
13 | 13 |
# Download the audio from a YouTube video. In this example I'm downloading "The Evolution of the Operating System" by Asionometry YouTube channel. I highly checking this channel out, the guy produces very high content. If you don't have `youtube-dl`, you'll have to install it. https://github.com/ytdl-org/youtube-dl |
14 | 14 |
youtube-dl --extract-audio --audio-format mp3 -o the-evolution-of-the-operating-system.mp3 'https://www.youtube.com/watch?v=1lG7lFLXBIs' |
Add a comment
Delete comment
Once you delete this comment, you won't be able to recover it. Are you sure you want to delete this comment?