You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+10-6Lines changed: 10 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -94,10 +94,11 @@ services:
94
94
- PUID=1000
95
95
- PGID=1000
96
96
- TZ=Etc/UTC
97
-
- WHISPER_MODEL=tiny-int8
97
+
- DEBUG= #optional
98
98
- LOCAL_ONLY= #optional
99
99
- WHISPER_BEAM=1 #optional
100
-
- WHISPER_LANG=en #optional
100
+
- WHISPER_LANG=auto #optional
101
+
- WHISPER_MODEL=auto #optional
101
102
volumes:
102
103
- /path/to/faster-whisper/data:/config
103
104
ports:
@@ -113,10 +114,11 @@ docker run -d \
113
114
-e PUID=1000 \
114
115
-e PGID=1000 \
115
116
-e TZ=Etc/UTC \
116
-
-e WHISPER_MODEL=tiny-int8 \
117
+
-e DEBUG= `#optional` \
117
118
-e LOCAL_ONLY= `#optional` \
118
119
-e WHISPER_BEAM=1 `#optional` \
119
-
-e WHISPER_LANG=en `#optional` \
120
+
-e WHISPER_LANG=auto `#optional` \
121
+
-e WHISPER_MODEL=auto `#optional` \
120
122
-p 10300:10300 \
121
123
-v /path/to/faster-whisper/data:/config \
122
124
--restart unless-stopped \
@@ -133,10 +135,11 @@ Containers are configured using parameters passed at runtime (such as those abov
133
135
|`-e PUID=1000`| for UserID - see below for explanation |
134
136
|`-e PGID=1000`| for GroupID - see below for explanation |
135
137
|`-e TZ=Etc/UTC`| specify a timezone to use, see this [list](https://en.wikipedia.org/wiki/List_of_tz_database_time_zones#List). |
136
-
|`-e WHISPER_MODEL=tiny-int8`|Whisper model that will be used for transcription. From `tiny`, `base`, `small` and `medium`, all with `-int8` compressed variants|
138
+
|`-e DEBUG=`|If set to `true`, or any other value, the container will output debug logs.|
137
139
|`-e LOCAL_ONLY=`| If set to `true`, or any other value, the container will not attempt to download models from HuggingFace and will only use locally-provided models. |
138
140
|`-e WHISPER_BEAM=1`| Number of candidates to consider simultaneously during transcription. |
139
-
|`-e WHISPER_LANG=en`| Language that you will speak to the add-on. |
141
+
|`-e WHISPER_LANG=auto`| Two character code for the language that you will speak to the add-on. |
142
+
|`-e WHISPER_MODEL=auto`| Whisper model that will be used for transcription. From [here](https://github.com/SYSTRAN/faster-whisper/blob/master/faster_whisper/utils.py#L11-L31). |
140
143
|`-v /config`| Local path for Whisper config files. |
141
144
|`--read-only=true`| Run container with a read-only filesystem. Please [read the docs](https://docs.linuxserver.io/misc/read-only/). |
142
145
@@ -302,6 +305,7 @@ Once registered you can define the dockerfile to use with `-f Dockerfile.aarch64
302
305
303
306
## Versions
304
307
308
+
***26.01.26:** - Default to `auto`for model and language if not set.
305
309
***20.08.25:** - Add gpu-legacy branch for pre-Turing cards.
306
310
***10.08.25:** - Add support for local-only mode.
307
311
***05.12.24:** - Build from Github releases rather than Pypi.
Copy file name to clipboardExpand all lines: readme-vars.yml
+5-4Lines changed: 5 additions & 4 deletions
Original file line number
Diff line number
Diff line change
@@ -18,9 +18,7 @@ development_versions_items:
18
18
# container parameters
19
19
common_param_env_vars_enabled: true
20
20
param_container_name: "{{ project_name }}"
21
-
param_usage_include_env: true
22
-
param_env_vars:
23
-
- {env_var: "WHISPER_MODEL", env_value: "tiny-int8", desc: "Whisper model that will be used for transcription. From `tiny`, `base`, `small` and `medium`, all with `-int8` compressed variants", env_options: ["tiny-int8", "tiny", "base-int8", "base", "small-int8", "small", "medium-int8"]}
- {env_var: "DEBUG", env_value: "", desc: "If set to `true`, or any other value, the container will output debug logs."}
33
32
- {env_var: "LOCAL_ONLY", env_value: "", desc: "If set to `true`, or any other value, the container will not attempt to download models from HuggingFace and will only use locally-provided models."}
34
33
- {env_var: "WHISPER_BEAM", env_value: "1", desc: "Number of candidates to consider simultaneously during transcription."}
35
-
- {env_var: "WHISPER_LANG", env_value: "en", desc: "Language that you will speak to the add-on."}
34
+
- {env_var: "WHISPER_LANG", env_value: "auto", desc: "Two character code for the language that you will speak to the add-on."}
35
+
- {env_var: "WHISPER_MODEL", env_value: "auto", desc: "Whisper model that will be used for transcription. From [here](https://github.com/SYSTRAN/faster-whisper/blob/master/faster_whisper/utils.py#L11-L31)."}
36
36
readonly_supported: true
37
37
# application setup block
38
38
app_setup_block_enabled: true
@@ -85,6 +85,7 @@ init_diagram: |
85
85
"faster-whisper:gpu-legacy" <- Base Images
86
86
# changelog
87
87
changelogs:
88
+
- {date: "26.01.26:", desc: "Default to `auto` for model and language if not set."}
88
89
- {date: "20.08.25:", desc: "Add gpu-legacy branch for pre-Turing cards."}
89
90
- {date: "10.08.25:", desc: "Add support for local-only mode."}
90
91
- {date: "05.12.24:", desc: "Build from Github releases rather than Pypi."}
0 commit comments