You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+4Lines changed: 4 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -95,6 +95,7 @@ services:
95
95
- PGID=1000
96
96
- TZ=Etc/UTC
97
97
- WHISPER_MODEL=tiny-int8
98
+
- LOCAL_ONLY= #optional
98
99
- WHISPER_BEAM=1 #optional
99
100
- WHISPER_LANG=en #optional
100
101
volumes:
@@ -113,6 +114,7 @@ docker run -d \
113
114
-e PGID=1000 \
114
115
-e TZ=Etc/UTC \
115
116
-e WHISPER_MODEL=tiny-int8 \
117
+
-e LOCAL_ONLY= `#optional` \
116
118
-e WHISPER_BEAM=1 `#optional` \
117
119
-e WHISPER_LANG=en `#optional` \
118
120
-p 10300:10300 \
@@ -132,6 +134,7 @@ Containers are configured using parameters passed at runtime (such as those abov
132
134
|`-e PGID=1000`| for GroupID - see below for explanation |
133
135
|`-e TZ=Etc/UTC`| specify a timezone to use, see this [list](https://en.wikipedia.org/wiki/List_of_tz_database_time_zones#List). |
134
136
|`-e WHISPER_MODEL=tiny-int8`| Whisper model that will be used for transcription. From [here](https://github.com/SYSTRAN/faster-whisper/blob/master/faster_whisper/utils.py#L12-L31), all with `-int8` compressed variants |
137
+
|`-e LOCAL_ONLY=`| If set to `true`, or any other value, the container will not attempt to download models from HuggingFace and will only use locally-provided models. |
135
138
|`-e WHISPER_BEAM=1`| Number of candidates to consider simultaneously during transcription. |
136
139
|`-e WHISPER_LANG=en`| Language that you will speak to the add-on. |
137
140
|`-v /config`| Local path for Whisper config files. |
@@ -299,6 +302,7 @@ Once registered you can define the dockerfile to use with `-f Dockerfile.aarch64
299
302
300
303
## Versions
301
304
305
+
***10.08.25:** - Add support for local-only mode.
302
306
***30.12.24:** - Add arm64 support for non-GPU builds.
303
307
***05.12.24:** - Build from Github releases rather than Pypi.
Copy file name to clipboardExpand all lines: readme-vars.yml
+4Lines changed: 4 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -16,6 +16,7 @@ development_versions: true
16
16
development_versions_items:
17
17
- {tag: "latest", desc: "Stable releases"}
18
18
- {tag: "gpu", desc: "Releases with Nvidia GPU support (amd64 only)"}
19
+
- {tag: "gpu-legacy", desc: "Legacy releases with Nvidia GPU support for pre-Turing cards (amd64 only)"}
19
20
# container parameters
20
21
common_param_env_vars_enabled: true
21
22
param_container_name: "{{ project_name }}"
@@ -31,6 +32,7 @@ param_ports:
31
32
# optional container parameters
32
33
opt_param_usage_include_env: true
33
34
opt_param_env_vars:
35
+
- {env_var: "LOCAL_ONLY", env_value: "", desc: "If set to `true`, or any other value, the container will not attempt to download models from HuggingFace and will only use locally-provided models."}
34
36
- {env_var: "WHISPER_BEAM", env_value: "1", desc: "Number of candidates to consider simultaneously during transcription."}
35
37
- {env_var: "WHISPER_LANG", env_value: "en", desc: "Language that you will speak to the add-on."}
36
38
readonly_supported: true
@@ -85,6 +87,8 @@ init_diagram: |
85
87
"faster-whisper:latest" <- Base Images
86
88
# changelog
87
89
changelogs:
90
+
- {date: "20.08.25:", desc: "Add gpu-legacy branch for pre-Turing cards."}
91
+
- {date: "10.08.25:", desc: "Add support for local-only mode."}
88
92
- {date: "30.12.24:", desc: "Add arm64 support for non-GPU builds."}
89
93
- {date: "05.12.24:", desc: "Build from Github releases rather than Pypi."}
90
94
- {date: "18.07.24:", desc: "Rebase to Ubuntu Noble."}
0 commit comments