forked from ahmetoner/whisper-asr-webservice
-
Notifications
You must be signed in to change notification settings - Fork 0
/
docker-compose.oldgpu.yml
117 lines (111 loc) · 3.46 KB
/
docker-compose.oldgpu.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
version: "3.4"
services:
traefik:
image: "traefik:v2.10"
container_name: "traefik"
command:
- "--log.level=DEBUG"
- "--api.insecure=true"
- "--providers.docker=true"
- "--providers.docker.exposedbydefault=false"
- "--entrypoints.web.address=:8070"
- "--accesslog"
- "--api"
ports:
- "8070:8070"
#- "8080:8080"
volumes:
- "/var/run/docker.sock:/var/run/docker.sock:ro"
whoami:
# A container that exposes an API to show its IP address
image: traefik/whoami
labels:
- traefik.http.routers.whoami.rule=Host(`whoami.localhost`)
- traefik.http.routers.whoami.entrypoints=web
- traefik.enable=true
whisper-asr-webservice-1:
image: allomediadocker/openai-whisper-asr-webservice:cuda-10.2
environment:
# https://github.com/openai/whisper#available-models-and-languages
- ASR_MODEL=large
- ASR_ENGINE=openai_whisper # faster_whisper can't work with cuda10 (only cuda 11)
env_file:
- .env
deploy:
mode: replicated
replicas: 1
resources:
reservations:
devices:
- driver: nvidia
capabilities: [gpu]
device_ids: ['0']
labels:
- traefik.enable=true
- traefik.http.routers.whisper-asr-webservice.rule=Host(`${TRAEFIK_HOSTNAME}`)
- traefik.http.routers.whisper-asr-webservice.entrypoints=web
- traefik.http.services.whisper-asr-webservice.loadbalancer.server.port=9000
volumes:
- ./app:/app/app
- cache-pip:/root/.cache/pip
- cache-poetry:/root/.cache/poetry
- cache-whisper:/root/.cache/whisper
whisper-asr-webservice-2:
image: allomediadocker/openai-whisper-asr-webservice:cuda-10.2
environment:
# https://github.com/openai/whisper#available-models-and-languages
- ASR_MODEL=large
- ASR_ENGINE=openai_whisper
env_file:
- .env
deploy:
mode: replicated
replicas: 1
resources:
reservations:
devices:
- driver: nvidia
capabilities: [gpu]
device_ids: ['1']
labels:
- traefik.enable=true
- traefik.http.routers.whisper-asr-webservice.rule=Host(`${TRAEFIK_HOSTNAME}`)
- traefik.http.routers.whisper-asr-webservice.entrypoints=web
- traefik.http.services.whisper-asr-webservice.loadbalancer.server.port=9000
volumes:
- ./app:/app/app
- cache-pip:/root/.cache/pip
- cache-poetry:/root/.cache/poetry
- cache-whisper:/root/.cache/whisper
whisper-asr-webservice-3:
image: allomediadocker/openai-whisper-asr-webservice:cuda-10.2
environment:
# https://github.com/openai/whisper#available-models-and-languages
- ASR_MODEL=large
- ASR_ENGINE=openai_whisper
env_file:
- .env
deploy:
mode: replicated
replicas: 1
resources:
reservations:
devices:
- driver: nvidia
capabilities: [gpu]
device_ids: ['2']
labels:
- traefik.enable=true
- traefik.http.routers.whisper-asr-webservice.rule=Host(`${TRAEFIK_HOSTNAME}`)
- traefik.http.routers.whisper-asr-webservice.entrypoints=web
- traefik.http.services.whisper-asr-webservice.loadbalancer.server.port=9000
volumes:
- ./app:/app/app
- cache-pip:/root/.cache/pip
- cache-poetry:/root/.cache/poetry
- cache-whisper:/root/.cache/whisper
volumes:
cache-pip:
cache-poetry:
cache-whisper:
cache-faster-whisper: