Skip to content

Commit 2375924

Browse files
authored
Merge pull request #375 from makaveli10/update_trt_docs
Update tensorrt_llm docker setup
2 parents 4ba576f + ae16924 commit 2375924

File tree

3 files changed

+9
-33
lines changed

3 files changed

+9
-33
lines changed

.github/workflows/ci.yml

Lines changed: 0 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -99,35 +99,6 @@ jobs:
9999
push: true
100100
tags: ghcr.io/collabora/whisperlive-cpu:latest
101101

102-
build-and-push-docker-tensorrt:
103-
needs: [run-tests, check-code-format]
104-
timeout-minutes: 60
105-
runs-on: ubuntu-22.04
106-
if: github.event_name == 'push' && (github.ref == 'refs/heads/main' || startsWith(github.ref, 'refs/tags/'))
107-
steps:
108-
- uses: actions/checkout@v2
109-
110-
- name: Log in to GitHub Container Registry
111-
uses: docker/login-action@v1
112-
with:
113-
registry: ghcr.io
114-
username: ${{ github.repository_owner }}
115-
password: ${{ secrets.GHCR_TOKEN }}
116-
117-
- name: Docker Prune
118-
run: docker system prune -af
119-
120-
- name: Set up Docker Buildx
121-
uses: docker/setup-buildx-action@v1
122-
123-
- name: Build and push Docker GPU image
124-
uses: docker/build-push-action@v2
125-
with:
126-
context: .
127-
file: docker/Dockerfile.tensorrt
128-
push: true
129-
tags: ghcr.io/collabora/whisperlive-tensorrt:latest
130-
131102
build-and-push-docker-gpu:
132103
needs: [run-tests, check-code-format, build-and-push-docker-cpu]
133104
timeout-minutes: 20

README.md

Lines changed: 7 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -142,12 +142,18 @@ client(hls_url="http://as-hls-ww-live.akamaized.net/pool_904/live/ww/bbc_1xtra/b
142142
## Browser Extensions
143143
- Run the server with your desired backend as shown [here](https://github.com/collabora/WhisperLive?tab=readme-ov-file#running-the-server).
144144
- Transcribe audio directly from your browser using our Chrome or Firefox extensions. Refer to [Audio-Transcription-Chrome](https://github.com/collabora/whisper-live/tree/main/Audio-Transcription-Chrome#readme) and https://github.com/collabora/WhisperLive/blob/main/TensorRT_whisper.md
145+
146+
## Whisper Live Server in Docker
147+
- GPU
148+
- Faster-Whisper
149+
```bash
145150
docker run -it --gpus all -p 9090:9090 ghcr.io/collabora/whisperlive-gpu:latest
146151
```
147152

148153
- TensorRT. Refer to [TensorRT_whisper readme](https://github.com/collabora/WhisperLive/blob/main/TensorRT_whisper.md) for setup and more tensorrt backend configurations.
149154
```bash
150-
docker run -p 9090:9090 --runtime=nvidia --gpus all --entrypoint /bin/bash -it ghcr.io/collabora/whisperlive-tensorrt
155+
docker build . -f docker/Dockerfile.tensorrt -t whisperlive-tensorrt
156+
docker run -p 9090:9090 --runtime=nvidia --gpus all --entrypoint /bin/bash -it whisperlive-tensorrt
151157

152158
# Build small.en engine
153159
bash build_whisper_tensorrt.sh /app/TensorRT-LLM-examples small.en # float16
@@ -173,8 +179,6 @@ client(hls_url="http://as-hls-ww-live.akamaized.net/pool_904/live/ww/bbc_1xtra/b
173179
docker run -it -p 9090:9090 ghcr.io/collabora/whisperlive-cpu:latest
174180
```
175181

176-
**Note**: By default we use "small" model size. To build docker image for a different model size, change the size in server.py and then build the docker image.
177-
178182
## Future Work
179183
- [ ] Add translation to other languages on top of transcription.
180184

TensorRT_whisper.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,8 @@ We have only tested the TensorRT backend in docker so, we recommend docker for a
88

99
- Run WhisperLive TensorRT in docker
1010
```bash
11-
docker run -p 9090:9090 --runtime=nvidia --gpus all --entrypoint /bin/bash -it ghcr.io/collabora/whisperlive-tensorrt:latest
11+
docker build . -f docker/Dockerfile.tensorrt -t whisperlive-tensorrt
12+
docker run -p 9090:9090 --runtime=nvidia --gpus all --entrypoint /bin/bash -it whisperlive-tensorrt
1213
```
1314

1415
## Whisper TensorRT Engine

0 commit comments

Comments
 (0)