Skip to content

Commit 0306c62

Browse files
Update TGI CPU image to latest official release 2.4.0 (#1035)
Signed-off-by: lvliang-intel <liang1.lv@intel.com> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
1 parent 3372b9d commit 0306c62

File tree

40 files changed

+49
-49
lines changed

40 files changed

+49
-49
lines changed

AudioQnA/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ services:
4141
environment:
4242
TTS_ENDPOINT: ${TTS_ENDPOINT}
4343
tgi-service:
44-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
44+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
4545
container_name: tgi-service
4646
ports:
4747
- "3006:80"

AudioQnA/docker_compose/intel/cpu/xeon/compose_multilang.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ services:
2626
https_proxy: ${https_proxy}
2727
restart: unless-stopped
2828
tgi-service:
29-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
29+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
3030
container_name: tgi-service
3131
ports:
3232
- "3006:80"

AudioQnA/kubernetes/intel/cpu/xeon/manifest/audioqna.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -247,7 +247,7 @@ spec:
247247
- envFrom:
248248
- configMapRef:
249249
name: audio-qna-config
250-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
250+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
251251
name: llm-dependency-deploy-demo
252252
securityContext:
253253
capabilities:

AvatarChatbot/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ services:
4242
environment:
4343
TTS_ENDPOINT: ${TTS_ENDPOINT}
4444
tgi-service:
45-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
45+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
4646
container_name: tgi-service
4747
ports:
4848
- "3006:80"

ChatQnA/docker_compose/intel/cpu/xeon/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -195,7 +195,7 @@ For users in China who are unable to download models directly from Huggingface,
195195
export HF_TOKEN=${your_hf_token}
196196
export HF_ENDPOINT="https://hf-mirror.com"
197197
model_name="Intel/neural-chat-7b-v3-3"
198-
docker run -p 8008:80 -v ./data:/data --name tgi-service -e HF_ENDPOINT=$HF_ENDPOINT -e http_proxy=$http_proxy -e https_proxy=$https_proxy --shm-size 1g ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu --model-id $model_name
198+
docker run -p 8008:80 -v ./data:/data --name tgi-service -e HF_ENDPOINT=$HF_ENDPOINT -e http_proxy=$http_proxy -e https_proxy=$https_proxy --shm-size 1g ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu --model-id $model_name
199199
```
200200

201201
2. Offline
@@ -209,7 +209,7 @@ For users in China who are unable to download models directly from Huggingface,
209209
```bash
210210
export HF_TOKEN=${your_hf_token}
211211
export model_path="/path/to/model"
212-
docker run -p 8008:80 -v $model_path:/data --name tgi_service --shm-size 1g ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu --model-id /data
212+
docker run -p 8008:80 -v $model_path:/data --name tgi_service --shm-size 1g ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu --model-id /data
213213
```
214214

215215
### Setup Environment Variables

ChatQnA/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ services:
7373
HF_HUB_ENABLE_HF_TRANSFER: 0
7474
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
7575
tgi-service:
76-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
76+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
7777
container_name: tgi-service
7878
ports:
7979
- "9009:80"

ChatQnA/docker_compose/intel/cpu/xeon/compose_qdrant.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ services:
7272
HF_HUB_ENABLE_HF_TRANSFER: 0
7373
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
7474
tgi-service:
75-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
75+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
7676
container_name: tgi-service
7777
ports:
7878
- "6042:80"

ChatQnA/docker_compose/intel/cpu/xeon/compose_without_rerank.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -57,7 +57,7 @@ services:
5757
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
5858
restart: unless-stopped
5959
tgi-service:
60-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
60+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
6161
container_name: tgi-service
6262
ports:
6363
- "9009:80"

ChatQnA/kubernetes/intel/README_gmc.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ The ChatQnA uses the below prebuilt images if you choose a Xeon deployment
1818
- tei_embedding_service: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
1919
- retriever: opea/retriever-redis:latest
2020
- tei_xeon_service: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
21-
- tgi-service: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
21+
- tgi-service: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
2222
- chaqna-xeon-backend-server: opea/chatqna:latest
2323

2424
Should you desire to use the Gaudi accelerator, two alternate images are used for the embedding and llm services.

ChatQnA/kubernetes/intel/cpu/xeon/manifest/chatqna-guardrails.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1100,7 +1100,7 @@ spec:
11001100
runAsUser: 1000
11011101
seccompProfile:
11021102
type: RuntimeDefault
1103-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
1103+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
11041104
imagePullPolicy: IfNotPresent
11051105
volumeMounts:
11061106
- mountPath: /data
@@ -1180,7 +1180,7 @@ spec:
11801180
runAsUser: 1000
11811181
seccompProfile:
11821182
type: RuntimeDefault
1183-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
1183+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
11841184
imagePullPolicy: IfNotPresent
11851185
volumeMounts:
11861186
- mountPath: /data

ChatQnA/kubernetes/intel/cpu/xeon/manifest/chatqna.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -922,7 +922,7 @@ spec:
922922
runAsUser: 1000
923923
seccompProfile:
924924
type: RuntimeDefault
925-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
925+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
926926
imagePullPolicy: IfNotPresent
927927
volumeMounts:
928928
- mountPath: /data

ChatQnA/kubernetes/intel/cpu/xeon/manifest/chatqna_bf16.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -925,7 +925,7 @@ spec:
925925
runAsUser: 1000
926926
seccompProfile:
927927
type: RuntimeDefault
928-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
928+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
929929
imagePullPolicy: IfNotPresent
930930
volumeMounts:
931931
- mountPath: /data

ChatQnA/tests/test_compose_on_xeon.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ function build_docker_images() {
2222
service_list="chatqna chatqna-ui chatqna-conversation-ui dataprep-redis retriever-redis nginx"
2323
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
2424

25-
docker pull ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
25+
docker pull ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
2626
docker pull ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
2727

2828
docker images && sleep 1s

CodeGen/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
tgi-service:
6-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
6+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
77
container_name: tgi-service
88
ports:
99
- "8028:80"

CodeGen/kubernetes/intel/cpu/xeon/manifest/codegen.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -404,7 +404,7 @@ spec:
404404
runAsUser: 1000
405405
seccompProfile:
406406
type: RuntimeDefault
407-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
407+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
408408
imagePullPolicy: IfNotPresent
409409
volumeMounts:
410410
- mountPath: /data

CodeGen/kubernetes/intel/cpu/xeon/manifest/codegen_react_ui.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,7 @@ spec:
126126
- name: no_proxy
127127
value:
128128
securityContext: {}
129-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
129+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
130130
imagePullPolicy: IfNotPresent
131131
volumeMounts:
132132
- mountPath: /data

CodeGen/tests/test_compose_on_xeon.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ function build_docker_images() {
2222
service_list="codegen codegen-ui llm-tgi"
2323
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
2424

25-
docker pull ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
25+
docker pull ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
2626
docker images && sleep 1s
2727
}
2828

CodeTrans/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
tgi-service:
6-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
6+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
77
container_name: codetrans-tgi-service
88
ports:
99
- "8008:80"

CodeTrans/kubernetes/intel/cpu/xeon/manifest/codetrans.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -404,7 +404,7 @@ spec:
404404
runAsUser: 1000
405405
seccompProfile:
406406
type: RuntimeDefault
407-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
407+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
408408
imagePullPolicy: IfNotPresent
409409
volumeMounts:
410410
- mountPath: /data

CodeTrans/tests/test_compose_on_xeon.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ function build_docker_images() {
2222
service_list="codetrans codetrans-ui llm-tgi nginx"
2323
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
2424

25-
docker pull ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
25+
docker pull ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
2626
docker images && sleep 1s
2727
}
2828

DocSum/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
tgi-service:
6-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
6+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
77
container_name: tgi-service
88
ports:
99
- "8008:80"

DocSum/kubernetes/intel/README_gmc.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ Install GMC in your Kubernetes cluster, if you have not already done so, by foll
88
The DocSum application is defined as a Custom Resource (CR) file that the above GMC operator acts upon. It first checks if the microservices listed in the CR yaml file are running, if not it starts them and then proceeds to connect them. When the DocSum RAG pipeline is ready, the service endpoint details are returned, letting you use the application. Should you use "kubectl get pods" commands you will see all the component microservices, in particular embedding, retriever, rerank, and llm.
99

1010
The DocSum pipeline uses prebuilt images. The Xeon version uses the prebuilt image `llm-docsum-tgi:latest` which internally leverages the
11-
the image `ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu`. The service is called tgi-svc. Meanwhile, the Gaudi version launches the
11+
the image `ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu`. The service is called tgi-svc. Meanwhile, the Gaudi version launches the
1212
service tgi-gaudi-svc, which uses the image `ghcr.io/huggingface/tgi-gaudi:2.0.5`. Both TGI model services serve the model specified in the LLM_MODEL_ID variable that is exported by you. In the below example we use `Intel/neural-chat-7b-v3-3`.
1313

1414
[NOTE]

DocSum/kubernetes/intel/cpu/xeon/manifest/docsum.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -404,7 +404,7 @@ spec:
404404
runAsUser: 1000
405405
seccompProfile:
406406
type: RuntimeDefault
407-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
407+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
408408
imagePullPolicy: IfNotPresent
409409
volumeMounts:
410410
- mountPath: /data

DocSum/kubernetes/intel/cpu/xeon/manifest/ui/react-docsum.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,7 @@ spec:
126126
- name: no_proxy
127127
value:
128128
securityContext: {}
129-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
129+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
130130
imagePullPolicy: IfNotPresent
131131
volumeMounts:
132132
- mountPath: /data

FaqGen/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
tgi-service:
6-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
6+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
77
container_name: tgi-xeon-server
88
ports:
99
- "8008:80"

FaqGen/kubernetes/intel/cpu/xeon/manifest/faqgen_react_ui.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,7 @@ spec:
126126
- name: no_proxy
127127
value:
128128
securityContext: {}
129-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
129+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
130130
imagePullPolicy: IfNotPresent
131131
volumeMounts:
132132
- mountPath: /data

ProductivitySuite/kubernetes/intel/cpu/xeon/manifest/chatqna.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -993,7 +993,7 @@ spec:
993993
name: chatqna-tgi-config
994994
securityContext:
995995
{}
996-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
996+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
997997
imagePullPolicy: IfNotPresent
998998
volumeMounts:
999999
- mountPath: /data

ProductivitySuite/kubernetes/intel/cpu/xeon/manifest/codegen.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -229,7 +229,7 @@ spec:
229229
name: codegen-tgi-config
230230
securityContext:
231231
{}
232-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
232+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
233233
imagePullPolicy: IfNotPresent
234234
volumeMounts:
235235
- mountPath: /data

ProductivitySuite/kubernetes/intel/cpu/xeon/manifest/docsum.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -229,7 +229,7 @@ spec:
229229
name: docsum-tgi-config
230230
securityContext:
231231
{}
232-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
232+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
233233
imagePullPolicy: IfNotPresent
234234
volumeMounts:
235235
- mountPath: /data

ProductivitySuite/kubernetes/intel/cpu/xeon/manifest/faqgen.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -138,7 +138,7 @@ spec:
138138
- configMapRef:
139139
name: faqgen-tgi-config
140140
securityContext: {}
141-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
141+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
142142
imagePullPolicy: IfNotPresent
143143
volumeMounts:
144144
- mountPath: /data

SearchQnA/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ services:
7373
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
7474
restart: unless-stopped
7575
tgi-service:
76-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
76+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
7777
container_name: tgi-service
7878
ports:
7979
- "3006:80"

SearchQnA/tests/test_compose_on_xeon.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ function build_docker_images() {
2323
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
2424

2525
docker pull ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
26-
docker pull ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
26+
docker pull ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
2727
docker images && sleep 1s
2828
}
2929

Translation/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
tgi-service:
6-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
6+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
77
container_name: tgi-service
88
ports:
99
- "8008:80"

Translation/kubernetes/intel/cpu/xeon/manifest/translation.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -361,7 +361,7 @@ spec:
361361
runAsUser: 1000
362362
seccompProfile:
363363
type: RuntimeDefault
364-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
364+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
365365
imagePullPolicy: IfNotPresent
366366
volumeMounts:
367367
- mountPath: /data

Translation/tests/test_compose_on_xeon.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ function build_docker_images() {
2222
service_list="translation translation-ui llm-tgi nginx"
2323
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
2424

25-
docker pull ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
25+
docker pull ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
2626
docker images && sleep 1s
2727
}
2828

VisualQnA/docker_compose/intel/cpu/xeon/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -67,12 +67,12 @@ docker build --no-cache -t opea/visualqna-ui:latest --build-arg https_proxy=$htt
6767
### 4. Pull TGI Xeon Image
6868

6969
```bash
70-
docker pull ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
70+
docker pull ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
7171
```
7272

7373
Then run the command `docker images`, you will have the following 5 Docker Images:
7474

75-
1. `ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu`
75+
1. `ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu`
7676
2. `opea/lvm-tgi:latest`
7777
3. `opea/visualqna:latest`
7878
4. `opea/visualqna-ui:latest`

VisualQnA/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
llava-tgi-service:
6-
image: ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
6+
image: ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
77
container_name: tgi-llava-xeon-server
88
ports:
99
- "8399:80"

VisualQnA/kubernetes/intel/cpu/xeon/manifest/visualqna.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -216,7 +216,7 @@ spec:
216216
name: visualqna-tgi-config
217217
securityContext:
218218
{}
219-
image: "ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu"
219+
image: "ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu"
220220
imagePullPolicy: IfNotPresent
221221
volumeMounts:
222222
- mountPath: /data

VisualQnA/tests/test_compose_on_xeon.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ function build_docker_images() {
2121
echo "Build all the images with --no-cache, check docker_image_build.log for details..."
2222
docker compose -f build.yaml build --no-cache > ${LOG_PATH}/docker_image_build.log
2323

24-
docker pull ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu
24+
docker pull ghcr.io/huggingface/text-generation-inference:2.4.0-intel-cpu
2525
docker images && sleep 1s
2626
}
2727

0 commit comments

Comments
 (0)