76
76
no_proxy : " "
77
77
LOGFLAG : " "
78
78
vLLM_ENDPOINT : " insert-your-remote-vllm-inference-endpoint"
79
- LLM_MODEL : " meta-llama/Meta-Llama-3.1-8B -Instruct"
80
- MODEL_ID : " meta-llama/Meta-Llama-3.1-8B -Instruct"
79
+ LLM_MODEL : " meta-llama/Meta-Llama-3.1-70B -Instruct"
80
+ MODEL_ID : " meta-llama/Meta-Llama-3.1-70B -Instruct"
81
81
CLIENTID : " "
82
82
CLIENT_SECRET : " "
83
83
TOKEN_URL : " "
@@ -174,6 +174,10 @@ data:
174
174
proxy_set_header X-Real-IP $remote_addr;
175
175
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
176
176
proxy_set_header X-Forwarded-Proto $scheme;
177
+ proxy_buffering off;
178
+ proxy_cache off;
179
+ proxy_request_buffering off;
180
+ gzip off;
177
181
}
178
182
179
183
location /v1/dataprep {
@@ -459,6 +463,9 @@ spec:
459
463
{}
460
464
containers :
461
465
- name : chatqna-ui
466
+ env :
467
+ - name : MODEL_ID
468
+ value : " meta-llama/Meta-Llama-3.1-70B-Instruct"
462
469
securityContext :
463
470
{}
464
471
image : " opea/chatqna-ui:latest"
@@ -981,7 +988,7 @@ spec:
981
988
- name : EMBEDDING_SERVICE_HOST_IP
982
989
value : chatqna-embedding-usvc
983
990
- name : MODEL_ID
984
- value : " meta-llama/Meta-Llama-3.1-8B -Instruct"
991
+ value : " meta-llama/Meta-Llama-3.1-70B -Instruct"
985
992
securityContext :
986
993
allowPrivilegeEscalation : false
987
994
capabilities :
@@ -993,7 +1000,7 @@ spec:
993
1000
seccompProfile :
994
1001
type : RuntimeDefault
995
1002
image : " opea/chatqna-wrapper:latest"
996
- imagePullPolicy : IfNotPresent
1003
+ imagePullPolicy : Always
997
1004
volumeMounts :
998
1005
- mountPath : /tmp
999
1006
name : tmp
0 commit comments