@@ -53,20 +53,20 @@ function start_services() {
53
53
export TGI_LLM_ENDPOINT_CODEGEN=" http://${ip_address} :8028"
54
54
export TGI_LLM_ENDPOINT_FAQGEN=" http://${ip_address} :9009"
55
55
export TGI_LLM_ENDPOINT_DOCSUM=" http://${ip_address} :9009"
56
- export BACKEND_SERVICE_ENDPOINT_CHATQNA=" http://${host_ip } :8888/v1/chatqna"
57
- export BACKEND_SERVICE_ENDPOINT_FAQGEN=" http://${host_ip } :8889/v1/faqgen"
58
- export DATAPREP_DELETE_FILE_ENDPOINT=" http://${host_ip } :6009/v1/dataprep/delete_file"
59
- export BACKEND_SERVICE_ENDPOINT_CODEGEN=" http://${host_ip } :7778/v1/codegen"
60
- export BACKEND_SERVICE_ENDPOINT_DOCSUM=" http://${host_ip } :8890/v1/docsum"
61
- export DATAPREP_SERVICE_ENDPOINT=" http://${host_ip } :6007/v1/dataprep"
62
- export DATAPREP_GET_FILE_ENDPOINT=" http://${host_ip } :6008/v1/dataprep/get_file"
63
- export CHAT_HISTORY_CREATE_ENDPOINT=" http://${host_ip } :6012/v1/chathistory/create"
64
- export CHAT_HISTORY_CREATE_ENDPOINT=" http://${host_ip } :6012/v1/chathistory/create"
65
- export CHAT_HISTORY_DELETE_ENDPOINT=" http://${host_ip } :6012/v1/chathistory/delete"
66
- export CHAT_HISTORY_GET_ENDPOINT=" http://${host_ip } :6012/v1/chathistory/get"
67
- export PROMPT_SERVICE_GET_ENDPOINT=" http://${host_ip } :6015/v1/prompt/get"
68
- export PROMPT_SERVICE_CREATE_ENDPOINT=" http://${host_ip } :6015/v1/prompt/create"
69
- export KEYCLOAK_SERVICE_ENDPOINT=" http://${host_ip } :8080"
56
+ export BACKEND_SERVICE_ENDPOINT_CHATQNA=" http://${ip_address } :8888/v1/chatqna"
57
+ export BACKEND_SERVICE_ENDPOINT_FAQGEN=" http://${ip_address } :8889/v1/faqgen"
58
+ export DATAPREP_DELETE_FILE_ENDPOINT=" http://${ip_address } :6009/v1/dataprep/delete_file"
59
+ export BACKEND_SERVICE_ENDPOINT_CODEGEN=" http://${ip_address } :7778/v1/codegen"
60
+ export BACKEND_SERVICE_ENDPOINT_DOCSUM=" http://${ip_address } :8890/v1/docsum"
61
+ export DATAPREP_SERVICE_ENDPOINT=" http://${ip_address } :6007/v1/dataprep"
62
+ export DATAPREP_GET_FILE_ENDPOINT=" http://${ip_address } :6008/v1/dataprep/get_file"
63
+ export CHAT_HISTORY_CREATE_ENDPOINT=" http://${ip_address } :6012/v1/chathistory/create"
64
+ export CHAT_HISTORY_CREATE_ENDPOINT=" http://${ip_address } :6012/v1/chathistory/create"
65
+ export CHAT_HISTORY_DELETE_ENDPOINT=" http://${ip_address } :6012/v1/chathistory/delete"
66
+ export CHAT_HISTORY_GET_ENDPOINT=" http://${ip_address } :6012/v1/chathistory/get"
67
+ export PROMPT_SERVICE_GET_ENDPOINT=" http://${ip_address } :6015/v1/prompt/get"
68
+ export PROMPT_SERVICE_CREATE_ENDPOINT=" http://${ip_address } :6015/v1/prompt/create"
69
+ export KEYCLOAK_SERVICE_ENDPOINT=" http://${ip_address } :8080"
70
70
export MONGO_HOST=${ip_address}
71
71
export MONGO_PORT=27017
72
72
export DB_NAME=" opea"
@@ -235,23 +235,23 @@ function validate_microservices() {
235
235
236
236
# FAQGen llm microservice
237
237
validate_service \
238
- " ${ip_address} :${LLM_SERVICE_HOST_PORT_FAQGEN} /v1/faqgen" \
238
+ " ${ip_address} :9002 /v1/faqgen" \
239
239
" data: " \
240
240
" llm_faqgen" \
241
241
" llm-faqgen-server" \
242
242
' {"query":"Text Embeddings Inference (TEI) is a toolkit for deploying and serving open source text embeddings and sequence classification models. TEI enables high-performance extraction for the most popular models, including FlagEmbedding, Ember, GTE and E5."}'
243
243
244
244
# Docsum llm microservice
245
245
validate_service \
246
- " ${ip_address} :${LLM_SERVICE_HOST_PORT_DOCSUM} /v1/chat/docsum" \
246
+ " ${ip_address} :9003 /v1/chat/docsum" \
247
247
" data: " \
248
248
" llm_docsum" \
249
249
" llm-docsum-server" \
250
250
' {"query":"Text Embeddings Inference (TEI) is a toolkit for deploying and serving open source text embeddings and sequence classification models. TEI enables high-performance extraction for the most popular models, including FlagEmbedding, Ember, GTE and E5."}'
251
251
252
252
# CodeGen llm microservice
253
253
validate_service \
254
- " ${ip_address} :${LLM_SERVICE_HOST_PORT_CODEGEN} /v1/chat/completions" \
254
+ " ${ip_address} :9001 /v1/chat/completions" \
255
255
" data: " \
256
256
" llm_codegen" \
257
257
" llm-tgi-server-codegen" \
0 commit comments