@@ -8,9 +8,9 @@ services:
8
8
image : ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
9
9
container_name : search-tei-embedding-server
10
10
ports :
11
- - " 3001:80"
11
+ - " ${SEARCH_TEI_EMBEDDING_PORT:- 3001} :80"
12
12
volumes :
13
- - " ${MODEL_PATH :-./data}:/data"
13
+ - " ${MODEL_CACHE :-./data}:/data"
14
14
shm_size : 1g
15
15
environment :
16
16
no_proxy : ${no_proxy}
@@ -20,13 +20,14 @@ services:
20
20
HUGGING_FACE_HUB_TOKEN : ${SEARCH_HUGGINGFACEHUB_API_TOKEN}
21
21
HUGGINGFACEHUB_API_TOKEN : ${SEARCH_HUGGINGFACEHUB_API_TOKEN}
22
22
command : --model-id ${SEARCH_EMBEDDING_MODEL_ID} --auto-truncate
23
+
23
24
search-embedding :
24
25
image : ${REGISTRY:-opea}/embedding:${TAG:-latest}
25
26
container_name : search-embedding-server
26
27
depends_on :
27
28
- search-tei-embedding-service
28
29
ports :
29
- - " 3002:6000"
30
+ - " ${SEARCH_EMBEDDING_SERVICE_PORT:- 3002} :6000"
30
31
ipc : host
31
32
environment :
32
33
no_proxy : ${no_proxy}
@@ -36,11 +37,12 @@ services:
36
37
TEI_EMBEDDING_ENDPOINT : ${SEARCH_TEI_EMBEDDING_ENDPOINT}
37
38
HF_TOKEN : ${SEARCH_HUGGINGFACEHUB_API_TOKEN}
38
39
restart : unless-stopped
40
+
39
41
search-web-retriever :
40
42
image : ${REGISTRY:-opea}/web-retriever:${TAG:-latest}
41
43
container_name : search-web-retriever-server
42
44
ports :
43
- - " 3003:7077"
45
+ - " ${SEARCH_WEB_RETRIEVER_SERVICE_PORT:- 3003} :7077"
44
46
ipc : host
45
47
environment :
46
48
no_proxy : ${no_proxy}
@@ -50,26 +52,28 @@ services:
50
52
GOOGLE_API_KEY : ${SEARCH_GOOGLE_API_KEY}
51
53
GOOGLE_CSE_ID : ${SEARCH_GOOGLE_CSE_ID}
52
54
restart : unless-stopped
55
+
53
56
search-tei-reranking-service :
54
57
image : ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
55
58
container_name : search-tei-reranking-server
56
59
ports :
57
- - " 3004:80"
60
+ - " ${SEARCH_TEI_RERANKING_PORT:- 3004} :80"
58
61
volumes :
59
- - " ${MODEL_PATH :-./data}:/data"
62
+ - " ${MODEL_CACHE :-./data}:/data"
60
63
shm_size : 1g
61
64
environment :
62
65
no_proxy : ${no_proxy}
63
66
http_proxy : ${http_proxy}
64
67
https_proxy : ${https_proxy}
65
68
command : --model-id ${SEARCH_RERANK_MODEL_ID} --auto-truncate
69
+
66
70
search-reranking :
67
71
image : ${REGISTRY:-opea}/reranking:${TAG:-latest}
68
72
container_name : search-reranking-server
69
73
depends_on :
70
74
- search-tei-reranking-service
71
75
ports :
72
- - " 3005:8000"
76
+ - " ${SEARCH_RERANK_SERVICE_PORT:- 3005} :8000"
73
77
ipc : host
74
78
environment :
75
79
no_proxy : ${no_proxy}
@@ -80,13 +84,14 @@ services:
80
84
HUGGING_FACE_HUB_TOKEN : ${SEARCH_HUGGINGFACEHUB_API_TOKEN}
81
85
HUGGINGFACEHUB_API_TOKEN : ${SEARCH_HUGGINGFACEHUB_API_TOKEN}
82
86
restart : unless-stopped
87
+
83
88
search-tgi-service :
84
89
image : ghcr.io/huggingface/text-generation-inference:2.3.1-rocm
85
90
container_name : search-tgi-service
86
91
ports :
87
- - " 3006:80"
92
+ - " ${SEARCH_TGI_SERVICE_PORT:- 3006} :80"
88
93
volumes :
89
- - " ${MODEL_PATH :-./data}:/data"
94
+ - " ${MODEL_CACHE :-./data}:/data"
90
95
environment :
91
96
no_proxy : ${no_proxy}
92
97
http_proxy : ${http_proxy}
@@ -96,7 +101,7 @@ services:
96
101
shm_size : 1g
97
102
devices :
98
103
- /dev/kfd:/dev/kfd
99
- - /dev/dri/ :/dev/dri/
104
+ - /dev/dri:/dev/dri
100
105
cap_add :
101
106
- SYS_PTRACE
102
107
group_add :
@@ -105,25 +110,26 @@ services:
105
110
- seccomp:unconfined
106
111
ipc : host
107
112
command : --model-id ${SEARCH_LLM_MODEL_ID} --max-input-length 1024 --max-total-tokens 2048
113
+
108
114
search-llm :
109
115
image : ${REGISTRY:-opea}/llm-textgen:${TAG:-latest}
110
116
container_name : search-llm-server
111
117
depends_on :
112
118
- search-tgi-service
113
119
ports :
114
- - " 3007:9000"
120
+ - " ${SEARCH_LLM_SERVICE_PORT:- 3007} :9000"
115
121
ipc : host
116
122
environment :
117
123
no_proxy : ${no_proxy}
118
124
http_proxy : ${http_proxy}
119
125
https_proxy : ${https_proxy}
120
- TGI_LLM_ENDPOINT : ${SEARCH_TGI_LLM_ENDPOINT}
121
126
HUGGINGFACEHUB_API_TOKEN : ${SEARCH_HUGGINGFACEHUB_API_TOKEN}
122
127
LLM_ENDPOINT : ${SEARCH_TGI_LLM_ENDPOINT}
123
128
LLM_MODEL_ID : ${SEARCH_LLM_MODEL_ID}
124
129
LLM_MODEL : ${SEARCH_LLM_MODEL_ID}
125
130
HF_TOKEN : ${SEARCH_HUGGINGFACEHUB_API_TOKEN}
126
- OPENAI_API_KEY : ${SEARCH_OPENAI_API_KEY}
131
+ LLM_COMPONENT_NAME : " OpeaTextGenService"
132
+
127
133
restart : unless-stopped
128
134
search-backend-server :
129
135
image : ${REGISTRY:-opea}/searchqna:${TAG:-latest}
@@ -139,18 +145,18 @@ services:
139
145
ports :
140
146
- " ${SEARCH_BACKEND_SERVICE_PORT:-3008}:8888"
141
147
environment :
142
- - no_proxy= ${no_proxy}
143
- - https_proxy= ${https_proxy}
144
- - http_proxy= ${http_proxy}
145
- - MEGA_SERVICE_HOST_IP= ${SEARCH_MEGA_SERVICE_HOST_IP}
146
- - EMBEDDING_SERVICE_HOST_IP= ${SEARCH_EMBEDDING_SERVICE_HOST_IP}
147
- - WEB_RETRIEVER_SERVICE_HOST_IP= ${SEARCH_WEB_RETRIEVER_SERVICE_HOST_IP}
148
- - RERANK_SERVICE_HOST_IP= ${SEARCH_RERANK_SERVICE_HOST_IP}
149
- - LLM_SERVICE_HOST_IP= ${SEARCH_LLM_SERVICE_HOST_IP}
150
- - EMBEDDING_SERVICE_PORT= ${SEARCH_EMBEDDING_SERVICE_PORT}
151
- - WEB_RETRIEVER_SERVICE_PORT= ${SEARCH_WEB_RETRIEVER_SERVICE_PORT}
152
- - RERANK_SERVICE_PORT= ${SEARCH_RERANK_SERVICE_PORT}
153
- - LLM_SERVICE_PORT= ${SEARCH_LLM_SERVICE_PORT}
148
+ no_proxy : ${no_proxy}
149
+ https_proxy : ${https_proxy}
150
+ http_proxy : ${http_proxy}
151
+ MEGA_SERVICE_HOST_IP : ${SEARCH_MEGA_SERVICE_HOST_IP}
152
+ EMBEDDING_SERVICE_HOST_IP : ${SEARCH_EMBEDDING_SERVICE_HOST_IP}
153
+ WEB_RETRIEVER_SERVICE_HOST_IP : ${SEARCH_WEB_RETRIEVER_SERVICE_HOST_IP}
154
+ RERANK_SERVICE_HOST_IP : ${SEARCH_RERANK_SERVICE_HOST_IP}
155
+ LLM_SERVICE_HOST_IP : ${SEARCH_LLM_SERVICE_HOST_IP}
156
+ EMBEDDING_SERVICE_PORT : ${SEARCH_EMBEDDING_SERVICE_PORT}
157
+ WEB_RETRIEVER_SERVICE_PORT : ${SEARCH_WEB_RETRIEVER_SERVICE_PORT}
158
+ RERANK_SERVICE_PORT : ${SEARCH_RERANK_SERVICE_PORT}
159
+ LLM_SERVICE_PORT : ${SEARCH_LLM_SERVICE_PORT}
154
160
ipc : host
155
161
restart : always
156
162
search-ui-server :
@@ -161,10 +167,10 @@ services:
161
167
ports :
162
168
- " ${SEARCH_FRONTEND_SERVICE_PORT:-5173}:5173"
163
169
environment :
164
- - no_proxy= ${no_proxy}
165
- - https_proxy= ${https_proxy}
166
- - http_proxy= ${http_proxy}
167
- - BACKEND_BASE_URL= ${SEARCH_BACKEND_SERVICE_ENDPOINT}
170
+ no_proxy : ${no_proxy}
171
+ https_proxy : ${https_proxy}
172
+ http_proxy : ${http_proxy}
173
+ BACKEND_BASE_URL : ${SEARCH_BACKEND_SERVICE_ENDPOINT}
168
174
ipc : host
169
175
restart : always
170
176
0 commit comments