File tree Expand file tree Collapse file tree 3 files changed +3
-6
lines changed
AudioQnA/docker_compose/intel/cpu/xeon
ChatQnA/docker_compose/intel/cpu/xeon
CodeGen/docker_compose/intel/cpu/xeon Expand file tree Collapse file tree 3 files changed +3
-6
lines changed Original file line number Diff line number Diff line change @@ -25,14 +25,13 @@ services:
25
25
https_proxy : ${https_proxy}
26
26
restart : unless-stopped
27
27
vllm-service :
28
- image : openeuler/vllm-cpu:0.9 .1-oe2403lts
28
+ image : openeuler/vllm-cpu:0.10 .1-oe2403lts
29
29
container_name : vllm-service
30
30
ports :
31
31
- ${LLM_SERVER_PORT:-3006}:80
32
32
volumes :
33
33
- " ${MODEL_CACHE:-./data}:/root/.cache/huggingface/hub"
34
34
shm_size : 128g
35
- privileged : true
36
35
environment :
37
36
no_proxy : ${no_proxy}
38
37
http_proxy : ${http_proxy}
Original file line number Diff line number Diff line change @@ -88,14 +88,13 @@ services:
88
88
HF_HUB_ENABLE_HF_TRANSFER : 0
89
89
command : --model-id ${RERANK_MODEL_ID} --auto-truncate
90
90
vllm-service :
91
- image : openeuler/vllm-cpu:0.9 .1-oe2403lts
91
+ image : openeuler/vllm-cpu:0.10 .1-oe2403lts
92
92
container_name : vllm-service
93
93
ports :
94
94
- " 9009:80"
95
95
volumes :
96
96
- " ${MODEL_CACHE:-./data}:/root/.cache/huggingface/hub"
97
97
shm_size : 128g
98
- privileged : true
99
98
environment :
100
99
no_proxy : ${no_proxy}
101
100
http_proxy : ${http_proxy}
Original file line number Diff line number Diff line change 4
4
services :
5
5
6
6
vllm-service :
7
- image : openeuler/vllm-cpu:0.9 .1-oe2403lts
7
+ image : openeuler/vllm-cpu:0.10 .1-oe2403lts
8
8
container_name : vllm-server
9
9
ports :
10
10
- " 8028:80"
11
11
volumes :
12
12
- " ${MODEL_CACHE:-./data}:/root/.cache/huggingface/hub"
13
13
shm_size : 1g
14
- privileged : true
15
14
environment :
16
15
no_proxy : ${no_proxy}
17
16
http_proxy : ${http_proxy}
You can’t perform that action at this time.
0 commit comments