File tree 11 files changed +58
-113
lines changed
11 files changed +58
-113
lines changed Original file line number Diff line number Diff line change 6
6
7
7
# export host_ip=<your External Public IP> # export host_ip=$(hostname -I | awk '{print $1}')
8
8
9
- export host_ip=" 192.165.1.21 "
10
- export HUGGINGFACEHUB_API_TOKEN=${YOUR_HUGGINGFACEHUB_API_TOKEN }
9
+ export host_ip=${ip_address}
10
+ export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN }
11
11
# <token>
12
12
13
13
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
Original file line number Diff line number Diff line change 6
6
7
7
# export host_ip=<your External Public IP> # export host_ip=$(hostname -I | awk '{print $1}')
8
8
9
- export host_ip=" "
10
- export external_host_ip=" "
9
+ export host_ip=${ip_address}
10
+ export external_host_ip=${ip_address}
11
11
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
12
12
export HF_CACHE_DIR=" ./data"
13
13
export LLM_MODEL_ID=" Intel/neural-chat-7b-v3-3"
Original file line number Diff line number Diff line change @@ -14,7 +14,8 @@ export MEGA_SERVICE_HOST_IP=${host_ip}
14
14
export WHISPER_SERVER_HOST_IP=${host_ip}
15
15
export SPEECHT5_SERVER_HOST_IP=${host_ip}
16
16
export LLM_SERVER_HOST_IP=${host_ip}
17
-
17
+ export GPT_SOVITS_SERVER_HOST_IP=${host_ip}
18
+ export GPT_SOVITS_SERVER_PORT=9880
18
19
export WHISPER_SERVER_PORT=7066
19
20
export SPEECHT5_SERVER_PORT=7055
20
21
export LLM_SERVER_PORT=3006
Original file line number Diff line number Diff line change
1
+ # AudioQnA E2E test scripts
2
+
3
+ ## Set the required environment variable
4
+
5
+ ``` bash
6
+ export HUGGINGFACEHUB_API_TOKEN=" Your_Huggingface_API_Token"
7
+ ```
8
+
9
+ ## Run test
10
+
11
+ On Intel Xeon with TGI:
12
+
13
+ ``` bash
14
+ bash test_compose_tgi_on_xeon.sh
15
+ ```
16
+
17
+ On Intel Xeon with vLLM:
18
+
19
+ ``` bash
20
+ bash test_compose_on_xeon.sh
21
+ ```
22
+
23
+ On Intel Gaudi with TGI:
24
+
25
+ ``` bash
26
+ bash test_compose_tgi_on_gaudi.sh
27
+ ```
28
+
29
+ On Intel Gaudi with vLLM:
30
+
31
+ ``` bash
32
+ bash test_compose_on_gaudi.sh
33
+ ```
34
+
35
+ On AMD ROCm with TGI:
36
+
37
+ ``` bash
38
+ bash test_compose_on_rocm.sh
39
+ ```
40
+
41
+ On AMD ROCm with vLLM:
42
+
43
+ ``` bash
44
+ bash test_compose_vllm_on_rocm.sh
45
+ ```
Original file line number Diff line number Diff line change @@ -40,21 +40,8 @@ function build_docker_images() {
40
40
41
41
function start_services() {
42
42
cd $WORKPATH /docker_compose/intel/cpu/xeon/
43
- export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
44
- export LLM_MODEL_ID=meta-llama/Meta-Llama-3-8B-Instruct
45
-
46
- export MEGA_SERVICE_HOST_IP=${ip_address}
47
- export WHISPER_SERVER_HOST_IP=${ip_address}
48
- export GPT_SOVITS_SERVER_HOST_IP=${ip_address}
49
- export LLM_SERVER_HOST_IP=${ip_address}
50
-
51
- export WHISPER_SERVER_PORT=7066
52
- export GPT_SOVITS_SERVER_PORT=9880
53
- export LLM_SERVER_PORT=3006
54
-
55
- export BACKEND_SERVICE_ENDPOINT=http://${ip_address} :3008/v1/audioqna
56
43
export host_ip=${ip_address}
57
-
44
+ source set_env.sh
58
45
# sed -i "s/backend_address/$ip_address/g" $WORKPATH/ui/svelte/.env
59
46
60
47
# Start Docker Containers
Original file line number Diff line number Diff line change @@ -40,24 +40,8 @@ function build_docker_images() {
40
40
41
41
function start_services() {
42
42
cd $WORKPATH /docker_compose/intel/hpu/gaudi
43
- export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
44
- export LLM_MODEL_ID=meta-llama/Meta-Llama-3-8B-Instruct
45
- export NUM_CARDS=1
46
- export BLOCK_SIZE=128
47
- export MAX_NUM_SEQS=256
48
- export MAX_SEQ_LEN_TO_CAPTURE=2048
49
-
50
- export MEGA_SERVICE_HOST_IP=${ip_address}
51
- export WHISPER_SERVER_HOST_IP=${ip_address}
52
- export SPEECHT5_SERVER_HOST_IP=${ip_address}
53
- export LLM_SERVER_HOST_IP=${ip_address}
54
-
55
- export WHISPER_SERVER_PORT=7066
56
- export SPEECHT5_SERVER_PORT=7055
57
- export LLM_SERVER_PORT=3006
58
-
59
- export BACKEND_SERVICE_ENDPOINT=http://${ip_address} :3008/v1/audioqna
60
43
export host_ip=${ip_address}
44
+ source set_env.sh
61
45
# sed -i "s/backend_address/$ip_address/g" $WORKPATH/ui/svelte/.env
62
46
63
47
# Start Docker Containers
Original file line number Diff line number Diff line change @@ -35,20 +35,7 @@ function build_docker_images() {
35
35
36
36
function start_services() {
37
37
cd $WORKPATH /docker_compose/amd/gpu/rocm/
38
- export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
39
- export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
40
-
41
- export MEGA_SERVICE_HOST_IP=${ip_address}
42
- export WHISPER_SERVER_HOST_IP=${ip_address}
43
- export SPEECHT5_SERVER_HOST_IP=${ip_address}
44
- export LLM_SERVER_HOST_IP=${ip_address}
45
-
46
- export WHISPER_SERVER_PORT=7066
47
- export SPEECHT5_SERVER_PORT=7055
48
- export LLM_SERVER_PORT=3006
49
-
50
- export BACKEND_SERVICE_ENDPOINT=http://${ip_address} :3008/v1/audioqna
51
-
38
+ source set_env.sh
52
39
# Start Docker Containers
53
40
docker compose up -d > ${LOG_PATH} /start_services_with_compose.log
54
41
n=0
Original file line number Diff line number Diff line change @@ -40,21 +40,8 @@ function build_docker_images() {
40
40
41
41
function start_services() {
42
42
cd $WORKPATH /docker_compose/intel/cpu/xeon/
43
- export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
44
- export LLM_MODEL_ID=meta-llama/Meta-Llama-3-8B-Instruct
45
-
46
- export MEGA_SERVICE_HOST_IP=${ip_address}
47
- export WHISPER_SERVER_HOST_IP=${ip_address}
48
- export SPEECHT5_SERVER_HOST_IP=${ip_address}
49
- export LLM_SERVER_HOST_IP=${ip_address}
50
-
51
- export WHISPER_SERVER_PORT=7066
52
- export SPEECHT5_SERVER_PORT=7055
53
- export LLM_SERVER_PORT=3006
54
-
55
- export BACKEND_SERVICE_ENDPOINT=http://${ip_address} :3008/v1/audioqna
56
43
export host_ip=${ip_address}
57
-
44
+ source set_env.sh
58
45
# sed -i "s/backend_address/$ip_address/g" $WORKPATH/ui/svelte/.env
59
46
60
47
# Start Docker Containers
Original file line number Diff line number Diff line change @@ -34,21 +34,8 @@ function build_docker_images() {
34
34
35
35
function start_services() {
36
36
cd $WORKPATH /docker_compose/intel/hpu/gaudi
37
- export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
38
- export LLM_MODEL_ID=meta-llama/Meta-Llama-3-8B-Instruct
39
-
40
- export MEGA_SERVICE_HOST_IP=${ip_address}
41
- export WHISPER_SERVER_HOST_IP=${ip_address}
42
- export SPEECHT5_SERVER_HOST_IP=${ip_address}
43
- export LLM_SERVER_HOST_IP=${ip_address}
44
-
45
- export WHISPER_SERVER_PORT=7066
46
- export SPEECHT5_SERVER_PORT=7055
47
- export LLM_SERVER_PORT=3006
48
-
49
- export BACKEND_SERVICE_ENDPOINT=http://${ip_address} :3008/v1/audioqna
50
37
export host_ip=${ip_address}
51
-
38
+ source set_env.sh
52
39
# Start Docker Containers
53
40
docker compose -f compose_tgi.yaml up -d > ${LOG_PATH} /start_services_with_compose.log
54
41
n=0
Original file line number Diff line number Diff line change @@ -34,21 +34,8 @@ function build_docker_images() {
34
34
35
35
function start_services() {
36
36
cd $WORKPATH /docker_compose/intel/cpu/xeon/
37
- export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
38
- export LLM_MODEL_ID=meta-llama/Meta-Llama-3-8B-Instruct
39
-
40
- export MEGA_SERVICE_HOST_IP=${ip_address}
41
- export WHISPER_SERVER_HOST_IP=${ip_address}
42
- export SPEECHT5_SERVER_HOST_IP=${ip_address}
43
- export LLM_SERVER_HOST_IP=${ip_address}
44
-
45
- export WHISPER_SERVER_PORT=7066
46
- export SPEECHT5_SERVER_PORT=7055
47
- export LLM_SERVER_PORT=3006
48
-
49
- export BACKEND_SERVICE_ENDPOINT=http://${ip_address} :3008/v1/audioqna
50
37
export host_ip=${ip_address}
51
-
38
+ source set_env.sh
52
39
# Start Docker Containers
53
40
docker compose -f compose_tgi.yaml up -d > ${LOG_PATH} /start_services_with_compose.log
54
41
n=0
Original file line number Diff line number Diff line change @@ -33,27 +33,7 @@ function build_docker_images() {
33
33
34
34
function start_services() {
35
35
cd $WORKPATH /docker_compose/amd/gpu/rocm/
36
-
37
- export host_ip=${ip_address}
38
- export external_host_ip=${ip_address}
39
- export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
40
- export HF_CACHE_DIR=" ./data"
41
- export LLM_MODEL_ID=" Intel/neural-chat-7b-v3-3"
42
- export VLLM_SERVICE_PORT=" 8081"
43
-
44
- export MEGA_SERVICE_HOST_IP=${host_ip}
45
- export WHISPER_SERVER_HOST_IP=${host_ip}
46
- export SPEECHT5_SERVER_HOST_IP=${host_ip}
47
- export LLM_SERVER_HOST_IP=${host_ip}
48
-
49
- export WHISPER_SERVER_PORT=7066
50
- export SPEECHT5_SERVER_PORT=7055
51
- export LLM_SERVER_PORT=${VLLM_SERVICE_PORT}
52
- export BACKEND_SERVICE_PORT=3008
53
- export FRONTEND_SERVICE_PORT=5173
54
-
55
- export BACKEND_SERVICE_ENDPOINT=http://${external_host_ip} :${BACKEND_SERVICE_PORT} /v1/audioqna
56
-
36
+ source set_env_vllm.sh
57
37
sed -i " s/backend_address/$ip_address /g" $WORKPATH /ui/svelte/.env
58
38
59
39
# Start Docker Containers
You can’t perform that action at this time.
0 commit comments