Skip to content

Commit

Permalink
update test scripts and bug fix (#464)
Browse files Browse the repository at this point in the history
Signed-off-by: chensuyue <[email protected]>
Signed-off-by: zepan <[email protected]>
  • Loading branch information
ZePan110 authored Aug 13, 2024
1 parent 2891cc6 commit ab8ebc4
Show file tree
Hide file tree
Showing 15 changed files with 57 additions and 120 deletions.
1 change: 1 addition & 0 deletions comps/dataprep/redis/langchain_ray/requirements.txt
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,7 @@ pyarrow
pymupdf
python-bidi==0.4.2
python-docx
python-multipart
python-pptx
ray
redis
Expand Down
22 changes: 11 additions & 11 deletions tests/test_agent_langchain.sh
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ function build_docker_images() {
echo "Building the docker images"
cd $WORKPATH
echo $WORKPATH
docker build -t opea/comps-agent-langchain:latest -f comps/agent/langchain/docker/Dockerfile .
docker build -t opea/comps-agent-langchain:comps -f comps/agent/langchain/docker/Dockerfile .

}

Expand All @@ -24,27 +24,27 @@ function start_service() {

#single card
echo "start tgi gaudi service"
docker run -d --runtime=habana --name "comps-tgi-gaudi-service" -p 8080:80 -v ./data:/data -e HF_TOKEN=$HF_TOKEN -e HABANA_VISIBLE_DEVICES=all -e OMPI_MCA_btl_vader_single_copy_mechanism=none --cap-add=sys_nice --ipc=host ghcr.io/huggingface/tgi-gaudi:latest --model-id $model --max-input-tokens 4096 --max-total-tokens 8092
docker run -d --runtime=habana --name "test-comps-tgi-gaudi-service" -p 8080:80 -v ./data:/data -e HF_TOKEN=$HF_TOKEN -e HABANA_VISIBLE_DEVICES=all -e OMPI_MCA_btl_vader_single_copy_mechanism=none --cap-add=sys_nice --ipc=host ghcr.io/huggingface/tgi-gaudi:latest --model-id $model --max-input-tokens 4096 --max-total-tokens 8092
sleep 5s
docker logs comps-tgi-gaudi-service
docker logs test-comps-tgi-gaudi-service

echo "Starting agent microservice"
docker run -d --runtime=runc --name="comps-langchain-agent-endpoint" -v $WORKPATH/comps/agent/langchain/tools:/home/user/comps/agent/langchain/tools -p 9090:9090 --ipc=host -e HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -e model=${model} -e strategy=react -e llm_endpoint_url=http://${ip_address}:8080 -e llm_engine=tgi -e recursion_limit=5 -e require_human_feedback=false -e tools=/home/user/comps/agent/langchain/tools/custom_tools.yaml opea/comps-agent-langchain:latest
docker run -d --runtime=runc --name="test-comps-langchain-agent-endpoint" -v $WORKPATH/comps/agent/langchain/tools:/home/user/comps/agent/langchain/tools -p 9090:9090 --ipc=host -e HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -e model=${model} -e strategy=react -e llm_endpoint_url=http://${ip_address}:8080 -e llm_engine=tgi -e recursion_limit=5 -e require_human_feedback=false -e tools=/home/user/comps/agent/langchain/tools/custom_tools.yaml opea/comps-agent-langchain:comps
sleep 5s
docker logs comps-langchain-agent-endpoint
docker logs test-comps-langchain-agent-endpoint

echo "Waiting tgi gaudi ready"
n=0
until [[ "$n" -ge 100 ]] || [[ $ready == true ]]; do
docker logs comps-tgi-gaudi-service > ${WORKPATH}/tests/tgi-gaudi-service.log
docker logs test-comps-tgi-gaudi-service > ${WORKPATH}/tests/tgi-gaudi-service.log
n=$((n+1))
if grep -q Connected ${WORKPATH}/tests/tgi-gaudi-service.log; then
break
fi
sleep 5s
done
sleep 5s
docker logs comps-tgi-gaudi-service
docker logs test-comps-tgi-gaudi-service
echo "Service started successfully"
}

Expand Down Expand Up @@ -72,17 +72,17 @@ function validate_microservice() {
local EXIT_CODE="${EXIT_CODE:0-1}"
echo "return value is $EXIT_CODE"
if [ "$EXIT_CODE" == "1" ]; then
docker logs comps-tgi-gaudi-service &> ${LOG_PATH}/test-comps-tgi-gaudi-service.log
docker logs comps-langchain-agent-endpoint &> ${LOG_PATH}/test-comps-langchain-agent-endpoint.log
docker logs test-comps-tgi-gaudi-service &> ${LOG_PATH}/test-comps-tgi-gaudi-service.log
docker logs test-comps-langchain-agent-endpoint &> ${LOG_PATH}/test-comps-langchain-agent-endpoint.log
exit 1
fi
}

function stop_docker() {
cid=$(docker ps -aq --filter "name=comps-tgi-gaudi-service")
cid=$(docker ps -aq --filter "name=test-comps-tgi-gaudi-service")
echo "Stopping the docker containers "${cid}
if [[ ! -z "$cid" ]]; then docker rm $cid -f && sleep 1s; fi
cid=$(docker ps -aq --filter "name=comps-langchain-agent-endpoint")
cid=$(docker ps -aq --filter "name=test-comps-langchain-agent-endpoint")
echo "Stopping the docker containers "${cid}
if [[ ! -z "$cid" ]]; then docker rm $cid -f && sleep 1s; fi
echo "Docker containers stopped successfully"
Expand Down
25 changes: 12 additions & 13 deletions tests/test_dataprep_pgvector.sh
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
# Copyright (C) 2024 Intel Corporation
# SPDX-License-Identifier: Apache-2.0

set -xe
set -x

WORKPATH=$(dirname "$PWD")
LOG_PATH="$WORKPATH/tests"
Expand All @@ -16,20 +16,19 @@ function build_docker_images() {
docker pull pgvector/pgvector:0.7.0-pg16

# build dataprep image for pgvector
docker build -t opea/dataprep-pgvector:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f $WORKPATH/comps/dataprep/pgvector/langchain/docker/Dockerfile .
docker build -t opea/dataprep-pgvector:comps --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f $WORKPATH/comps/dataprep/pgvector/langchain/docker/Dockerfile .
}

function start_service() {
export POSTGRES_USER=testuser
export POSTGRES_PASSWORD=testpwd
export POSTGRES_DB=vectordb


docker run --name vectorstore-postgres -e POSTGRES_USER=${POSTGRES_USER} -e POSTGRES_HOST_AUTH_METHOD=trust -e POSTGRES_DB=${POSTGRES_DB} -e POSTGRES_PASSWORD=${POSTGRES_PASSWORD} -p 5432:5432 -d -v $WORKPATH/comps/vectorstores/langchain/pgvector/init.sql:/docker-entrypoint-initdb.d/init.sql pgvector/pgvector:0.7.0-pg16
docker run --name test-comps-vectorstore-postgres -e POSTGRES_USER=${POSTGRES_USER} -e POSTGRES_HOST_AUTH_METHOD=trust -e POSTGRES_DB=${POSTGRES_DB} -e POSTGRES_PASSWORD=${POSTGRES_PASSWORD} -p 5432:5432 -d -v $WORKPATH/comps/vectorstores/langchain/pgvector/init.sql:/docker-entrypoint-initdb.d/init.sql pgvector/pgvector:0.7.0-pg16

sleep 10s

docker run -d --name="dataprep-pgvector" -p ${dataprep_service_port}:6007 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy -e PG_CONNECTION_STRING=postgresql+psycopg2://${POSTGRES_USER}:${POSTGRES_PASSWORD}@$ip_address:5432/${POSTGRES_DB} opea/dataprep-pgvector:latest
docker run -d --name="test-comps-dataprep-pgvector" -p ${dataprep_service_port}:6007 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy -e PG_CONNECTION_STRING=postgresql+psycopg2://${POSTGRES_USER}:${POSTGRES_PASSWORD}@$ip_address:5432/${POSTGRES_DB} opea/dataprep-pgvector:comps

sleep 3m
}
Expand All @@ -50,12 +49,12 @@ function validate_microservice() {
echo "[ dataprep ] Content is as expected."
else
echo "[ dataprep ] Content does not match the expected result: $CONTENT"
docker logs dataprep-pgvector >> ${LOG_PATH}/dataprep.log
docker logs test-comps-dataprep-pgvector >> ${LOG_PATH}/dataprep.log
exit 1
fi
else
echo "[ dataprep ] HTTP status is not 200. Received status was $HTTP_STATUS"
docker logs dataprep-pgvector >> ${LOG_PATH}/dataprep.log
docker logs test-comps-dataprep-pgvector >> ${LOG_PATH}/dataprep.log
exit 1
fi

Expand All @@ -70,12 +69,12 @@ function validate_microservice() {
echo "[ dataprep - file ] Content is as expected."
else
echo "[ dataprep - file ] Content does not match the expected result: $CONTENT"
docker logs dataprep-pgvector >> ${LOG_PATH}/dataprep_file.log
docker logs test-comps-dataprep-pgvector >> ${LOG_PATH}/dataprep_file.log
exit 1
fi
else
echo "[ dataprep - file ] HTTP status is not 200. Received status was $HTTP_STATUS"
docker logs dataprep-pgvector >> ${LOG_PATH}/dataprep_file.log
docker logs test-comps-dataprep-pgvector >> ${LOG_PATH}/dataprep_file.log
exit 1
fi

Expand All @@ -84,19 +83,19 @@ function validate_microservice() {
HTTP_STATUS=$(curl -s -o /dev/null -w "%{http_code}" -X POST -d '{"file_path": "dataprep_file.txt"}' -H 'Content-Type: application/json' "$URL")
if [ "$HTTP_STATUS" -eq 200 ]; then
echo "[ dataprep - del ] HTTP status is 200."
docker logs dataprep-pgvector >> ${LOG_PATH}/dataprep_del.log
docker logs test-comps-dataprep-pgvector >> ${LOG_PATH}/dataprep_del.log
else
echo "[ dataprep - del ] HTTP status is not 200. Received status was $HTTP_STATUS"
docker logs dataprep-pgvector >> ${LOG_PATH}/dataprep_del.log
docker logs test-comps-dataprep-pgvector >> ${LOG_PATH}/dataprep_del.log
exit 1
fi
}

function stop_docker() {
cid=$(docker ps -aq --filter "name=vectorstore-postgres*")
cid=$(docker ps -aq --filter "name=test-comps-vectorstore-postgres*")
if [[ ! -z "$cid" ]]; then docker stop $cid && docker rm $cid && sleep 1s; fi

cid=$(docker ps -aq --filter "name=dataprep-pgvector*")
cid=$(docker ps -aq --filter "name=test-comps-dataprep-pgvector*")
if [[ ! -z "$cid" ]]; then docker stop $cid && docker rm $cid && sleep 1s; fi
}

Expand Down
4 changes: 2 additions & 2 deletions tests/test_dataprep_redis_langchain_ray.sh
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ ip_address=$(hostname -I | awk '{print $1}')
function build_docker_images() {
echo "Building the docker images"
cd $WORKPATH
docker build -t opea/dataprep-on-ray-redis:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/redis/langchain_ray/docker/Dockerfile .
docker build -t opea/dataprep-on-ray-redis:comps --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/redis/langchain_ray/docker/Dockerfile .
echo "Docker image built successfully"
}

Expand All @@ -24,7 +24,7 @@ function start_service() {
export REDIS_URL="redis://${ip_address}:6382"
export INDEX_NAME="rag-redis"
echo "Starting dataprep-redis-server"
docker run -d --name="test-comps-dataprep-redis-ray-server" --runtime=runc -p 6009:6007 -p 6010:6008 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy -e REDIS_URL=$REDIS_URL -e INDEX_NAME=$INDEX_NAME -e TEI_ENDPOINT=$TEI_ENDPOINT -e TIMEOUT_SECONDS=600 opea/dataprep-on-ray-redis:latest
docker run -d --name="test-comps-dataprep-redis-ray-server" --runtime=runc -p 6009:6007 -p 6010:6008 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy -e REDIS_URL=$REDIS_URL -e INDEX_NAME=$INDEX_NAME -e TEI_ENDPOINT=$TEI_ENDPOINT -e TIMEOUT_SECONDS=600 opea/dataprep-on-ray-redis:comps

sleep 10
echo "Service started successfully"
Expand Down
12 changes: 6 additions & 6 deletions tests/test_embeddings_llama_index.sh
Original file line number Diff line number Diff line change
Expand Up @@ -11,24 +11,24 @@ ip_address=$(hostname -I | awk '{print $1}')
function build_docker_images() {
cd $WORKPATH
echo $(pwd)
docker build --no-cache -t opea/embedding-tei:comps --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/llama_index/docker/Dockerfile .
docker build --no-cache -t opea/embedding-tei-llamaindex:comps --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/llama_index/docker/Dockerfile .
}

function start_service() {
tei_endpoint=5001
model="BAAI/bge-large-en-v1.5"
revision="refs/pr/5"
docker run -d --name="test-comps-embedding-tei-endpoint" -p $tei_endpoint:80 -v ./data:/data -e http_proxy=$http_proxy -e https_proxy=$https_proxy --pull always ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 --model-id $model --revision $revision
docker run -d --name="test-comps-embedding-tei-llamaindex-endpoint" -p $tei_endpoint:80 -v ./data:/data -e http_proxy=$http_proxy -e https_proxy=$https_proxy --pull always ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 --model-id $model --revision $revision
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:${tei_endpoint}"
tei_service_port=5010
docker run -d --name="test-comps-embedding-tei-server" -e http_proxy=$http_proxy -e https_proxy=$https_proxy -p ${tei_service_port}:6000 --ipc=host -e TEI_EMBEDDING_ENDPOINT=$TEI_EMBEDDING_ENDPOINT opea/embedding-tei:comps
docker run -d --name="test-comps-embedding-tei-llamaindex-server" -e http_proxy=$http_proxy -e https_proxy=$https_proxy -p ${tei_service_port}:6000 --ipc=host -e TEI_EMBEDDING_ENDPOINT=$TEI_EMBEDDING_ENDPOINT opea/embedding-tei-llamaindex:comps
sleep 3m
}

function validate_microservice() {
tei_service_port=5010
URL="http://${ip_address}:$tei_service_port/v1/embeddings"
docker logs test-comps-embedding-tei-server >> ${LOG_PATH}/embedding.log
docker logs test-comps-embedding-tei-llamaindex-server >> ${LOG_PATH}/embedding.log
HTTP_STATUS=$(curl -s -o /dev/null -w "%{http_code}" -X POST -d '{"text":"What is Deep Learning?"}' -H 'Content-Type: application/json' "$URL")
if [ "$HTTP_STATUS" -eq 200 ]; then
echo "[ embedding - llama_index ] HTTP status is 200. Checking content..."
Expand All @@ -38,12 +38,12 @@ function validate_microservice() {
echo "[ embedding - llama_index ] Content is as expected."
else
echo "[ embedding - llama_index ] Content does not match the expected result: $CONTENT"
docker logs test-comps-embedding-tei-server >> ${LOG_PATH}/embedding.log
docker logs test-comps-embedding-tei-llamaindex-server >> ${LOG_PATH}/embedding.log
exit 1
fi
else
echo "[ embedding - llama_index ] HTTP status is not 200. Received status was $HTTP_STATUS"
docker logs test-comps-embedding-tei-server >> ${LOG_PATH}/embedding.log
docker logs test-comps-embedding-tei-llamaindex-server >> ${LOG_PATH}/embedding.log
exit 1
fi
}
Expand Down
12 changes: 6 additions & 6 deletions tests/test_guardrails_llama_guard.sh
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ function build_docker_images() {
echo "Start building docker images for microservice"
cd $WORKPATH
docker pull ghcr.io/huggingface/tgi-gaudi:2.0.1
docker build --no-cache -t opea/guardrails-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/guardrails/llama_guard/docker/Dockerfile .
docker build --no-cache -t opea/guardrails-tgi:comps --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/guardrails/llama_guard/docker/Dockerfile .
echo "Docker images built"
}

Expand All @@ -21,9 +21,9 @@ function start_service() {
export SAFETY_GUARD_MODEL_ID="meta-llama/Meta-Llama-Guard-2-8B"
export SAFETY_GUARD_ENDPOINT=http://${ip_address}:8088/v1/chat/completions

docker run -d --name="test-guardrails-langchain-tgi-server" -p 8088:80 --runtime=habana -e HF_TOKEN=$HF_TOKEN -e HABANA_VISIBLE_DEVICES=all -e OMPI_MCA_btl_vader_single_copy_mechanism=none --cap-add=sys_nice --ipc=host -e HTTPS_PROXY=$https_proxy -e HTTP_PROXY=$https_proxy ghcr.io/huggingface/tgi-gaudi:2.0.1 --model-id $model_id --max-input-length 1024 --max-total-tokens 2048
docker run -d --name="test-comps-guardrails-langchain-tgi-server" -p 8088:80 --runtime=habana -e HF_TOKEN=$HF_TOKEN -e HABANA_VISIBLE_DEVICES=all -e OMPI_MCA_btl_vader_single_copy_mechanism=none --cap-add=sys_nice --ipc=host -e HTTPS_PROXY=$https_proxy -e HTTP_PROXY=$https_proxy ghcr.io/huggingface/tgi-gaudi:2.0.1 --model-id $model_id --max-input-length 1024 --max-total-tokens 2048
sleep 4m
docker run -d --name="test-guardrails-langchain-service" -p 9090:9090 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy -e no_proxy=$no_proxy -e SAFETY_GUARD_MODEL_ID=$SAFETY_GUARD_MODEL_ID -e SAFETY_GUARD_ENDPOINT=$SAFETY_GUARD_ENDPOINT -e HUGGINGFACEHUB_API_TOKEN=$HF_TOKEN opea/guardrails-tgi:latest
docker run -d --name="test-comps-guardrails-langchain-service" -p 9090:9090 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy -e no_proxy=$no_proxy -e SAFETY_GUARD_MODEL_ID=$SAFETY_GUARD_MODEL_ID -e SAFETY_GUARD_ENDPOINT=$SAFETY_GUARD_ENDPOINT -e HUGGINGFACEHUB_API_TOKEN=$HF_TOKEN opea/guardrails-tgi:comps
sleep 10s

echo "Microservice started"
Expand All @@ -32,8 +32,8 @@ function start_service() {
function validate_microservice() {
echo "Validate microservice started"
echo "test 1 - violated policies"
docker logs test-guardrails-langchain-tgi-server
docker logs test-guardrails-langchain-service
docker logs test-comps-guardrails-langchain-tgi-server
docker logs test-comps-guardrails-langchain-service
result=$(http_proxy= curl http://localhost:9090/v1/guardrails -X POST -d '{"text":"How do you buy a tiger in the US?","parameters":{"max_new_tokens":32}}' -H 'Content-Type: application/json')
if [[ $result == *"Violated"* ]]; then
echo "Result correct."
Expand All @@ -53,7 +53,7 @@ function validate_microservice() {
}

function stop_docker() {
cid=$(docker ps -aq --filter "name=test-guardrails-langchain*")
cid=$(docker ps -aq --filter "name=test-comps-guardrails-langchain*")
echo "Shutdown legacy containers "$cid
if [[ ! -z "$cid" ]]; then docker stop $cid && docker rm $cid && sleep 1s; fi
}
Expand Down
8 changes: 4 additions & 4 deletions tests/test_guardrails_pii_detection.sh
Original file line number Diff line number Diff line change
Expand Up @@ -10,13 +10,13 @@ ip_address=$(hostname -I | awk '{print $1}')
function build_docker_images() {
echo "Start building docker images for microservice"
cd $WORKPATH
docker build -t opea/guardrails-pii-detection:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/guardrails/pii_detection/docker/Dockerfile .
docker build -t opea/guardrails-pii-detection:comps --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/guardrails/pii_detection/docker/Dockerfile .
echo "Docker images built"
}

function start_service() {
echo "Starting microservice"
docker run -d --runtime=runc --name="test-guardrails-pii-detection-endpoint" -p 6357:6357 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy opea/guardrails-pii-detection:latest
docker run -d --runtime=runc --name="test-comps-guardrails-pii-detection-endpoint" -p 6357:6357 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy opea/guardrails-pii-detection:comps
sleep 5
echo "Microservice started"
}
Expand All @@ -34,11 +34,11 @@ function validate_microservice() {
echo "test 4 - 20 tasks in parallel - ml"
python comps/guardrails/pii_detection/test.py --test_text --batch_size 20 --ip_addr $ip_address --strategy ml
echo "Validate microservice completed"
docker logs test-guardrails-pii-detection-endpoint
docker logs test-comps-guardrails-pii-detection-endpoint
}

function stop_docker() {
cid=$(docker ps -aq --filter "name=test-guardrails-pii-detection-endpoint")
cid=$(docker ps -aq --filter "name=test-comps-guardrails-pii-detection-endpoint")
echo "Shutdown legacy containers "$cid
if [[ ! -z "$cid" ]]; then docker stop $cid && docker rm $cid && sleep 1s; fi
}
Expand Down
2 changes: 1 addition & 1 deletion tests/test_llms_text-generation_ray_serve.sh
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ function build_docker_images() {
docker build \
-f comps/llms/text-generation/ray_serve/docker/Dockerfile.rayserve \
--network=host \
-t ray_serve:habana .
-t ray_serve_habana:comps .

## Build OPEA microservice docker
cd $WORKPATH
Expand Down
4 changes: 2 additions & 2 deletions tests/test_llms_text-generation_vllm-openvino.sh
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ WORKPATH="$( cd "$( dirname "$0" )" && pwd )"
# Define variables
port=8123
HF_CACHE_DIR=$HOME/.cache/huggingface
DOCKER_IMAGE="vllm:openvino"
DOCKER_IMAGE="vllm-openvino:comps"
CONTAINER_NAME="vllm-openvino-container"

function build_container() {
Expand All @@ -34,7 +34,7 @@ start_container() {
-e HTTPS_PROXY=$https_proxy \
-e HTTP_PROXY=$https_proxy \
-v $HF_CACHE_DIR:/root/.cache/huggingface \
vllm:openvino /bin/bash -c "\
vllm-openvino:comps /bin/bash -c "\
cd / && \
export VLLM_CPU_KVCACHE_SPACE=50 && \
python3 -m vllm.entrypoints.openai.api_server \
Expand Down
2 changes: 1 addition & 1 deletion tests/test_llms_text-generation_vllm-ray.sh
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ function build_docker_images() {
cd $WORKPATH
docker build \
-f comps/llms/text-generation/vllm-ray/docker/Dockerfile.vllmray \
-t opea/vllm_ray:habana --network=host .
-t opea/vllm_ray-habana:comps --network=host .

## Build OPEA microservice docker
cd $WORKPATH
Expand Down
Loading

0 comments on commit ab8ebc4

Please sign in to comment.