The Embedding Microservice is designed to efficiently convert textual strings into vectorized embeddings, facilitating seamless integration into various machine learning and data processing workflows. This service utilizes advanced algorithms to generate high-quality embeddings that capture the semantic essence of the input text, making it ideal for applications in natural language processing, information retrieval, and similar fields.
Key Features:
High Performance: Optimized for quick and reliable conversion of textual data into vector embeddings.
Scalability: Built to handle high volumes of requests simultaneously, ensuring robust performance even under heavy loads.
Ease of Integration: Provides a simple and intuitive API, allowing for straightforward integration into existing systems and workflows.
Customizable: Supports configuration and customization to meet specific use case requirements, including different embedding models and preprocessing techniques.
Users are able to configure and build embedding-related services according to their actual needs.
Currently, we provide three ways to implement the embedding service:
-
Build the embedding model locally from the server, which is faster, but takes up memory on the local server.
-
Build it based on the TEI endpoint, which provides more flexibility, but may bring some network latency.
-
Build it based on the Prediction Guard endpoint, which provides performant, hosted embedding models on top of Gaudi, but needs an API key.
Regardless of the implementation, you need to install requirements first.
# run with langchain
pip install -r langchain/requirements.txt
# run with llama_index
pip install -r llama_index/requirements.txt
# run with predictionguard
pip install -r predictionguard/requirements.txt
You can select one of following ways to start the embedding service:
First, you need to start a PredictionGuard service.
docker run -d --name="embedding-predictionguard" \
-e http_proxy=$http_proxy -e https_proxy=$https_proxy \
-p 6000:6000 --ipc=host \
-e PREDICTIONGUARD_API_KEY=<your_api_key> \
opea/embedding-predictionguard:latest
Then you need to test your PredictionGuard service using the following commands:
curl http://localhost:6000/v1/embeddings\
-X POST \
-d '{"input":"Hello, world!"}' \
-H 'Content-Type: application/json'
Start the embedding service
# run with predictionguard
cd predictionguard
export PREDICTIONGUARD_API_KEY=${your_api_key}
python embedding_pg.py
First, you need to start a TEI service.
your_port=8090
model="BAAI/bge-large-en-v1.5"
docker run -p $your_port:80 -v ./data:/data --name tei_server -e http_proxy=$http_proxy -e https_proxy=$https_proxy --pull always ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 --model-id $model
Then you need to test your TEI service using the following commands:
curl localhost:$your_port/embed \
-X POST \
-d '{"inputs":"What is Deep Learning?"}' \
-H 'Content-Type: application/json'
Start the embedding service with the TEI_EMBEDDING_ENDPOINT.
# run with langchain
cd langchain
# run with llama_index
cd llama_index
export TEI_EMBEDDING_ENDPOINT="http://localhost:$yourport"
export TEI_EMBEDDING_MODEL_NAME="BAAI/bge-large-en-v1.5"
python embedding_tei.py
# run with langchain
cd langchain
# run with llama_index
cd llama_index
python local_embedding.py
First, you need to start a TEI service.
your_port=8090
model="BAAI/bge-large-en-v1.5"
docker run -p $your_port:80 -v ./data:/data --name tei_server -e http_proxy=$http_proxy -e https_proxy=$https_proxy --pull always ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 --model-id $model
Then you need to test your TEI service using the following commands:
curl localhost:$your_port/embed \
-X POST \
-d '{"inputs":"What is Deep Learning?"}' \
-H 'Content-Type: application/json'
Export the TEI_EMBEDDING_ENDPOINT
for later usage:
export TEI_EMBEDDING_ENDPOINT="http://localhost:$yourport"
export TEI_EMBEDDING_MODEL_NAME="BAAI/bge-large-en-v1.5"
First, build the Docker image for the PredictionGuard embedding microservice:
docker build -t opea/embedding-predictionguard:latest --build-arg http_proxy=$http_proxy --build-arg https_proxy=$https_proxy -f comps/embeddings/predictionguard/docker/Dockerfile .
Start the Docker container for the PredictionGuard embedding microservice. Replace <your_api_key> with your PredictionGuard API key.
docker run -d --name="embedding-predictionguard" \
-e http_proxy=$http_proxy -e https_proxy=$https_proxy \
-p 6000:6000 --ipc=host \
-e PREDICTIONGUARD_API_KEY=$PREDICTIONGUARD_API_KEY \
opea/embedding-predictionguard:latest
Then you need to test your PredictionGuard service using the following commands:
curl http://localhost:6000/v1/embeddings\
-X POST \
-d '{"input":"Hello, world!"}' \
-H 'Content-Type: application/json'
cd ../../
docker build -t opea/embedding-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile .
cd ../../
docker build -t opea/embedding-tei-llama-index:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/llama_index/docker/Dockerfile .
docker build -t opea/embedding-predictionguard:latest --build-arg http_proxy=$http_proxy --build-arg https_proxy=$https_proxy -f comps/embeddings/predictionguard/docker/Dockerfile .
# run with langchain docker
docker run -d --name="embedding-tei-server" -p 6000:6000 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy -e TEI_EMBEDDING_ENDPOINT=$TEI_EMBEDDING_ENDPOINT -e TEI_EMBEDDING_MODEL_NAME=$TEI_EMBEDDING_MODEL_NAME opea/embedding-tei:latest
# run with llama-index docker
docker run -d --name="embedding-tei-llama-index-server" -p 6000:6000 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy -e TEI_EMBEDDING_ENDPOINT=$TEI_EMBEDDING_ENDPOINT -e TEI_EMBEDDING_MODEL_NAME=$TEI_EMBEDDING_MODEL_NAME opea/embedding-tei-llama-index:latest
docker run -d --name="embedding-predictionguard" -e http_proxy=$http_proxy -e https_proxy=$https_proxy -p 6000:6000 --ipc=host -e PREDICTIONGUARD_API_KEY=$PREDICTIONGUARD_API_KEY opea/embedding-predictionguard:latest
cd docker
docker compose -f docker_compose_embedding.yaml up -d
curl http://localhost:6000/v1/health_check\
-X GET \
-H 'Content-Type: application/json'
curl http://localhost:6000/v1/embeddings\
-X POST \
-d '{"text":"Hello, world!"}' \
-H 'Content-Type: application/json'