-
Notifications
You must be signed in to change notification settings - Fork 235
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ray Serve Mistral LLM GPU Deploy Worker fails readiness check due to /home/ray/anaconda3/lib/libtinfo.so.6: no version information available (required by bash)
#661
Comments
Hi @calvinraveenthran , kubectl get pods -n rayserve-vllm Can you please share the KubeRay and Ray service logs? You can use the following command to get more details: |
Mistral-7B-Instruct-v0.2 is a gated model, can you confirm you have been granted access to the model in HF? |
I do have access to the model in HF. @shivam-dubey-1 will test this today and get back to you. |
When I retry the build I get another error:
|
Look at the #669 - pin |
This issue has been automatically marked as stale because it has been open 30 days |
Issue closed due to inactivity. |
Description
I am following this doc: https://awslabs.github.io/data-on-eks/docs/gen-ai/inference/GPUs/vLLM-rayserve
Once I run
I notice that gpu worker pod continually fails the readiness check.
Here are the events:
If your request is for a new feature, please use the
Feature request
template.Before you submit an issue, please perform the following for Terraform examples:
.terraform
directory (! ONLY if state is stored remotely, which hopefully you are following that best practice!):rm -rf .terraform/
terraform init
Versions
Module version [Required]:
Terraform version:
Reproduction Code [Required]
Steps to reproduce the behavior:
Expected behavior
Worker pod should be running.
Actual behavior
Worker pod fails readiness check.
Terminal Output Screenshot(s)
Additional context
The text was updated successfully, but these errors were encountered: