Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: Support for SageMaker-required endpoints #11557

Open
1 task done
nathan-az opened this issue Dec 27, 2024 · 0 comments · May be fixed by #11576
Open
1 task done

[Feature]: Support for SageMaker-required endpoints #11557

nathan-az opened this issue Dec 27, 2024 · 0 comments · May be fixed by #11576

Comments

@nathan-az
Copy link

nathan-az commented Dec 27, 2024

🚀 The feature, motivation and pitch

This was discussed before and was not supported due to AWS needing to manage the images.

I'm wondering if there is interest in at least including routing sourcecode for the required SageMaker endpoints (/invocations and /ping) to the vLLM source.

The main benefit would be the standard openai vLLM image should be automatically compatible with SageMaker endpoints. Currently, interested users have to do so through LMI, or fork vLLM and add these.

If there is interest and support from vLLM maintainers, I'm happy to contribute this to the openai entrypoints:

  • a ping endpoint rerouting to /health
  • an invocations endpoint that routes to the expected existing endpoint (or with an additional parameter so the user dictates the target)

My understanding is that these are the only two requirements for SageMaker support.

Before submitting a new issue...

  • Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
1 participant