Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[RFE] Add openai text generation API compatibility layer in kserve/TGIS #226

Open
ashishkamra opened this issue Feb 19, 2024 · 2 comments
Open

Comments

@ashishkamra
Copy link

Request
The ask is to introduce a openai text generation API compatibility layer (chat completion endpoint) to kserve/TGIS.

Why
Having an openai API compatibility layer will allow more open source tools such as 'https://github.com/EleutherAI/lm-evaluation-harness' to interoperate with our model serving stack

Suggested Implemention
Use litellm openai proxy server - https://litellm.vercel.app/docs/providers/huggingface

@danielezonca
Copy link
Contributor

One idea to explore is to use a similar proxy as KServe transformer around TGIS image so that we can easily plug it in the ServingRuntime.
The existing HF provided will not probably work OOTB with TGIS but the work to adapt should not be too much.

@Xaenalt @dtrifiro
Wdyt? (I'm just thinking for now at a community/research spike, nothing long term/supported)

@danielezonca
Copy link
Contributor

FYI KServe project itself is considering to add OpenAI API support for chat completition as part of open-inference-protocol spec

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants