New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
HfAPI().create_inference_endpoint errors and does not follow documentation #2277
Comments
I see this too with 0.23.0. Copying and pasting the example in https://huggingface.co/blog/tgi-messages-api doesn't work.
I tried GCP too, same result. I used the parameters for the curl call in the IE page.
@philschmid any idea? |
The naming was adjust. pinging @co42 here. |
The naming here should be correct: https://huggingface.co/docs/inference-endpoints/pricing can you try |
The Google example works. My mistake was the vendor name: "gcp", not "google" :) The blog post example works when changed to:
|
Thanks everyone for reporting/fixing this! Just to be sure, is there still something to fix on |
Describe the bug
Using hf_api.create_inference_endpoint with configuration in documentation raises error.
https://huggingface.co/docs/huggingface_hub/en/package_reference/hf_api#huggingface_hub.HfApi.create_inference_endpoint.task
Instance types appear different in the available vendors too:
https://api.endpoints.huggingface.cloud/v2/provider
https://huggingface.co/docs/inference-endpoints/en/pricing
Terminology in vendor list does not match the API:
Eg.
instanceSize
"small" != "x1"Reproduction
Logs
System info
The text was updated successfully, but these errors were encountered: