[FEEDBACK] Inference Providers
Any inference provider you love, and that you'd like to be able to access directly from the Hub?
Love that I can call DeepSeek R1 directly from the Hub ๐ฅ
from huggingface_hub import InferenceClient
client = InferenceClient(
provider="together",
api_key="xxxxxxxxxxxxxxxxxxxxxxxx"
)
messages = [
{
"role": "user",
"content": "What is the capital of France?"
}
]
completion = client.chat.completions.create(
model="deepseek-ai/DeepSeek-R1",
messages=messages,
max_tokens=500
)
print(completion.choices[0].message)
Is it possible to set a monthly payment budget or rate limits for all the external providers? I don't see such options in billings tab. In case a key is or session token is stolen, it can be quite dangerous to my thin wallet:(
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
Thanks for your quick reply, good to know!
Would be great if you could add Nebius AI Studio to the list :) New inference provider on the market, with the absolute cheapest prices and the highest rate limits...
Could be good to add featherless.ai
TitanML !!
OpenRouter!
Hi everyone and first of all I really thank huggingface team for releasing this feature.
One another suggestion that I need is :
to give more granular parameters for the user to share the deployment configuration file (as an terraform hcl configuration) in order to define more optimum infrastructure for inference and let them define the inference model.
requesty.ai
groq
TypeError: InferenceClient.__init__()
got an unexpected keyword argument 'provider'
Add RunPod as an inference provider!
Hyperbolic!
Runpod!
@llamameta
please, make sure you use the latest version of huggingface_hub
pip install --upgrade huggingface_hub
Hello. Congratulations on great feature. How we can proceed on adding deepinfra.com to the providers list?
That's a great news, and if you provide more measurable details on what's free inference with a small quota
is about for signed-in free users, you'll build more trust around the community, hence more PRO subscribers!
requesty.ai !
Runpod is excellent.
Let's add it to inference provider list
runpod!
Runpod would be very welcome!
Yes, Runpod.io would be great indeed
is there a way to add https://nineteen.ai/ as a provider ? it allows free access to top models like Deepseek R1 & the llama family
there is also : https://chutes.ai/ which can allow users to deploy any model on demand
Can we Add https://avian.io as a provider @julien-c
Currently have the fastest inference on Nvidia hardware, as well as highest throughput, allowing users to deploy any model on demand
ai/ml api - dudes have a lot of models from HF!
try Runware.
Runware plz โ it's ~5x cheaper than other providers and still one of the fastest
Try Scaleway.