[FEEDBACK] Inference Providers

#49
by julien-c HF staff - opened
Hugging Face org

Any inference provider you love, and that you'd like to be able to access directly from the Hub?

Hugging Face org
โ€ข
edited 2 days ago

Love that I can call DeepSeek R1 directly from the Hub ๐Ÿ”ฅ

from huggingface_hub import InferenceClient

client = InferenceClient(
    provider="together",
    api_key="xxxxxxxxxxxxxxxxxxxxxxxx"
)

messages = [
    {
        "role": "user",
        "content": "What is the capital of France?"
    }
]

completion = client.chat.completions.create(
    model="deepseek-ai/DeepSeek-R1", 
    messages=messages, 
    max_tokens=500
)

print(completion.choices[0].message)

Is it possible to set a monthly payment budget or rate limits for all the external providers? I don't see such options in billings tab. In case a key is or session token is stolen, it can be quite dangerous to my thin wallet:(

Hugging Face org

@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future

@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future

Thanks for your quick reply, good to know!

Would be great if you could add Nebius AI Studio to the list :) New inference provider on the market, with the absolute cheapest prices and the highest rate limits...

Could be good to add featherless.ai

TitanML !!

OpenRouter!

Hi everyone and first of all I really thank huggingface team for releasing this feature.
One another suggestion that I need is :
to give more granular parameters for the user to share the deployment configuration file (as an terraform hcl configuration) in order to define more optimum infrastructure for inference and let them define the inference model.

requesty.ai
groq

TypeError: InferenceClient.__init__() got an unexpected keyword argument 'provider'

Add RunPod as an inference provider!

Hyperbolic!

Runpod!

Hugging Face org

@llamameta please, make sure you use the latest version of huggingface_hub

pip install --upgrade huggingface_hub

Hello. Congratulations on great feature. How we can proceed on adding deepinfra.com to the providers list?

That's a great news, and if you provide more measurable details on what's free inference with a small quota is about for signed-in free users, you'll build more trust around the community, hence more PRO subscribers!

Runpod is excellent.
Let's add it to inference provider list

Runpod would be very welcome!

Yes, Runpod.io would be great indeed

This comment has been hidden

@julien-c

is there a way to add https://nineteen.ai/ as a provider ? it allows free access to top models like Deepseek R1 & the llama family
there is also : https://chutes.ai/ which can allow users to deploy any model on demand

Can we Add https://avian.io as a provider @julien-c

Currently have the fastest inference on Nvidia hardware, as well as highest throughput, allowing users to deploy any model on demand

The inference API is not working for HF Inference (โˆž requests).

My Settings:
Screenshot at Jan 30 09-36-54.png

The Issue:
Screenshot at Jan 30 09-37-55.png

ai/ml api - dudes have a lot of models from HF!

try Runware.

Runware plz โ€“ it's ~5x cheaper than other providers and still one of the fastest

Try Scaleway.

Sign up or log in to comment