[FEEDBACK] Inference Providers

#49
by julien-c - opened
Hugging Face org

Any inference provider you love, and that you'd like to be able to access directly from the Hub?

Hugging Face org
โ€ข
edited Jan 28

Love that I can call DeepSeek R1 directly from the Hub ๐Ÿ”ฅ

from huggingface_hub import InferenceClient

client = InferenceClient(
    provider="together",
    api_key="xxxxxxxxxxxxxxxxxxxxxxxx"
)

messages = [
    {
        "role": "user",
        "content": "What is the capital of France?"
    }
]

completion = client.chat.completions.create(
    model="deepseek-ai/DeepSeek-R1", 
    messages=messages, 
    max_tokens=500
)

print(completion.choices[0].message)

Is it possible to set a monthly payment budget or rate limits for all the external providers? I don't see such options in billings tab. In case a key is or session token is stolen, it can be quite dangerous to my thin wallet:(

Hugging Face org

@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future

@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future

Thanks for your quick reply, good to know!

Would be great if you could add Nebius AI Studio to the list :) New inference provider on the market, with the absolute cheapest prices and the highest rate limits...

Could be good to add featherless.ai

TitanML !!

Since yesterday, the ZeroGPU version of SD3.5 Large Turbo freezes out with ERROR (All the time...) Please Support, fix this problem as soon as you can...

Hugging Face org

@gulyasdavid1999 this is unrelated to Inference Providers though

I get a token 403 unexpected error: 403 Forbidden: This authentication method does not have sufficient permissions to call Inference Providers on behalf of user XXX

Screenshot 2025-03-24 103849.png

Screenshot 2025-03-24 104202.png

Code to reproduce :

from huggingface_hub import InferenceClient
client = InferenceClient()
client.list_deployed_models("text-generation-inference")

question around enterprise accounts:

  • each user gets a $2 quota?
  • in order to apply the quota they have to use a key from their ent account and the X-HF-Bill-To header?
  • how much does a call cost? I can't find any documentation...

Would be great to add Clarifai to the list. The platform is vendor-agnostic, supporting AWS, GCP, Vultr, and Oracle. We are planning to add a lot more providers soon.

Our company wants to provide some private models.
Is it possible in Model Mapping [ https://huggingface.co/docs/inference-providers/en/register-as-a-provider#3-model-mapping-api ] to have hfModel as a "stub" only and providerModel as a real model?

Just signed up with HF and had some questions for the general community to help us get started. We plan to use the Cerebras Inference Provider using direct calls rather than routing through HF itself.

With a Pro subscription, are there any limits to token usage or queuing constraints when using a custom API key and direct calls? The free tier on Cerebras did have such constraints.

Thanks in advance

Hey all, I'd like to make nCompass (https://docs.ncompass.tech/api-reference/quickstart) an inference provider on HF. We build GPU optimizations to be able to support an API without rate limits by maximizing GPU utilization. I would really appreciate it if someone could help us with the process of becoming an inference provider.

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment