[FEEDBACK] Inference Providers

#49
by julien-c - opened
Hugging Face org

Any inference provider you love, and that you'd like to be able to access directly from the Hub?

Hugging Face org
edited Jan 28

Love that I can call DeepSeek R1 directly from the Hub 🔥

from huggingface_hub import InferenceClient

client = InferenceClient(
    provider="together",
    api_key="xxxxxxxxxxxxxxxxxxxxxxxx"
)

messages = [
    {
        "role": "user",
        "content": "What is the capital of France?"
    }
]

completion = client.chat.completions.create(
    model="deepseek-ai/DeepSeek-R1", 
    messages=messages, 
    max_tokens=500
)

print(completion.choices[0].message)

Is it possible to set a monthly payment budget or rate limits for all the external providers? I don't see such options in billings tab. In case a key is or session token is stolen, it can be quite dangerous to my thin wallet:(

Hugging Face org

@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future

@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future

Thanks for your quick reply, good to know!

Would be great if you could add Nebius AI Studio to the list :) New inference provider on the market, with the absolute cheapest prices and the highest rate limits...

Could be good to add featherless.ai

TitanML !!

I am sad, too 🥹💔💔

Hi @julien-c , we at Dat1.co want to be an inference provider and have submitted a PR for the initial step. Could anyone help us move forward? https://github.com/huggingface/huggingface.js/pull/1460

Please add FPT AI Inference http://marketplace.fptcloud.com/
It's a newcomer but already has hundreds of users, thanks to its speed, stability, and competitive price.

black-forest-labs/FLUX.1-schnell is not properly working under HFInference provider, is this going to be permanent or there is an issue?

Actually it seems that HFInference is not working at all, do we need to use models now only via external providers????

The message for any HFInference is: "Our latest automated health check on this model for this provider did not complete successfully." Is this temporary or HFInference won't process certain models anymore, or maybe it is a bug?

Hi Team,

I want to register as a Inference Providers, Can you please suggest us the way forward process.

Thanks
Cyfuture

Dear Hugging Face Team,

Greetings from Cyfuture AI!

We are reaching out to explore a potential collaboration with Hugging Face. As a rapidly growing enterprise-grade AI solutions provider, Cyfuture AI offers robust and scalable inference capabilities powered by high-performance GPU infrastructure.
We would be keen to join Hugging Face as an official inference provider to support model deployment and inference workloads for your global community. We believe this integration would bring mutual value—enhancing access to affordable, high-speed inference while expanding our reach within the AI ecosystem.
Please let us know the next steps or any prerequisites required to move forward with this partnership.
Looking forward to your response.
Regards,

Cyfuture.ai
Email us at - [email protected]

Sign up or log in to comment