Skip to content

Commit bb1017e

Browse files
SBrandeismerveenoyanWauplinhanouticelinaVaibhavs10
authored
[Inference Providers] Featherless release blogpost (#2883)
* blogpost: featherless as a provider * Update inference-providers-feeatherless.md Co-authored-by: Merve Noyan <[email protected]> * Update inference-providers-feeatherless.md Co-authored-by: Lucain <[email protected]> * Update inference-providers-feeatherless.md Co-authored-by: Lucain <[email protected]> * Changes from code review Co-authored-by: Wauplin <[email protected]> Co-authored-by: hanouticelina <[email protected]> * add Lucain as author * add to _blog.yml * add inference tag * banner / thumbnail * fix: _blog + spelling * change thumbnail * vb' wording * Apply suggestions from code review Co-authored-by: burtenshaw <[email protected]> * suggestions from wauplin * Apply suggestions from code review * New inference-providers tag * Update inference-providers-featherless.md Co-authored-by: Pedro Cuenca <[email protected]> --------- Co-authored-by: Merve Noyan <[email protected]> Co-authored-by: Lucain <[email protected]> Co-authored-by: Wauplin <[email protected]> Co-authored-by: hanouticelina <[email protected]> Co-authored-by: vb <[email protected]> Co-authored-by: burtenshaw <[email protected]> Co-authored-by: Pedro Cuenca <[email protected]>
1 parent b9dfd89 commit bb1017e

File tree

4 files changed

+147
-1
lines changed

4 files changed

+147
-1
lines changed

_blog.yml

Lines changed: 18 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5417,6 +5417,7 @@
54175417
tags:
54185418
- announcement
54195419
- hub
5420+
- inference-providers
54205421

54215422
- local: deepseek-r1-aws
54225423
title: "How to deploy and fine-tune DeepSeek models on AWS"
@@ -5540,6 +5541,7 @@
55405541
tags:
55415542
- announcement
55425543
- hub
5544+
- inference-providers
55435545

55445546
- local: inference-providers-nebius-novita-hyperbolic
55455547
title: "Introducing Three New Serverless Inference Providers: Hyperbolic, Nebius AI Studio, and Novita 🔥"
@@ -5549,6 +5551,7 @@
55495551
tags:
55505552
- announcement
55515553
- hub
5554+
- inference-providers
55525555

55535556
- local: paligemma2mix
55545557
title: "PaliGemma 2 Mix - New Instruction Vision Language Models by Google"
@@ -5881,7 +5884,7 @@
58815884
date: April 16, 2025
58825885
tags:
58835886
- llms
5884-
- inference
5887+
- inference-providers
58855888
- cohere
58865889
- enterprise
58875890
- partnerships
@@ -6148,3 +6151,17 @@
61486151
- training
61496152
- partnerships
61506153
- announcement
6154+
6155+
- local: inference-providers-featherless
6156+
title: "Featherless AI on Hugging Face Inference Providers 🔥"
6157+
author: sbrandeis
6158+
thumbnail: /blog/assets/inference-providers/welcome-featherless.jpg
6159+
date: Jun 12, 2025
6160+
tags:
6161+
- llms
6162+
- inference-providers
6163+
- featherless
6164+
- recursal
6165+
- enterprise
6166+
- partnerships
6167+
- hub

_tags.yml

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -45,3 +45,6 @@
4545

4646
- value: lerobot
4747
label: LeRobot
48+
49+
- value: inference-providers
50+
label: Inference Providers
Loading

inference-providers-featherless.md

Lines changed: 126 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,126 @@
1+
---
2+
title: "Featherless AI on Hugging Face Inference Providers 🔥"
3+
thumbnail: /blog/assets/inference-providers/welcome-featherless.jpg
4+
authors:
5+
- user: wxgeorge
6+
guest: true
7+
org: featherless-ai
8+
- user: pohnean-recursal
9+
guest: true
10+
org: featherless-ai
11+
- user: picocreator
12+
guest: true
13+
org: featherless-ai
14+
- user: celinah
15+
- user: Wauplin
16+
- user: sbrandeis
17+
---
18+
19+
![banner image](https://huggingface.co/blog/assets/inference-providers/welcome-featherless.jpg)
20+
21+
# Featherless AI on Hugging Face Inference Providers 🔥
22+
23+
We're thrilled to share that **Featherless AI** is now a supported Inference Provider on the Hugging Face Hub!
24+
Featherless AI joins our growing ecosystem, enhancing the breadth and capabilities of serverless inference directly on the Hub’s model pages. Inference Providers are also seamlessly integrated into our client SDKs (for both JS and Python), making it super easy to use a wide variety of models with your preferred providers.
25+
26+
[Featherless AI](https://featherless.ai) supports a wide variety of text and conversational models, including the latest open-source models from DeepSeek, Meta, Google, Qwen, and much more.
27+
28+
Featherless AI is a serverless AI inference provider with unique model loading and GPU orchestration abilities that makes an exceptionally large catalog of models available for users. Providers often offer either a low cost of access to a limited set of models, or an unlimited range of models with users managing servers and the associated costs of operation. Featherless provides the best of both worlds offering unmatched model range and variety but with serverless pricing. Find the full list of supported models on the [models page](https://huggingface.co/models?inference_provider=featherless-ai&sort=trending).
29+
30+
We're super excited to see what you'll build with this new provider!
31+
32+
Read more about how to use Featherless as an Inference Provider in its dedicated [documentation page](https://huggingface.co/docs/inference-providers/providers/featherless-ai).
33+
34+
## How it works
35+
36+
### In the website UI
37+
38+
39+
1. In your user account settings, you are able to:
40+
- Set your own API keys for the providers you’ve signed up with. If no custom key is set, your requests will be routed through HF. Learn more about request types in the [docs](https://huggingface.co/docs/inference-providers/en/pricing#routed-requests-vs-direct-calls).
41+
- Order providers by preference. This applies to the widget and code snippets in the model pages.
42+
43+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/inference-providers/user-settings-updated.png" alt="Inference Providers"/>
44+
45+
46+
2. As [mentioned](https://huggingface.co/docs/inference-providers/en/pricing), there are two modes when calling Inference Providers:
47+
- Custom key (calls go directly to the inference provider, using your own API key of the corresponding inference provider)
48+
- Routed by HF (in that case, you don't need a token from the provider, and the charges are applied directly to your HF account rather than the provider's account)
49+
50+
51+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/inference-providers/explainer.png" alt="Inference Providers"/>
52+
53+
54+
3. Model pages showcase third-party inference providers (the ones that are compatible with the current model, sorted by user preference)
55+
56+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/inference-providers/model-widget-updated.png" alt="Inference Providers"/>
57+
58+
59+
### From the client SDKs
60+
61+
#### from Python, using huggingface_hub
62+
63+
The following example shows how to use DeepSeek-R1 using Featherless AI as the inference provider. You can use a [Hugging Face token](https://huggingface.co/settings/tokens) for automatic routing through Hugging Face, or your own Featherless AI API key if you have one.
64+
65+
Install or upgrade `huggingface_hub` to ensure you have version v0.33.0 or better: `pip install --upgrade huggingface-hub`
66+
67+
```python
68+
import os
69+
from huggingface_hub import InferenceClient
70+
71+
client = InferenceClient(
72+
provider="featherless-ai",
73+
api_key=os.environ["HF_TOKEN"]
74+
)
75+
76+
messages = [
77+
{
78+
"role": "user",
79+
"content": "What is the capital of France?"
80+
}
81+
]
82+
83+
completion = client.chat.completions.create(
84+
model="deepseek-ai/DeepSeek-R1-0528",
85+
messages=messages,
86+
)
87+
88+
print(completion.choices[0].message)
89+
```
90+
91+
#### from JS using @huggingface/inference
92+
93+
```js
94+
import { InferenceClient } from "@huggingface/inference";
95+
96+
const client = new InferenceClient(process.env.HF_TOKEN);
97+
98+
const chatCompletion = await client.chatCompletion({
99+
model: "deepseek-ai/DeepSeek-R1-0528",
100+
messages: [
101+
{
102+
role: "user",
103+
content: "What is the capital of France?"
104+
}
105+
],
106+
provider: "featherless-ai",
107+
});
108+
109+
console.log(chatCompletion.choices[0].message);
110+
```
111+
112+
## Billing
113+
114+
For direct requests, i.e. when you use the key from an inference provider, you are billed by the corresponding provider. For instance, if you use a Featherless AI API key you're billed on your Featherless AI account.
115+
116+
For routed requests, i.e. when you authenticate via the Hugging Face Hub, you'll only pay the standard provider API rates. There's no additional markup from us, we just pass through the provider costs directly. (In the future, we may establish revenue-sharing agreements with our provider partners.)
117+
118+
**Important Note** ‼️ PRO users get $2 worth of Inference credits every month. You can use them across providers. 🔥
119+
120+
Subscribe to the [Hugging Face PRO plan](https://hf.co/subscribe/pro) to get access to Inference credits, ZeroGPU, Spaces Dev Mode, 20x higher limits, and more.
121+
122+
We also provide free inference with a small quota for our signed-in free users, but please upgrade to PRO if you can!
123+
124+
## Feedback and next steps
125+
126+
We would love to get your feedback! Share your thoughts and/or comments here: https://huggingface.co/spaces/huggingface/HuggingDiscussions/discussions/49

0 commit comments

Comments
 (0)