Inference Providers documentation
Featherless AI
Featherless AI
All supported Featherless AI models can be found here
Featherless AI is a serverless AI inference platform that offers access to thousands of open-source models.
Our goal is to make all AI models available for serverless inference. We provide inference via API to a continually expanding library of open-weight models.
Supported tasks
Chat Completion (LLM)
Find out more about Chat Completion (LLM) here.
Language
Client
Provider
Copied
import os
from openai import OpenAI
client = OpenAI(
base_url="/static-proxy?url=https%3A%2F%2Frouter.huggingface.co%2Fv1%26quot%3B%3C%2Fspan%3E%2C
api_key=os.environ["HF_TOKEN"],
)
completion = client.chat.completions.create(
model="inclusionAI/Ling-1T:featherless-ai",
messages=[
{
"role": "user",
"content": "What is the capital of France?"
}
],
)
print(completion.choices[0].message)
Chat Completion (VLM)
Find out more about Chat Completion (VLM) here.
Language
Client
Provider
Copied
import os
from openai import OpenAI
client = OpenAI(
base_url="/static-proxy?url=https%3A%2F%2Frouter.huggingface.co%2Fv1%26quot%3B%3C%2Fspan%3E%2C
api_key=os.environ["HF_TOKEN"],
)
completion = client.chat.completions.create(
model="google/gemma-3-27b-it:featherless-ai",
messages=[
{
"role": "user",
"content": [
{
"type": "text",
"text": "Describe this image in one sentence."
},
{
"type": "image_url",
"image_url": {
"url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg"
}
}
]
}
],
)
print(completion.choices[0].message)
Text Generation
Find out more about Text Generation here.
Language
Client
Provider
Copied
import os
from openai import OpenAI
client = OpenAI(
base_url="/static-proxy?url=https%3A%2F%2Frouter.huggingface.co%2Ffeatherless-ai%2Fv1%2Fcompletions%26quot%3B%3C%2Fspan%3E%2C
api_key=os.environ["HF_TOKEN"],
)
completion = client.chat.completions.create(
model="inclusionAI/Ling-1T",
messages="\"Can you please let us know more details about your \"",
)
print(completion.choices[0].message)