Inference Providers documentation
Featherless AI
Featherless AI
All supported Featherless AI models can be found here
Featherless AI is a serverless AI inference platform that offers access to thousands of open-source models.
Our goal is to make all AI models available for serverless inference. We provide inference via API to a continually expanding library of open-weight models.
Supported tasks
Chat Completion (LLM)
Find out more about Chat Completion (LLM) here.
Language
Client
Provider
Copied
import os from openai import OpenAI client = OpenAI( base_url="https://router.huggingface.co/v1", api_key=os.environ["HF_TOKEN"], ) completion = client.chat.completions.create( model="meta-llama/Llama-3.1-8B-Instruct:featherless-ai", messages=[ { "role": "user", "content": "What is the capital of France?" } ], ) print(completion.choices[0].message)Chat Completion (VLM)
Find out more about Chat Completion (VLM) here.
Language
Client
Provider
Copied
import os from openai import OpenAI client = OpenAI( base_url="https://router.huggingface.co/v1", api_key=os.environ["HF_TOKEN"], ) completion = client.chat.completions.create( model="google/gemma-3-27b-it:featherless-ai", messages=[ { "role": "user", "content": [ { "type": "text", "text": "Describe this image in one sentence." }, { "type": "image_url", "image_url": { "url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg" } } ] } ], ) print(completion.choices[0].message)Text Generation
Find out more about Text Generation here.
Language
Client
Provider
Copied
import os from openai import OpenAI client = OpenAI( base_url="https://router.huggingface.co/featherless-ai/v1/completions", api_key=os.environ["HF_TOKEN"], ) completion = client.chat.completions.create( model="meta-llama/Meta-Llama-3.1-8B-Instruct", messages="\"Can you please let us know more details about your \"", ) print(completion.choices[0].message)
