Supported Models
| Model | ID | Context | Notes |
|---|---|---|---|
| Llama 3.1 70B | accounts/fireworks/models/llama-v3p1-70b-instruct | 128K | Default |
| Llama 3.1 8B | accounts/fireworks/models/llama-v3p1-8b-instruct | 128K | Fast/cheap |
| Mixtral 8x22B | accounts/fireworks/models/mixtral-8x22b-instruct | 65K | Large MoE |
| DeepSeek R1 | accounts/fireworks/models/deepseek-r1 | 64K | Reasoning |
Setup
Get an API key
Sign up at fireworks.ai. Free trial credits available.
Environment Variables
Your Fireworks AI API key. Format:
fw_...Configuration Example
- .env
- settings.yml
Model Aliases
| Alias | Model |
|---|---|
fireworks | accounts/fireworks/models/llama-v3p1-70b-instruct |
Usage Examples
Notes
- API endpoint:
https://api.fireworks.ai/inference/v1(OpenAI-compatible) - Status: Beta
- Fireworks supports fine-tuned model deployment and dedicated instances.
- Model IDs use the
accounts/fireworks/models/prefix format.
Related
- AI Providers Overview - Compare all 37 supported providers
- Together AI - Similar open-source model hosting option
- Groq - Ultra-fast LPU inference for low-latency workloads
- profclaw provider - Add and test providers from the CLI