Supported Models
| Model | ID | Context | Max Output | Tools | Input $/1M | Output $/1M |
|---|---|---|---|---|---|---|
| Llama 3.3 70B | llama-3.3-70b-versatile | 128K | 32K | Yes | $0.59 | $0.79 |
| Llama 3.1 8B Instant | llama-3.1-8b-instant | 128K | 8K | Yes | $0.05 | $0.08 |
| Mixtral 8x7B | mixtral-8x7b-32768 | 32K | 8K | Yes | $0.24 | $0.24 |
Setup
Get an API key
Sign up at console.groq.com. Free tier available.
Environment Variables
Your Groq API key. Format:
gsk_...Configuration Example
- .env
- settings.yml
Model Aliases
| Alias | Model |
|---|---|
groq | llama-3.3-70b-versatile |
groq-fast | llama-3.1-8b-instant |
groq-mixtral | mixtral-8x7b-32768 |
Usage Examples
Notes
- Groq is ranked 5th in auto-selection priority after Anthropic, OpenAI, Azure, and Google.
llama-3.1-8b-instantis one of the cheapest available models at $0.05/1M input tokens.- Groq has a generous free tier with rate limits per day.
- API is OpenAI-compatible - endpoint:
https://api.groq.com/openai/v1
Related
- AI Providers Overview - Compare all 37 supported providers
- Cerebras - Wafer-scale inference for even faster token speeds
- Together AI - Hundreds of open-source models via one API
- profclaw provider - Add and test providers from the CLI