Skip to main content
Ollama lets you run open-source models locally. profClaw connects to Ollama automatically - it’s always enabled and is the fallback when no cloud API keys are configured.

Supported Models

Ollama supports hundreds of models. Popular choices in profClaw:
AliasModelBest For
local / llamallama3.2General purpose
deepseek-localdeepseek-r1:7bReasoning tasks
qwenqwen2.5:14bMultilingual
mistral-localmistral:7bFast inference
Any model available in ollama list can be used by its full name.
Most local Ollama models do not support native tool calling. profClaw automatically falls back to manual tool prompting for these models.

Setup

1

Install Ollama

# macOS
brew install ollama

# Linux
curl -fsSL https://ollama.com/install.sh | sh

# Windows: download from https://ollama.com
2

Pull a model

ollama pull llama3.2
ollama pull deepseek-r1:7b
ollama pull qwen2.5:14b
3

Start the Ollama server

ollama serve
# Runs at http://localhost:11434 by default
4

profClaw connects automatically

No API key needed. profClaw connects to Ollama at http://localhost:11434.
profclaw doctor --provider ollama

Environment Variables

OLLAMA_BASE_URL
string
Ollama server URL. Defaults to http://localhost:11434. Override for remote Ollama instances.

Configuration Example

OLLAMA_BASE_URL=http://192.168.1.100:11434

Model Aliases

AliasModel
localllama3.2
llamallama3.2
deepseek-localdeepseek-r1:7b
qwenqwen2.5:14b
mistral-localmistral:7b

Usage Examples

# Use default local model
profclaw chat --model local "What is dependency injection?"

# Use a specific model by name
profclaw chat --model llama3.2 "Explain this code"

# Use any installed model
profclaw chat --model codellama:13b "Write a binary search"

Notes

  • Ollama is always lowest priority in auto-selection. If any cloud key is set, it takes precedence.
  • Local models work without internet access - useful for air-gapped environments.
  • GPU acceleration significantly improves performance. Ollama auto-detects CUDA/Metal.
  • Tool calling is available via manual prompting fallback for models that don’t support it natively.