AI Assistant Model Configuration Guide
Quick Start
The AI Assistant supports multiple AI models via PydanticAI’s multi-provider support. Configure models using environment variables in your .env file.
Configuration File Location
Place your .env file in the repository root (same location as .env.example):
cd /path/to/qubinode_navigator
cp .env.example .env
# Edit .env with your preferred models and API keys
Model Configuration Variables
Core Model Variables
# Manager Model (orchestration & planning)
MANAGER_MODEL=google-gla:gemini-2.0-flash
# Developer Model (code generation & execution)
DEVELOPER_MODEL=google-gla:gemini-2.0-flash
# Deployment Model (infrastructure provisioning)
PYDANTICAI_MODEL=google-gla:gemini-2.0-flash
Model Format
PydanticAI uses provider:model format (note the colon, not slash):
# ✅ Correct
MANAGER_MODEL=google-gla:gemini-2.0-flash
MANAGER_MODEL=openrouter:anthropic/claude-3.5-sonnet
MANAGER_MODEL=openai:gpt-4o
# ❌ Incorrect
MANAGER_MODEL=google-gla/gemini-2.0-flash # Wrong separator
MANAGER_MODEL=gemini/gemini-2.0-flash # LiteLLM format, not PydanticAI
Supported Providers
1. Google Gemini (Recommended - Fast & Cheap)
MANAGER_MODEL=google-gla:gemini-2.0-flash
GEMINI_API_KEY=your-api-key
Get API key: https://makersuite.google.com/app/apikey
Available Models:
google-gla:gemini-2.0-flash- Fast, recommended for most tasksgoogle-gla:gemini-2.5-flash- Latest versiongoogle-gla:gemini-1.5-pro- More capable, higher latency
2. OpenRouter (100+ Models via Single API)
MANAGER_MODEL=openrouter:anthropic/claude-3.5-sonnet
DEVELOPER_MODEL=openrouter:google/gemini-2.0-flash-exp
PYDANTICAI_MODEL=openrouter:openai/gpt-4o
OPENROUTER_API_KEY=sk-or-your-api-key
Get API key: https://openrouter.ai/keys
Popular Models:
openrouter:anthropic/claude-3.5-sonnet- Claude via OpenRouteropenrouter:google/gemini-2.0-flash-exp- Gemini via OpenRouteropenrouter:openai/gpt-4o- GPT-4o via OpenRouteropenrouter:meta-llama/llama-3.3-70b-instruct- Llama 3.3 70Bopenrouter:deepseek/deepseek-chat- DeepSeek Chat
3. Anthropic Claude
MANAGER_MODEL=anthropic:claude-3-5-sonnet-latest
ANTHROPIC_API_KEY=sk-ant-your-api-key
Get API key: https://console.anthropic.com/
Available Models:
anthropic:claude-3-5-sonnet-latest- Claude 3.5 Sonnetanthropic:claude-3-haiku-20240307- Fast & cheapanthropic:claude-3-opus-latest- Most capable
4. OpenAI
MANAGER_MODEL=openai:gpt-4o
OPENAI_API_KEY=sk-your-api-key
Available Models:
openai:gpt-4o- GPT-4oopenai:gpt-4-turbo- GPT-4 Turbo
5. Local Models (Ollama)
MANAGER_MODEL=ollama:granite3.3:8b
OLLAMA_BASE_URL=http://localhost:11434
Setup: https://ollama.com/download
Popular Models:
ollama:llama3.2:latest- Meta Llama 3.2ollama:granite3.3:8b- IBM Granite 3.3 8Bollama:mistral:7b- Mistral 7B
6. Groq (Fast Inference)
MANAGER_MODEL=groq:llama-3.3-70b-versatile
GROQ_API_KEY=gsk-your-api-key
Configuration Examples
Example 1: Google Gemini (Default - Fast & Free)
# .env
MANAGER_MODEL=google-gla:gemini-2.0-flash
DEVELOPER_MODEL=google-gla:gemini-2.0-flash
PYDANTICAI_MODEL=google-gla:gemini-2.0-flash
GEMINI_API_KEY=your-gemini-api-key
Example 2: OpenRouter with Mixed Models
# .env
MANAGER_MODEL=openrouter:anthropic/claude-3.5-sonnet
DEVELOPER_MODEL=openrouter:google/gemini-2.0-flash-exp
PYDANTICAI_MODEL=openrouter:openai/gpt-4o
OPENROUTER_API_KEY=sk-or-your-api-key
Example 3: Anthropic Claude for All Tasks
# .env
MANAGER_MODEL=anthropic:claude-3-5-sonnet-latest
DEVELOPER_MODEL=anthropic:claude-3-haiku-20240307
PYDANTICAI_MODEL=anthropic:claude-3-5-sonnet-latest
ANTHROPIC_API_KEY=sk-ant-your-api-key
Example 4: Local Ollama (No API Key Required)
# .env
MANAGER_MODEL=ollama:granite3.3:8b
DEVELOPER_MODEL=ollama:llama3.2
PYDANTICAI_MODEL=ollama:granite3.3:8b
OLLAMA_BASE_URL=http://localhost:11434
Deployment
After configuring your .env file, deploy with:
./scripts/development/deploy-qubinode.sh
The script will:
- Load your
.envfile from the repository root - Validate model configuration
- Log the models being used
- Start the AI Assistant container with your configuration
Verification
Check that your models are loaded correctly:
# During deployment, look for:
[INFO] AI Model Configuration:
[INFO] Manager Model: openrouter:anthropic/claude-3.5-sonnet
[INFO] Developer Model: openrouter:google/gemini-2.0-flash-exp
[INFO] Deployment Model: openrouter:openai/gpt-4o
[INFO] OpenRouter API Key: [SET]
# After deployment, verify via API:
curl http://localhost:8080/orchestrator/status
Expected response:
{
"status": "ready",
"models": {
"manager": "openrouter:anthropic/claude-3.5-sonnet",
"developer": "openrouter:google/gemini-2.0-flash-exp",
"deployment": "openrouter:openai/gpt-4o"
},
"api_keys": {
"openrouter": true
}
}
Troubleshooting
Issue: Container uses default models instead of .env configuration
Solution: Ensure .env file is in the repository root:
# Check .env location
ls -la /path/to/qubinode_navigator/.env
# If missing, copy from example
cp .env.example .env
Issue: Model format error
Error: Invalid model format: google-gla/gemini-2.0-flash
Solution: Use colon separator (PydanticAI format), not slash:
# ✅ Correct
MANAGER_MODEL=google-gla:gemini-2.0-flash
# ❌ Wrong
MANAGER_MODEL=google-gla/gemini-2.0-flash
Issue: API key not being used
Solution: Verify API key is exported:
# Check if API key is loaded
./scripts/development/deploy-qubinode.sh 2>&1 | grep "API Key"
# Should see:
# OpenRouter API Key: [SET]
Issue: Container fails to start
Solution: Check container logs:
podman logs qubinode-ai-assistant
Common issues:
- Invalid API key format
- Network connectivity issues
- Missing required environment variables
Advanced Configuration
Using Different Models for Different Agents
# Smart manager, fast worker
MANAGER_MODEL=anthropic:claude-3-5-sonnet-latest # Planning
DEVELOPER_MODEL=google-gla:gemini-2.0-flash # Execution
PYDANTICAI_MODEL=openrouter:meta-llama/llama-3.3-70b-instruct
PYDANTICAI_MODEL Default Behavior
If PYDANTICAI_MODEL is not set, it defaults to MANAGER_MODEL:
# Only set MANAGER_MODEL
MANAGER_MODEL=anthropic:claude-3-5-sonnet-latest
# PYDANTICAI_MODEL automatically becomes:
# anthropic:claude-3-5-sonnet-latest
Reference
.env.example- Full configuration template with all options- PydanticAI Docs: https://ai.pydantic.dev/models/
- OpenRouter Models: https://openrouter.ai/models
- ADR-0049: PydanticAI Integration
- ADR-0063: Multi-Agent Architecture