Troubleshooting
model not found With Ollama
Common error:
model 'gemma4:e4b' not found (status code: 404)
Check:
ollama list
The model name must match the name shown by Ollama exactly.
OLLAMA_API_KEY Is Set, But I Want Local Ollama
Use ollama:<model>. The harness forces http://localhost:11434 by default:
agno-deep-agent --model ollama:gemma4:e4b "Say hello"
For cloud, use ollama-cloud:<model>.
Ollama Runs On Another Host
Pass --ollama-host:
agno-deep-agent --model ollama:gemma4:e4b \
--ollama-host http://localhost:11434 \
"Say hello"
OPENAI_API_KEY is not set
This warning appears when the provider is OpenAI:
export OPENAI_API_KEY="your-key-here"
Or use local Ollama:
agno-deep-agent --model ollama:gemma4:e4b "Say hello"
The Agent Tried To Run A Blocked Command
CodingTools blocks commands outside the allowlist and dangerous shell
operators in restricted mode. For a more conservative run, disable shell:
agno-deep-agent --no-shell "Analyze the files without running commands"
Memory Does Not Seem To Persist
Confirm that you are using the same --db-file, --user-id, and --session-id.
Example:
agno-deep-agent --user-id dev@example.com --session-id local-project \
"Remember that I prefer concise answers"
agno-deep-agent --user-id dev@example.com --session-id local-project \
"Continue the session"
GitHub Pages Did Not Update
Check:
Settings > Pagespoints to branchmainand folder/docs.docs/index.mdexists.- The latest deploy in
Actionscompleted successfully. - Your browser cache is not showing an old version.