Your terminal.
Your models.
Your rules.
Open-source AI coding agent built in Rust. Switch between Claude, GPT-4o, DeepSeek, Gemini, or any local Ollama model with one flag — 23 tools, full permission controls, MCP server built in.
curl -sSfL https://cuervo.cloud/install.sh | sh AI that works the way you do
Halcon CLI was built out of frustration with AI assistants that only chat. We built an agent that actually runs code, edits files, commits to git, searches your codebase — and does it safely, with your permission.
It runs entirely on your machine. No cloud intermediary, no data sent anywhere except directly to the AI provider of your choice. Open-source under Apache 2.0 — inspect every line, modify anything, contribute back.
Every feature was designed by developers who use it daily. If it doesn't feel right in the terminal, it doesn't ship.
2,226 tests, atomic file writes, circuit breakers, tool loop guards — because your codebase deserves better than a chatbot wrapper.
Every tool call, every permission decision, every model invocation is logged. Know exactly what Cuervo did and why.
See what Cuervo can do
From simple questions to complex multi-step agentic tasks — from your terminal.
items.iter()
.filter(|i| i.score > 0.7)
.map(ProcessedItem::from)
.collect()
Everything you need.
Nothing you don't.
Built for developers who want a serious AI coding partner — not a chatbot wrapper.
Multi-Model Intelligence
Switch between Claude, GPT-4, DeepSeek, Gemini, and local Ollama — one flag, no reconfiguration. Automatic fallback and latency-aware routing built in.
23 Built-in Tools
File ops, bash, git, grep, web search, directory tree, symbol search, background tasks — all with three-tier permission controls and human-in-the-loop authorization.
TUI Cockpit
Full-featured terminal UI with live activity zone, side panel for plan/metrics/context, real-time token tracking, session browser, and clickable stop button.
Permission Controls
ReadOnly / ReadWrite / Destructive permission model. Every sensitive tool requires user consent — with deny-always and non-interactive modes for CI/automation.
MCP Server
Exposes all 23 tools as an MCP server for IDE integration. Works with Cursor, VS Code, and any MCP-compatible client over stdio JSON-RPC — zero config.
Self-Correcting Agent
Detects when it's stuck before you do. Bayesian anomaly detection + ARIMA resource forecasting + reflexion loop — the agent corrects itself, so you don't have to restart.
Works with every major AI provider
Switch with a single flag — same interface, every model.
Any OpenAI-compatible endpoint works out of the box —
--provider custom --base-url https://...
Production-grade
agent architecture
Cuervo runs a multi-round agent loop with FSM state tracking, parallel tool batching, and a 5-tier context memory pipeline (L0–L4). Not a thin API wrapper.
Perceptual color science,
live in your browser
Halcón's UI system is powered by momoto-ui-core — a Rust/WASM engine that derives perceptually accurate UI state tokens from OKLCH color science. WCAG 2.1 + APCA validated in real time.
Brand Palette
Derived State Tokens
A11y Validation
Built in the open.
For everyone.
Apache 2.0 licensed. Every feature is public, every bug is trackable, every contribution is welcome.
Every command at a glance
12 top-level commands. All with --help for full options.
halcon chat "<prompt>" Start AI-powered agentic chat halcon chat --tui Launch the TUI cockpit interface halcon -p deepseek chat Use a specific AI provider halcon auth login anthropic Store API key in system keychain halcon auth status Show all configured providers halcon tools list Show all 23 available tools halcon tools validate Validate tool configurations halcon tools add <name> Add a custom tool from manifest halcon chat --full Enable orchestration, reflexion, tasks halcon memory search "<query>" Search across all session memory halcon update Self-update to latest release halcon mcp-server Start as MCP server for IDEs halcon doctor Check system health and config halcon cuervo <command>lt;commandcuervo <command>gt; --help
for full options. Full documentation ↗ --provider <name> Override AI provider (anthropic, openai, deepseek, ollama, gemini) --model <name> Override model for this session --tui Launch the TUI cockpit interface --no-tools Disable all tool use (chat only) --expert Enable verbose expert mode output --dry-run Preview actions without executing Up and running in 60 seconds
Three commands. That's all it takes.
curl -sSfL https://cuervo.cloud/install.sh | sh halcon auth login anthropic halcon chat --tui
Ready to ship
faster?
Open source. Self-hostable. No cloud required.
Just you, your terminal, and the best AI models.