The terminal AI agent ▋
Halcón connects your terminal to Claude, GPT-4o, DeepSeek, Gemini, or any local Ollama model — with 21 built-in tools, layered permissions, and a native MCP server for your IDE.
SHA-256 verified · no sudo · installs to ~/.local/bin
Multi-model
Works with every major AI provider
Switch providers and models with a single flag. No reconfiguration, no friction.
Models
Models
Models
Models
Models
Switch with one flag
halcon -p openai chat
halcon -p ollama -m llama3 chat
Capabilities
Everything you need to automate your workflow
Production-grade with 2,200+ test coverage. Multi-model. Zero native dependencies.
21 built-in tools
File read/write/edit, bash, git, web search, regex grep, code symbols, HTTP requests, and background jobs — all wired into the agent loop.
Permission-first
Every tool is classified. ReadOnly tools run silently. ReadWrite and Destructive tools require explicit confirmation.
TUI Cockpit
3-zone terminal UI: live token counters, agent FSM state, plan progress, and real-time activity stream. Pause, step, or cancel mid-execution.
halcon chat --tui Episodic memory
Remembers decisions, file paths, and learnings across sessions using BM25 semantic search and temporal decay scoring.
MCP Server built in
Run halcon mcp-server to expose all 21 tools via JSON-RPC over stdio. Wire into VS Code, Cursor, or any MCP-compatible IDE.
Multi-provider routing
Automatic fallback between providers, latency-aware model selection, and cost tracking per session. Set balanced, fast, or cheap strategies.
p95 latency · real session data · automatic routing
Full situational
awareness.
The 3-zone TUI gives you live token counters, real-time agent FSM state, execution plan progress, side panel metrics, and a scrollable activity stream — all updating as the agent works.
- ✓ Pause, step, or cancel the agent mid-execution
- ✓ Live token cost tracked per round
- ✓ Tool execution timeline with durations
- ✓ Provider fallback indicators
- ✓ Context compression events in real time
halcon chat --tui
Plan
Tokens
In action
One agent, every workflow
From quick questions to complex multi-step engineering tasks.
get_posts_by_user() loads all posts then fetches author per row. list_comments() makes one query per post. Use JOIN or batch WHERE id IN (?) instead. src/scraper.rs — async scraper using reqwest, tokio, concurrent requests via JoinSet, exponential backoff, and configurable rate limiting. Get started
Up and running in 60 seconds
Install Halcón CLI
Installs to ~/.local/bin/halcon. SHA-256 verified.
Configure your API key
Stores your key securely in the OS keychain.
Start coding with AI
Or launch the full TUI cockpit with --tui
Supports Anthropic, OpenAI, DeepSeek, Gemini, and local Ollama.
Start shipping faster today.
One command install. No configuration. Every major AI provider.
macOS · Linux · Windows.