Overview
Mental Model
- Models = raw intelligence
- CLIs = how you talk to models
- Agent runtimes = how models think and act
- Orchestrators = how multiple things work together
Scope
Power-user tools for LLM orchestration.
Included:
- CLI / desktop tools
- multi-model / multi-provider
- agent / orchestration capable
Excluded:
- web apps
- libraries / frameworks
- SDKs
Core Tools
Orchestrators (Multi-agent coordination tools)
Coordinate multiple models, agents, or tools into workflows.
OpenCode
Auth:
- Usually API keys for model providers
- Can also use provider-specific login or OAuth plugins
- Remote MCP servers can use OAuth or bearer tokens
Features:
- CLI or desktop agent runner
- Supports multiple providers, such as OpenAI, Claude, Gemini, etc.
- Acts as orchestrator, not just chat
Good for:
- multi-model workflows
- agent execution
- coding and automation
- chaining tools
Common in power-user setups.
AgentPipe
Auth:
- No single auth model
- Uses each underlying CLI's auth
- OpenRouter uses API key
- AgentPipe Web bridge uses API key
Features:
- Multi-agent CLI orchestrator
- Runs multiple AI CLIs together
- Supports multiple providers
- Orchestrates conversations between multiple AI agents like Claude, Gemini, Ollama, Qwen, Cursor in one shared session
Use cases:
- run Claude + local model + Gemini together
- compare outputs
- route tasks
- multi-agent workflows
Acts as orchestration shell.
Maestro
Auth:
- Inherits auth from connected CLIs and providers
- API keys for providers
- Optional service tokens for integrations
Features:
- Workflow orchestrator for agents and tools
- Declarative pipelines for multi-step tasks
- Scheduling and retries
- Observability hooks for runs
Good for:
- repeatable pipelines
- productionizing agent workflows
- chaining tools with control over flow
- monitoring and debugging runs
Often used when moving from experiments to repeatable workflows.
Agent Runtimes (Autonomous agent tools)
Run agent logic like planning, memory, and tool usage.
deepagents
Auth:
- Uses each provider's API key
Features:
- LangChain or LangGraph agent harness
- Planning + sub-agents + filesystem tools
- Designed for complex multi-step agents
- Can run as CLI or SDK
Good for:
- advanced agent workflows
- research or coding agents
- custom orchestration
- building your own agent runtime
Used in advanced agent stacks.
OpenClaw
Auth:
- Provider API keys
- Native Ollama support for local models
- Local model endpoints optional
Features:
- Agent runtime focused on autonomy
- Tool use with iterative planning loops
- Memory and context management
- Designed for long-running tasks
- Native Ollama integration with streaming and tool calling
Good for:
- autonomous agents
- long multi-step tasks
- experiments with self-directed workflows
- combining tools with planning loops
- local-model setups through Ollama
Often explored in experimental or research-heavy setups.
Ollama fit:
- first-class support
- uses Ollama's native API, not just OpenAI-compatible mode
- strong option if you want local models as the main serving layer
NanoClaw
Auth:
- Claude is the main orchestrator model in the documented Ollama setup
- Uses the auth of the underlying agent or model setup
- Messaging platform setup required for connected channels
- Local runtime and container environment required
Features:
- Lightweight self-hosted personal agent runtime
- Messaging-based interface, such as WhatsApp or Telegram
- Runs sessions in isolated containers
- Scheduled tasks, memory, web access, and skills
- Supports multi-agent or swarm-style workflows
- Can offload selected tasks to Ollama via an MCP server
Good for:
- personal AI agents
- messaging-first workflows
- local and self-hosted execution
- users who want a small and hackable runtime
- hybrid setups where Claude orchestrates and Ollama handles cheaper subtasks
Feels more like a personal agent harness than an IDE or CLI workflow tool.
Ollama fit:
- auxiliary support, not the main documented orchestrator path
- delegates summarization, translation, and similar cheaper tasks to local Ollama models
- Claude remains the orchestrator
Hermes Agent
Auth:
- Uses the auth of the underlying models and tools
- OpenAI-compatible local endpoints supported, including Ollama
- Messaging or gateway integrations require their own setup
- Local or server runtime required
Features:
- Persistent personal agent runtime that lives on your machine or server
- CLI plus messaging interfaces
- Memory, profiles, and auto-generated skills
- Scheduled automations and recurring tasks
- Delegation to isolated subagents and sandboxed execution backends
- Broad tool support, including MCP-style integrations
- Works with local models through OpenAI-compatible endpoints such as Ollama
Good for:
- long-running personal agents
- automation and recurring workflows
- messaging-based agent access
- users who want a more complete personal agent system
Feels more full-featured and productized than the lighter personal-agent runtimes.
Ollama fit:
- supported through custom or OpenAI-compatible endpoint configuration
- workable for local setups, but less native and explicit than OpenClaw's Ollama integration
Model Runtimes (Local Inference Backends)
Run models locally and expose them to other tools.
Ollama
Auth:
- No provider account required for local models
- Optional API keys when routing to external providers
Features:
- Simple local model serving
- Pull and switch models quickly
- Local API endpoint for other tools
- Broad ecosystem support
Good for:
- local-first workflows
- privacy or offline usage
- powering Continue, aichat, OpenCode, AgentPipe
- fast experimentation with local models
Common default choice for local runtimes.
Model Access CLIs (General-purpose model CLIs)
Let you interact with models directly from the terminal.
aichat
Auth:
- API key in config
Features:
- All-in-one LLM CLI
- Multi-provider support
- REPL, shell, agent, RAG
Good for:
- terminal workflow
- scripting
- switching models fast
Easy-LLM-CLI
Auth:
- Google login for default Gemini flow
- Custom providers use API key
Features:
- Model-agnostic CLI clients
- OpenAI-compatible APIs
- Custom providers supported
Good for:
- model-agnostic usage
- custom provider access
- OpenAI-compatible endpoint usage
- multi-provider CLI workflows
Common in builder setups.
ShellGPT
Auth:
- API key
Features:
- Simple CLI client
- Supports multiple providers via API keys
- Lightweight
Good for:
- lightweight terminal usage
- API-key based provider switching
- scripting helpers
- use inside orchestrators
Often used inside orchestrators.
XandAI CLI
Auth:
- Local endpoint
- Ollama or LM Studio
- No subscription login shown
Features:
- Hybrid local + cloud tools
- Supports Ollama, LM Studio, APIs
Good for:
- local + remote model usage
- Ollama or LM Studio backends
- hybrid CLI workflows
Used in local-first stacks.
Coding Agent CLIs (Coding-focused agent tools)
Specialized tools for coding tasks with repo awareness.
Used together in orchestration setups.
Examples:
Good for:
- terminal coding-agent workflows
- combining specialized agents
- orchestration with scripts and wrappers
Power users combine them with scripts.
IDE Agent Extensions (Editor-native coding tools)
Bring coding agents directly into editors.
Continue
Features:
- VS Code integration
- Supports local and cloud models
- Codebase-aware chat and edits
- Works with Ollama and provider APIs
- Custom assistants and workflows
Good for:
- daily coding workflows
- repo-aware editing
- local coding setups
- fast in-editor iteration
Codex (IDE integrations)
Mentioned above as a primary CLI tool. Also relevant here when used through editor integrations and coding workflows.
Good for:
- in-editor coding assistance
- repo-aware generation and edits
- pairing CLI and editor workflows
Usage Patterns
How people combine these tools in practice.
Tool-integrated orchestration
Uses external tools, plugins, or protocol-based integrations, often through MCP.
Used for:
- tool calling
- external integrations
- multi-agent tool sharing
Common examples:
- Gemini CLI with MCP
- Claude Code with MCP
- OpenCode with MCP integrations
- custom MCP servers
- Goose / MCP tool runners
Pattern:
- model
- MCP
- tools
- orchestrator
Local-first orchestration setups
Focus on local models + optional cloud.
Common tools:
- Ollama
- LM Studio
- aichat
- OpenCode
- AgentPipe
- shell scripts
Pattern:
- local runtime
- CLI agent
- optional cloud model
- orchestrator
Example Stacks
Example stack 1
AgentPipe:
- Claude Code
- Gemini CLI
- Ollama
- OpenAI API
Example stack 2
OpenCode:
- GPT
- Claude
- Gemini
- local models
Example stack 3
aichat:
- OpenAI
- Anthropic
- custom endpoint