DB DevBrain

Ai

Local, CLI, and Agentic LLM Tools

Overview

Mental Model

  • Models = raw intelligence
  • CLIs = how you talk to models
  • Agent runtimes = how models think and act
  • Orchestrators = how multiple things work together

Scope

Power-user tools for LLM orchestration.
Included:
  • CLI / desktop tools
  • multi-model / multi-provider
  • agent / orchestration capable
Excluded:
  • web apps
  • libraries / frameworks
  • SDKs

Core Tools

Orchestrators (Multi-agent coordination tools)

Coordinate multiple models, agents, or tools into workflows.

OpenCode

Auth:
  • Usually API keys for model providers
  • Can also use provider-specific login or OAuth plugins
  • Remote MCP servers can use OAuth or bearer tokens
Features:
  • CLI or desktop agent runner
  • Supports multiple providers, such as OpenAI, Claude, Gemini, etc.
  • Acts as orchestrator, not just chat
Good for:
  • multi-model workflows
  • agent execution
  • coding and automation
  • chaining tools
Common in power-user setups.

AgentPipe

Auth:
  • No single auth model
  • Uses each underlying CLI's auth
  • OpenRouter uses API key
  • AgentPipe Web bridge uses API key
Features:
  • Multi-agent CLI orchestrator
  • Runs multiple AI CLIs together
  • Supports multiple providers
  • Orchestrates conversations between multiple AI agents like Claude, Gemini, Ollama, Qwen, Cursor in one shared session
Use cases:
  • run Claude + local model + Gemini together
  • compare outputs
  • route tasks
  • multi-agent workflows
Acts as orchestration shell.

Maestro

Auth:
  • Inherits auth from connected CLIs and providers
  • API keys for providers
  • Optional service tokens for integrations
Features:
  • Workflow orchestrator for agents and tools
  • Declarative pipelines for multi-step tasks
  • Scheduling and retries
  • Observability hooks for runs
Good for:
  • repeatable pipelines
  • productionizing agent workflows
  • chaining tools with control over flow
  • monitoring and debugging runs
Often used when moving from experiments to repeatable workflows.

Agent Runtimes (Autonomous agent tools)

Run agent logic like planning, memory, and tool usage.

deepagents

Auth:
  • Uses each provider's API key
Features:
  • LangChain or LangGraph agent harness
  • Planning + sub-agents + filesystem tools
  • Designed for complex multi-step agents
  • Can run as CLI or SDK
Good for:
  • advanced agent workflows
  • research or coding agents
  • custom orchestration
  • building your own agent runtime
Used in advanced agent stacks.

OpenClaw

Auth:
  • Provider API keys
  • Native Ollama support for local models
  • Local model endpoints optional
Features:
  • Agent runtime focused on autonomy
  • Tool use with iterative planning loops
  • Memory and context management
  • Designed for long-running tasks
  • Native Ollama integration with streaming and tool calling
Good for:
  • autonomous agents
  • long multi-step tasks
  • experiments with self-directed workflows
  • combining tools with planning loops
  • local-model setups through Ollama
Often explored in experimental or research-heavy setups.
Ollama fit:
  • first-class support
  • uses Ollama's native API, not just OpenAI-compatible mode
  • strong option if you want local models as the main serving layer

NanoClaw

Auth:
  • Claude is the main orchestrator model in the documented Ollama setup
  • Uses the auth of the underlying agent or model setup
  • Messaging platform setup required for connected channels
  • Local runtime and container environment required
Features:
  • Lightweight self-hosted personal agent runtime
  • Messaging-based interface, such as WhatsApp or Telegram
  • Runs sessions in isolated containers
  • Scheduled tasks, memory, web access, and skills
  • Supports multi-agent or swarm-style workflows
  • Can offload selected tasks to Ollama via an MCP server
Good for:
  • personal AI agents
  • messaging-first workflows
  • local and self-hosted execution
  • users who want a small and hackable runtime
  • hybrid setups where Claude orchestrates and Ollama handles cheaper subtasks
Feels more like a personal agent harness than an IDE or CLI workflow tool.
Ollama fit:
  • auxiliary support, not the main documented orchestrator path
  • delegates summarization, translation, and similar cheaper tasks to local Ollama models
  • Claude remains the orchestrator

Hermes Agent

Auth:
  • Uses the auth of the underlying models and tools
  • OpenAI-compatible local endpoints supported, including Ollama
  • Messaging or gateway integrations require their own setup
  • Local or server runtime required
Features:
  • Persistent personal agent runtime that lives on your machine or server
  • CLI plus messaging interfaces
  • Memory, profiles, and auto-generated skills
  • Scheduled automations and recurring tasks
  • Delegation to isolated subagents and sandboxed execution backends
  • Broad tool support, including MCP-style integrations
  • Works with local models through OpenAI-compatible endpoints such as Ollama
Good for:
  • long-running personal agents
  • automation and recurring workflows
  • messaging-based agent access
  • users who want a more complete personal agent system
Feels more full-featured and productized than the lighter personal-agent runtimes.
Ollama fit:
  • supported through custom or OpenAI-compatible endpoint configuration
  • workable for local setups, but less native and explicit than OpenClaw's Ollama integration

Model Runtimes (Local Inference Backends)

Run models locally and expose them to other tools.

Ollama

Auth:
  • No provider account required for local models
  • Optional API keys when routing to external providers
Features:
  • Simple local model serving
  • Pull and switch models quickly
  • Local API endpoint for other tools
  • Broad ecosystem support
Good for:
  • local-first workflows
  • privacy or offline usage
  • powering Continue, aichat, OpenCode, AgentPipe
  • fast experimentation with local models
Common default choice for local runtimes.

Model Access CLIs (General-purpose model CLIs)

Let you interact with models directly from the terminal.

aichat

Auth:
  • API key in config
Features:
  • All-in-one LLM CLI
  • Multi-provider support
  • REPL, shell, agent, RAG
Good for:
  • terminal workflow
  • scripting
  • switching models fast

Easy-LLM-CLI

Auth:
  • Google login for default Gemini flow
  • Custom providers use API key
Features:
  • Model-agnostic CLI clients
  • OpenAI-compatible APIs
  • Custom providers supported
Good for:
  • model-agnostic usage
  • custom provider access
  • OpenAI-compatible endpoint usage
  • multi-provider CLI workflows
Common in builder setups.

ShellGPT

Auth:
  • API key
Features:
  • Simple CLI client
  • Supports multiple providers via API keys
  • Lightweight
Good for:
  • lightweight terminal usage
  • API-key based provider switching
  • scripting helpers
  • use inside orchestrators
Often used inside orchestrators.

XandAI CLI

Auth:
  • Local endpoint
  • Ollama or LM Studio
  • No subscription login shown
Features:
  • Hybrid local + cloud tools
  • Supports Ollama, LM Studio, APIs
Good for:
  • local + remote model usage
  • Ollama or LM Studio backends
  • hybrid CLI workflows
Used in local-first stacks.

Coding Agent CLIs (Coding-focused agent tools)

Specialized tools for coding tasks with repo awareness.
Used together in orchestration setups.
Examples:
Good for:
  • terminal coding-agent workflows
  • combining specialized agents
  • orchestration with scripts and wrappers
Power users combine them with scripts.

IDE Agent Extensions (Editor-native coding tools)

Bring coding agents directly into editors.

Continue

Features:
  • VS Code integration
  • Supports local and cloud models
  • Codebase-aware chat and edits
  • Works with Ollama and provider APIs
  • Custom assistants and workflows
Good for:
  • daily coding workflows
  • repo-aware editing
  • local coding setups
  • fast in-editor iteration

Codex (IDE integrations)

Mentioned above as a primary CLI tool. Also relevant here when used through editor integrations and coding workflows.
Good for:
  • in-editor coding assistance
  • repo-aware generation and edits
  • pairing CLI and editor workflows

Usage Patterns

How people combine these tools in practice.

Tool-integrated orchestration

Uses external tools, plugins, or protocol-based integrations, often through MCP.
Used for:
  • tool calling
  • external integrations
  • multi-agent tool sharing
Common examples:
  • Gemini CLI with MCP
  • Claude Code with MCP
  • OpenCode with MCP integrations
  • custom MCP servers
  • Goose / MCP tool runners
Pattern:
  • model
  • MCP
  • tools
  • orchestrator

Local-first orchestration setups

Focus on local models + optional cloud.
Common tools:
  • Ollama
  • LM Studio
  • aichat
  • OpenCode
  • AgentPipe
  • shell scripts
Pattern:
  • local runtime
  • CLI agent
  • optional cloud model
  • orchestrator

Example Stacks

Example stack 1

AgentPipe:
  • Claude Code
  • Gemini CLI
  • Ollama
  • OpenAI API

Example stack 2

OpenCode:
  • GPT
  • Claude
  • Gemini
  • local models

Example stack 3

aichat:
  • OpenAI
  • Anthropic
  • custom endpoint