490 lines
28 KiB
Markdown
490 lines
28 KiB
Markdown
# Why Hermes
|
|
|
|
Hermes is a persistent, autonomous AI agent that runs on your server. It has layered memory that
|
|
accumulates across sessions, a cron scheduler that fires jobs while you're offline, and a
|
|
self-improving skills system that saves reusable procedures automatically. You reach it from a
|
|
terminal, a browser, or a messaging app — and it's the same agent with the same history every time.
|
|
|
|
This document explains the mental model, how Hermes compares to other tools honestly, and where
|
|
it is and is not the right choice.
|
|
|
|
---
|
|
|
|
## The real problem: most tools are excellent in the moment and weak over time
|
|
|
|
Memory is no longer a differentiator on its own. ChatGPT, Claude, Cursor, and GitHub Copilot all
|
|
have some form of memory now. Anthropic, OpenAI, and Microsoft are all shipping scheduling and
|
|
agent features. The category boundaries that existed twelve months ago are blurring fast.
|
|
|
|
Hermes is not the only tool with memory or automation. It is the tool that makes those
|
|
capabilities durable, self-hosted, cross-surface, and cumulative on your own server. The
|
|
distinction that matters is not "has memory" vs. "has no memory" — it's whether context persists
|
|
across sessions automatically, whether execution happens on hardware you control, whether you can
|
|
reach the same agent identity from any device, and whether the system gets meaningfully better at
|
|
your specific workflow over time without manual configuration.
|
|
|
|
```
|
|
Session-scoped: You -> [Tool] -> Answer -> Done
|
|
(some tools now carry memory, but the execution is stateless)
|
|
|
|
Persistent agent: You <-> [Hermes] <-> (memory, skills, schedule, tools, surfaces)
|
|
(runs on your server, accumulates context, acts on your behalf offline)
|
|
```
|
|
|
|
---
|
|
|
|
## A note on convergence
|
|
|
|
The market is converging. Chat assistants are adding task scheduling and file connectors. IDE
|
|
tools are launching cloud agent modes. CLI tools are adding skills systems and mobile surfaces.
|
|
The lines between "assistant," "editor," and "agent" are dissolving.
|
|
|
|
This makes comparisons harder but also makes the question sharper: what actually matters when
|
|
every tool is claiming some version of every feature? For Hermes, the answer is synthesis. Any
|
|
single feature — memory, scheduling, messaging — is available somewhere else. The value is
|
|
having all of them in one self-hosted system, running continuously, with a persistent identity
|
|
that accumulates real knowledge of your stack over time.
|
|
|
|
---
|
|
|
|
## The three pillars
|
|
|
|
### 1. Memory that compounds
|
|
|
|
Hermes has layered memory that survives every session, every reboot, and every model swap:
|
|
|
|
- User profile — who you are, your preferences, your communication style, things you've corrected Hermes on
|
|
- Agent memory — facts about your environment, your toolchain, your project conventions
|
|
- Skills — reusable procedures Hermes discovers and saves automatically; it never has to relearn how to deploy your app, run your tests, or review a PR
|
|
- Session history — every past conversation is searchable; Hermes can recall what you worked on last Tuesday
|
|
|
|
When you correct Hermes, it remembers. When it solves a tricky problem, it saves the approach.
|
|
When it learns your stack, that knowledge carries into every future session. You never configure
|
|
this manually — it happens in the background as a side effect of normal use.
|
|
|
|
### 2. Autonomous scheduling
|
|
|
|
Hermes can run jobs without you present — every hour, every morning, on any cron schedule. It
|
|
fires up a fresh session with full access to your memory and skills, runs the task, and delivers
|
|
the result wherever you want it: Telegram, Discord, Slack, Signal, WhatsApp, SMS, email, and more.
|
|
|
|
Things Hermes can do while you sleep:
|
|
|
|
- Review new pull requests on your GitHub repo and post a full verdict comment
|
|
- Send a morning briefing of news, markets, or anything else you track
|
|
- Run your test suite and alert you if something breaks
|
|
- Watch a competitor's blog for new posts and summarize them
|
|
- Monitor a datasource and notify you when a threshold is crossed
|
|
|
|
The difference from cloud-scheduled alternatives is that the job runs on your server, with your
|
|
memory and skills, and your data never leaves your hardware.
|
|
|
|
### 3. Reach it from anywhere
|
|
|
|
Hermes runs on your server and is reachable from every surface: terminal over SSH, the web UI
|
|
(this project), and messaging apps including Telegram, Discord, Slack, WhatsApp, Signal, and
|
|
Matrix. Start a task from your phone, check it from the browser on your laptop, continue it in
|
|
a terminal on a remote server. The same agent, memory, and history follow you across all of them.
|
|
|
|
---
|
|
|
|
## How AI tools are layered today
|
|
|
|
The old four-category model — chat, editor, CLI, agent — is too clean. These layers are actively
|
|
collapsing into each other. Here is a more honest picture:
|
|
|
|
Chat assistants (Claude.ai, ChatGPT) now have persistent memory, task scheduling, 50+ service
|
|
connectors, and in some cases full agent modes with computer use. They are no longer "just chat."
|
|
|
|
IDE tools (Cursor, Windsurf, Copilot) have shipped or are shipping cross-session memory,
|
|
cloud-based background agents, and in Cursor's case a full Automations platform with Slack
|
|
integration. Cursor v3.0 (April 2026) is explicitly agent-first.
|
|
|
|
CLI tools (Claude Code, Codex, OpenCode) have added hooks, skills, desktop app automations,
|
|
and multi-surface reach. Claude Code now spans terminal, IDE, desktop, and browser. Codex has
|
|
become a product family: CLI, IDE extension, desktop app, and Codex Cloud.
|
|
|
|
Persistent self-hosted agents (Hermes, OpenClaw) sit at the intersection: they combine the
|
|
tool-use power of CLI agents, the memory of chat assistants, the scheduling of automation
|
|
platforms, and the cross-surface reach of messaging integrations — running continuously on
|
|
hardware you own.
|
|
|
|
The question is not which category a tool belongs to. The question is which combination of
|
|
capabilities you actually need, where that execution lives, and whether the system gets better
|
|
at your specific context over time.
|
|
|
|
---
|
|
|
|
## How Hermes compares
|
|
|
|
### vs. OpenClaw
|
|
|
|
OpenClaw is the most direct comparison and the question most people ask first. Both are
|
|
open-source, self-hosted, always-on agents with persistent memory, cron scheduling, and messaging
|
|
app integration. If you're evaluating Hermes, evaluate OpenClaw too.
|
|
|
|
OpenClaw (MIT) is built around a Gateway control plane written in Node.js/TypeScript. It has the
|
|
widest messaging coverage in the space — 24+ channels including WhatsApp, Telegram, Signal,
|
|
iMessage, LINE, WeChat, Slack, Discord, Teams, Matrix, Google Chat, Feishu, Mattermost, IRC,
|
|
Nextcloud Talk, and more. It has native Chrome/Chromium control via CDP, voice wake words on
|
|
macOS and iOS, and a ClawHub marketplace with 10,700+ skills. The community is large (350k+
|
|
GitHub stars, 16,900+ commits) and growing.
|
|
|
|
Hermes is built in Python and centers on a self-improving agent loop rather than a gateway
|
|
control plane. The core architectural difference is in skills: OpenClaw skills are primarily
|
|
human-authored plugins installed from a marketplace. Hermes writes and saves its own skills
|
|
automatically as part of every session. When Hermes solves a problem a new way, it saves the
|
|
procedure and reuses it without any user effort. That's not a subtle distinction — it's the
|
|
reason Hermes gets meaningfully better at your workflow without you maintaining a plugin library.
|
|
|
|
Two practical differences worth knowing directly:
|
|
|
|
Stability. OpenClaw's GitHub issues and community forums document recurring update-breaking
|
|
regressions. Telegram integration was broken across multiple releases from early 2026 through
|
|
at least April 2026. The unofficial WhatsApp Web protocol OpenClaw relies on disconnects and
|
|
requires periodic re-pairing — this is in OpenClaw's own FAQ.
|
|
|
|
Security. ClawHub's open publishing model has been exploited at scale. Three separate audits in
|
|
early 2026 found serious problems: Koi Security (January 2026) linked 335 skills to a campaign
|
|
called "ClawHavoc" that delivered Atomic Stealer malware on macOS; Bitdefender found roughly
|
|
900 malicious packages representing about 20% of the ecosystem at the time; Snyk's "ToxicSkills"
|
|
report (February 2026) found malicious skills across roughly 4,000 scanned packages. China's
|
|
CNCERT issued a national warning about ClawHub. Hermes has no third-party marketplace and a
|
|
correspondingly smaller attack surface.
|
|
|
|
OpenClaw's genuine strengths are worth stating plainly: broader messaging coverage (iMessage,
|
|
LINE, WeChat, Teams, Google Chat — platforms Hermes does not support), native browser and
|
|
computer control via Chrome CDP, voice wake words, a larger community, and more third-party
|
|
integrations than Hermes. If those capabilities matter most, OpenClaw is worth a serious look.
|
|
|
|
Where Hermes fits better: you want an agent that self-improves from experience without managing
|
|
a plugin library, you work in Python and want the ML/data science ecosystem, you want a stable
|
|
deployment that doesn't break between updates, or you want a full web chat UI rather than a
|
|
control dashboard.
|
|
|
|
| | OpenClaw | Hermes |
|
|
|---|---|---|
|
|
| Persistent memory | Yes | Yes |
|
|
| Scheduled jobs (cron) | Yes | Yes |
|
|
| Messaging app access | Yes (24+ platforms, incl. iMessage/WeChat/LINE) | Yes (many platforms) |
|
|
| Web UI | Chat UI + control dashboard | Full three-panel chat UI |
|
|
| Self-hosted | Yes | Yes |
|
|
| Open source | Yes (MIT) | Yes |
|
|
| Self-improving skills | Partial (AI can generate; not the default loop) | Yes (automatic, first-class) |
|
|
| Browser / computer control | Yes (native Chrome CDP) | Via shell / tools |
|
|
| Voice wake words | Yes (macOS/iOS) | No |
|
|
| Python / ML ecosystem | No (Node.js) | Yes |
|
|
| Orchestrates Claude Code / Codex | No | Yes |
|
|
| Multi-profile support | Via binding-rule routing | Yes (first-class named profiles) |
|
|
| Provider-agnostic | Yes | Yes |
|
|
| Update reliability | Moderate (documented regressions) | High |
|
|
| Memory inspectability | Limited | Yes (markdown files, editable) |
|
|
| Self-hosted autonomous execution | Yes | Yes |
|
|
|
|
### vs. Claude Code (Anthropic)
|
|
|
|
Claude Code is Anthropic's official agentic tool and one of the strongest options for focused
|
|
coding sessions. It has deep code understanding, shell access, file editing, and multi-step
|
|
reasoning. It has been expanding rapidly — it now spans terminal, IDE plugin, desktop app, and
|
|
browser surfaces — and the gap is closing in several areas.
|
|
|
|
What Claude Code has that's worth knowing:
|
|
|
|
- Hooks system — 26 event types (SessionStart, PreToolUse, PostToolUse, Stop, and more) with
|
|
4 handler types (shell command, HTTP endpoint, LLM prompt, sub-agent); gives deterministic
|
|
non-LLM control over the agent lifecycle
|
|
- Plugins / Skills — installable via `/plugin install`, hot-reloaded from `~/.claude/skills`,
|
|
with a marketplace; includes the official ralph-wiggum plugin (`/ralph-loop`) for
|
|
autonomous iteration toward a completion goal (distinct from `/loop`)
|
|
- `/loop` — a native bundled skill, available in every session without any plugin, that runs
|
|
a prompt on a repeating schedule within an active CLI session (polling/monitoring use case);
|
|
session-scoped, dies when the terminal closes
|
|
- Scheduling — cloud-managed cron (Anthropic infrastructure, minimum 1-hour interval) and
|
|
desktop app scheduled tasks (run locally while the app is open, minimum 1-minute interval,
|
|
full local file access); no self-hosted cron
|
|
- Messaging channels — Telegram, Discord, and iMessage via the Channels feature (research
|
|
preview, requires Bun runtime); Slack is the most-requested addition and has not yet shipped
|
|
- Memory — CLAUDE.md and MEMORY.md for project-level context; auto-memory since v2.1.59+
|
|
- Claude Cowork — a separate knowledge-worker product connecting 38+ services via MCP
|
|
including Gmail, Microsoft Teams, Notion, Jira, Salesforce, and more
|
|
|
|
Claude Code's source was briefly and accidentally made public in March 2026 before being taken
|
|
down. The CLI ships as minified/bundled TypeScript compiled with Bun — it is not open source.
|
|
|
|
Key differences that remain:
|
|
|
|
- Scheduling requires cloud (Anthropic infrastructure, data off your hardware, 1-hour minimum)
|
|
or the desktop app (runs locally, but the app must stay open — not a headless server process);
|
|
neither runs as a server daemon the way Hermes cron does
|
|
- Memory is project-file-based (CLAUDE.md / MEMORY.md plus rolling auto-memory); it doesn't
|
|
automatically accumulate a cross-project knowledge graph the way Hermes does
|
|
- Not provider-agnostic — routes through Anthropic, Bedrock, Vertex, or Foundry, but always
|
|
a Claude model; you can't switch to GPT, Gemini, or a local model
|
|
- Messaging channels are still a research preview, not production
|
|
|
|
Hermes can use Claude Code as a sub-agent. For large implementation tasks, Hermes can spawn
|
|
Claude Code to handle the heavy lifting and fold the result back into its own memory and history.
|
|
|
|
| | Claude Code | Hermes |
|
|
|---|---|---|
|
|
| Persistent memory (automatic) | Partial (CLAUDE.md / MEMORY.md + auto-memory v2.1.59+) | Yes |
|
|
| Skills / hooks system | Yes (26-event Hooks + Plugin/Skills marketplace) | Yes (auto-generated from experience) |
|
|
| Scheduled jobs (self-hosted) | No (cloud or desktop-app only) | Yes |
|
|
| Messaging access | Partial (Telegram/Discord/iMessage research preview; Slack not yet) | Yes (many platforms, production) |
|
|
| Cowork connectors (Slack, Gmail, etc.) | Yes (via Claude Cowork, separate product) | Via agent tool use |
|
|
| Web UI | Yes (claude.ai/code, Anthropic-hosted) | Yes (self-hosted) |
|
|
| Provider-agnostic | No (Claude models only) | Yes (any provider) |
|
|
| Self-hosted scheduling | No | Yes |
|
|
| Open source | No | Yes |
|
|
| Background/cloud agent mode | Yes (cloud-scheduled) | Yes (self-hosted cron) |
|
|
| Runs as sub-agent of Hermes | Yes | N/A |
|
|
| Memory inspectability | Partial (CLAUDE.md readable; auto-memory less so) | Yes (markdown files) |
|
|
|
|
### vs. Codex CLI (OpenAI)
|
|
|
|
Codex CLI (Apache 2.0, ~60k GitHub stars) started as a straightforward terminal tool and has
|
|
expanded into a product family. It was rewritten from TypeScript to Rust. It now includes an IDE
|
|
extension, a desktop app with an Automations feature, and Codex Cloud for remote execution. A
|
|
Skills system is shared across surfaces. It supports 12+ built-in providers: OpenAI, Anthropic,
|
|
Google/Gemini, Mistral, Groq, Ollama, OpenRouter, LM Studio, Together AI, DeepSeek, xAI,
|
|
Azure OpenAI, and custom endpoints.
|
|
|
|
The CLI itself has no native scheduling (open feature request). Session continuity is available
|
|
via `codex resume`. Memory is session-history-based plus AGENTS.md project context — not a
|
|
living knowledge graph that accumulates across all your projects. No first-party messaging
|
|
integration. The Automations feature in the desktop app covers scheduled local tasks but doesn't
|
|
reach the cross-session, cross-surface continuity Hermes has.
|
|
|
|
| | Codex CLI | Hermes |
|
|
|---|---|---|
|
|
| Persistent memory | Partial (session history + AGENTS.md) | Yes (automatic, layered) |
|
|
| Scheduled jobs | Partial (desktop app Automations; CLI has none) | Yes |
|
|
| Messaging app access | No | Yes |
|
|
| Web UI | No (CLI + desktop app) | Yes (self-hosted) |
|
|
| Provider-agnostic | Yes (12+ providers) | Yes |
|
|
| Self-hosted | Yes | Yes |
|
|
| Open source | Yes (Apache 2.0) | Yes |
|
|
| Background/cloud agent mode | Yes (Codex Cloud) | Yes (self-hosted cron) |
|
|
| Self-improving skills | No | Yes |
|
|
|
|
### vs. OpenCode
|
|
|
|
OpenCode is an open-source TUI agentic coding assistant supporting 75+ providers. It has a WebUI
|
|
embedded in its binary, an official desktop app, SQLite session history, and AGENTS.md project
|
|
context. It supports CLAUDE.md as a fallback for users migrating from Claude Code. There are 30+
|
|
community plugins, and community messaging integrations exist for Telegram, Slack, Discord, and
|
|
Microsoft Teams — though none are first-party and all require manual setup.
|
|
|
|
OpenCode Go ($10/month) and OpenCode Zen (curated model service) are subscription tiers. The
|
|
GitHub Copilot official integration launched January 2026. There is no native scheduling; a
|
|
community background plugin exists. No automatic cross-session semantic memory.
|
|
|
|
| | OpenCode | Hermes |
|
|
|---|---|---|
|
|
| Persistent memory | Partial (session history + AGENTS.md) | Yes (automatic, layered) |
|
|
| Scheduled jobs | No (community plugin only) | Yes |
|
|
| Messaging app access | Community integrations only (Telegram/Slack/Discord/Teams) | Yes (first-party, many platforms) |
|
|
| Web UI | Yes (embedded + desktop app) | Yes (self-hosted) |
|
|
| Mobile access | No | Yes |
|
|
| Skills / plugins | Yes (30+ community plugins) | Yes (auto-generated, first-party) |
|
|
| Provider-agnostic | Yes (75+ providers) | Yes |
|
|
| Open source | Yes | Yes |
|
|
| Self-hosted autonomous execution | No | Yes |
|
|
|
|
### vs. Cursor
|
|
|
|
Cursor has changed substantially. The "no memory, no scheduling, no messaging" description was
|
|
accurate in 2024 and is wrong now.
|
|
|
|
Memories (per-project cross-session knowledge base) shipped in beta with v1.0 in June 2025.
|
|
Automations launched March 5, 2026 — time-based, event-based (GitHub/Linear/PagerDuty), and
|
|
communication-based (Slack) triggers that fire background agents on cloud VMs. The web app,
|
|
mobile agent, and Slack bot give it multi-surface reach. Cursor v3.0 (April 2, 2026) is
|
|
explicitly agent-first with Design Mode and 30+ marketplace plugins. Cursor acquired Supermaven
|
|
for autocomplete. As of early 2026 it's valued at $29.3B with $2B ARR. It is not a narrow editor
|
|
tool anymore.
|
|
|
|
Hermes still has a different profile: it's self-hosted and server-resident, the same persistent
|
|
identity follows you across every surface without cloud intermediation, and it works with any
|
|
model family rather than being cloud-VM-based. For workflows that require data sovereignty,
|
|
self-hosted scheduling, or deep Python/ML tooling on your own hardware, Cursor's cloud-agent
|
|
architecture is a fundamental mismatch. For teams that want editor-native agents with strong
|
|
IDE integration, Cursor's recent evolution is significant.
|
|
|
|
| | Cursor | Windsurf | Copilot | Hermes |
|
|
|---|---|---|---|---|
|
|
| In-editor autocomplete | Excellent (Supermaven) | Excellent (Cascade) | Excellent | No |
|
|
| Inline diff / refactor | Yes | Yes | Yes | Via shell |
|
|
| Cross-session memory | Yes (Memories, per-project) | Yes (Cascade Memories, workspace) | Yes (Agentic Memory, repo-scoped, 28-day expiry) | Yes (automatic, persistent) |
|
|
| Scheduled background jobs | Yes (Automations, cloud VM) | No | Via Coding Agent (issue-driven) | Yes (self-hosted cron) |
|
|
| Messaging app / multi-surface | Yes (Slack bot, web app, mobile) | No | Via Copilot CLI / fleet | Yes (many platforms) |
|
|
| Background/cloud agent mode | Yes (Automations on cloud VMs) | No | Yes (Coding Agent, GA Mar 2026) | Yes (self-hosted) |
|
|
| Terminal tool use | Limited | Limited | Limited | Full |
|
|
| Self-hosted | No | No | No | Yes |
|
|
| Self-hosted autonomous execution | No | No | No | Yes |
|
|
| Provider-agnostic | Partial | Partial | No (GitHub models) | Yes |
|
|
| Open source | No | No | No | Yes |
|
|
| Memory inspectability | Partial | Yes (stored locally) | Limited | Yes (markdown files) |
|
|
|
|
### vs. Claude.ai and ChatGPT
|
|
|
|
These are no longer simple chat tools. The description of "no memory, no scheduling, no
|
|
messaging" is inaccurate for both.
|
|
|
|
Claude Cowork (in Claude Desktop) launched scheduled tasks on February 25, 2026 — hourly,
|
|
daily, weekly, weekdays, and on-demand. It runs in an isolated VM with file and shell access.
|
|
Claude has 50+ service connectors as of February 2026 including Slack (launched January 26,
|
|
2026), Gmail, Google Calendar, Google Drive, Microsoft 365, Notion, Asana, Linear, and Jira.
|
|
Memory auto-generates from chat history, not just user-curated entries. Code execution and
|
|
file access in Artifacts is sandboxed, not the same as shell access on your own server.
|
|
|
|
ChatGPT has Agent Mode (launched July 17, 2025), Scheduled Tasks (January 2025, recurring
|
|
automated prompts), a computer-using agent, Projects, 50+ connectors including Gmail, GitHub,
|
|
and Google Drive, dual-mode memory (auto + manual), and ChatGPT Pulse for Pro users (daily
|
|
research briefings). It is not a passive Q&A interface.
|
|
|
|
Where Claude.ai and ChatGPT differ from Hermes: neither is self-hosted, neither is
|
|
provider-agnostic, and neither gives you execution on your own hardware. Connectors and
|
|
scheduling exist, but they run on Anthropic's or OpenAI's infrastructure. Your memory, session
|
|
history, and agent execution live on their servers, not yours. For many use cases that's fine
|
|
— they are capable and well-supported. For privacy-conscious users, regulated environments, or
|
|
workflows that require persistent server-side execution on controlled hardware, it's a
|
|
disqualifying constraint.
|
|
|
|
| | Claude.ai | ChatGPT | Hermes |
|
|
|---|---|---|---|
|
|
| Memory across conversations | Yes (auto-generated from history) | Yes (dual-mode: auto + manual) | Yes (deep, automatic) |
|
|
| Scheduled tasks | Yes (Cowork: hourly/daily/weekly) | Yes (since Jan 2025) | Yes (any cron, self-hosted) |
|
|
| Service connectors / messaging | Yes (50+ via Cowork) | Yes (50+ connectors) | Yes (many platforms, direct) |
|
|
| Runs shell commands | Sandboxed (Cowork VM) | Sandboxed | Yes (full shell) |
|
|
| Code execution | Sandboxed | Sandboxed | Yes (full shell) |
|
|
| Reads / writes files | Sandboxed | Sandboxed | Yes (full filesystem) |
|
|
| Web UI | Yes (Anthropic-hosted) | Yes (OpenAI-hosted) | Yes (self-hosted) |
|
|
| Self-hosted | No | No | Yes |
|
|
| Provider-agnostic | No | No | Yes |
|
|
| Open source | No | No | Yes |
|
|
| Self-hosted autonomous execution | No | No | Yes |
|
|
| Memory inspectability | Limited | Limited | Yes (markdown files) |
|
|
|
|
---
|
|
|
|
## The compounding advantage
|
|
|
|
What distinguishes Hermes from most of the tools above is that it gets meaningfully better at
|
|
your specific workflow over time without manual configuration.
|
|
|
|
Every time Hermes encounters a new environment, it saves facts to memory. Every time it solves
|
|
a problem a new way, it saves the approach as a skill. Every time you correct it, it updates its
|
|
profile of you. Every session, every scheduled job, every tool call adds to a body of knowledge
|
|
that is specific to you, stored on your hardware, and available to every future interaction.
|
|
|
|
A Claude Code session on day one and day one hundred are identical — it starts fresh. A Hermes
|
|
agent on day one and day one hundred knows your stack, your conventions, your preferences, and
|
|
the solutions that have worked before. That's the actual compounding.
|
|
|
|
---
|
|
|
|
## Who Hermes is for
|
|
|
|
Solo developers and power users who don't want to re-explain their stack every session and want
|
|
an AI that actually knows their environment.
|
|
|
|
Teams on a shared server where multiple people want capable AI access without each paying for
|
|
a separate subscription or running separate local tooling.
|
|
|
|
Automation-heavy workflows where you want an AI running tasks on a schedule, delivering results
|
|
to your phone, without babysitting it.
|
|
|
|
Privacy-conscious users who want their conversations, memory, and files on their own hardware.
|
|
|
|
Multi-model users who want to switch between OpenAI, Anthropic, Google, DeepSeek, and others
|
|
based on cost, capability, or rate limits, without rebuilding their workflow each time.
|
|
|
|
---
|
|
|
|
## What Hermes is not
|
|
|
|
Hermes is not the best in-editor autocomplete tool. Cursor and Windsurf do that job better.
|
|
Use one alongside Hermes.
|
|
|
|
It is not zero-setup. You are running a server. That means initial configuration, and it means
|
|
you're responsible for uptime, upgrades, and backups. The tradeoff is data sovereignty and
|
|
control; that only makes sense if you actually want it.
|
|
|
|
It does not make weaker models magical. Memory and skills help, but the underlying model still
|
|
determines reasoning quality. Hermes with a weak model is a well-organized weak model.
|
|
|
|
It still needs guardrails, approvals, and observability for high-stakes automations. Autonomous
|
|
execution on a schedule with shell access is powerful and requires judgment about what to
|
|
approve. Terminal commands can require confirmation before running; use that for anything
|
|
consequential.
|
|
|
|
If you need the absolute lowest-friction path to a one-off answer or a quick edit, a chat
|
|
interface or an in-editor tool is the right call. Hermes is for continuity and autonomy, not
|
|
minimum-friction one-shots.
|
|
|
|
---
|
|
|
|
## Scope and limits
|
|
|
|
Hermes lives in the terminal, browser, and messaging apps. For in-editor autocomplete and inline
|
|
diffs, use Cursor or Windsurf — they do that job better and work well alongside Hermes.
|
|
|
|
You run Hermes on your own server. That means initial setup, but your data stays on your
|
|
hardware and you control the schedule, the models, and the costs.
|
|
|
|
Hermes is an orchestration and memory layer. It makes whatever model you point at it more useful
|
|
over time. The models do the reasoning; Hermes makes sure that reasoning accumulates into
|
|
something durable.
|
|
|
|
---
|
|
|
|
## Security and control
|
|
|
|
Memory is stored locally on your server as readable, editable files: user profile, agent memory,
|
|
and skills are all markdown. Session history is in SQLite on your machine. You can inspect,
|
|
edit, or delete any of it directly.
|
|
|
|
If you want external memory providers, eight are supported: Mem0, Honcho, Hindsight, RetainDB,
|
|
ByteRover, Supermemory, Holographic, and others. These are optional and configurable.
|
|
|
|
Execution runs in configurable backends: local shell, Docker, SSH, Daytona, Singularity, or
|
|
Modal. You choose what execution environment Hermes operates in and what it can reach.
|
|
|
|
Terminal commands can require confirmation before running. For any automation that touches
|
|
production systems or makes external calls, enable approval controls.
|
|
|
|
Secrets stay on your hardware. Hermes does not phone home; it calls whatever model APIs you
|
|
configure directly.
|
|
|
|
Multiple profiles give isolation between users or projects. A shared server can have separate
|
|
profiles with separate memory, separate skills, and separate history.
|
|
|
|
---
|
|
|
|
## Quick reference
|
|
|
|
| | OpenClaw | Claude Code | Codex | OpenCode | Cursor | Copilot | Claude.ai | ChatGPT | Hermes |
|
|
|---|---|---|---|---|---|---|---|---|---|
|
|
| Persistent memory (auto) | Yes | Partial† | Partial | Partial | Yes (per-project) | Yes (repo-scoped‡) | Yes | Yes | Yes |
|
|
| Scheduled / background jobs | Yes | Partial§ | Partial¶ | No | Yes (Automations) | Via Coding Agent | Yes (Cowork) | Yes | Yes (self-hosted) |
|
|
| Messaging / multi-surface | Yes (24+ platforms) | Partial (preview) | No | Community only | Yes (Slack/web/mobile) | Via CLI/fleet | Yes (50+ connectors) | Yes (50+ connectors) | Yes (many platforms) |
|
|
| Web UI | Chat UI + control dashboard | Anthropic-hosted | No | Yes | Yes + mobile | github.com | Yes (Claude Desktop) | Yes | Yes (self-hosted) |
|
|
| Skills system | Yes (ClawHub marketplace) | Yes (Hooks + Plugins) | Partial (Skills) | Community plugins | Yes (marketplace) | No | No | No | Yes (auto-generated) |
|
|
| Self-improving skills | Partial | No | No | No | No | No | No | No | Yes |
|
|
| Browser / computer control | Yes (Chrome CDP) | No | No | No | No | No | No | Yes (CUA) | Via shell |
|
|
| In-editor autocomplete | No | No | Via extension | No | Excellent | Excellent | No | No | No |
|
|
| Orchestrates other agents | No | No | No | No | No | No | No | No | Yes |
|
|
| Provider-agnostic | Yes | No (Claude only) | Yes | Yes | Partial | No | No | No | Yes |
|
|
| Self-hosted | Yes | No | Yes (CLI) | Yes | No | No | No | No | Yes |
|
|
| Self-hosted autonomous execution | Yes | No | No | No | No | No | No | No | Yes |
|
|
| Background/cloud agent mode | Yes | Yes (cloud) | Yes (Codex Cloud) | No | Yes (cloud VMs) | Yes (Coding Agent) | Yes (Cowork VM) | Yes (Agent Mode) | Yes (self-hosted) |
|
|
| Memory inspectability | Limited | Partial | Partial | Partial | Partial | Limited | Limited | Limited | Yes (markdown files) |
|
|
| Open source | Yes (MIT) | No | Yes (Apache 2.0) | Yes | No | No | No | No | Yes |
|
|
| Always-on autonomous execution | Yes | No | No | No | No | No | No | No | Yes |
|
|
|
|
† Claude Code: CLAUDE.md / MEMORY.md project context plus auto-memory since v2.1.59+; no automatic cross-project accumulation
|
|
‡ Copilot Agentic Memory: public preview Jan 15, 2026; enabled by default Mar 4, 2026; repo-scoped, auto-expires after 28 days
|
|
§ Claude Code scheduling: cloud-managed (Anthropic infrastructure) or desktop-app only; no self-hosted cron
|
|
¶ Codex scheduling: desktop app Automations only; CLI has no native scheduling
|