A new open-source AI agent, Hermes, attracted over 22,000 GitHub stars by automating the two most tedious parts of agent development: skill creation and user-specific memory.
Back
A new open-source AI agent, Hermes, attracted over 22,000 GitHub stars by automating the two most tedious parts of agent development: skill creation and user-specific memory.

(P1) A new AI agent framework, Hermes Agent, has captured the top spot on GitHub's trending list for weeks, accumulating over 22,000 stars and challenging the dominance of its predecessor, OpenClaw. While both agents appear functionally similar, Hermes introduces a radical philosophy of full automation that is setting a new direction for the industry, pushing the value proposition from developer tool to autonomous partner.
(P2) The technical core of Hermes’s self-evolving skill system is based on the Genetic-Pareto Prompt Evolution (GEPA) algorithm, detailed in an ICLR 2026 paper by Lakshya Agrawal and others. "Reflective prompt evolution can outperform reinforcement learning," the paper argues, providing the academic backbone for Hermes's departure from traditional RL-based skill enhancement.
(P3) Hermes’s differentiation comes from two key automated systems. First, a self-evolving skill system silently generates new workflows when a tool is used more than five times or an error is recovered. These skills are then optimized offline using the GEPA algorithm. Second, a proactive memory system uses a "nudge" mechanism every 15 conversation turns to actively reflect on and save user preferences, a stark contrast to OpenClaw's passive, overflow-prevention memory saves.
(P4) The rapid ascent of Hermes signals a potential paradigm shift for the $25 billion AI development market. Its success suggests a strong developer appetite for agents that reduce manual configuration and actively learn. This could accelerate M&A activity around promising open-source projects and force established players like Anthropic, which was publicly called out for copying Hermes's features, to adopt a more automated, user-centric design philosophy.
The most significant structural difference between Hermes and its rivals is its closed-loop system for automatic skill evolution. While OpenClaw requires users to manually create, install, and authorize new skills, Hermes automates this process entirely. The system silently packages any successful workflow involving five or more tool calls into a new skill file.
Later, a separate offline process detailed in the hermes-agent-self-evolution repository uses the DSPy framework and the GEPA algorithm to refine these skills. This algorithm rests on three concepts: reflective mutation, where the model analyzes execution traces to make targeted changes; Pareto frontier selection, which preserves a diverse set of high-performing skill variations; and using natural language feedback as the primary signal for mutation. This process generates a pull request for human review, ensuring a human-in-the-loop maintains final control over the agent's core capabilities, debunking the myth of a system that operates with no user oversight.
Hermes's second key innovation is its aggressive and proactive memory system. While competitors like Claude Code have memory systems strictly isolated to a single project, and OpenClaw passively saves memories only before a context window overflows, Hermes takes a different approach. Approximately every 15 turns of conversation, a "nudge" mechanism forces the agent to reflect on the interaction and decide if any user preferences or facts are worth remembering permanently.
This high-frequency, active approach ensures a much richer user model is built over time. The system is further enhanced by a built-in SQLite FTS5 full-text search capability, allowing the agent to instantly recall past interactions without relying on external vector databases. While the advanced, AI-native memory backend Honcho was the default in earlier versions, the v0.7 update made it an optional plugin, prioritizing the stability of the simpler built-in system and giving users more control.
Hermes’s automation is achieved by replacing flexible model-based judgments with deterministic, hard-coded rules. The system's complexity is not eliminated but transferred from the user's responsibility to the agent's underlying code. Decisions like when to generate a skill (5 tool calls) or when to reflect on memory (15 turns) are governed by rigid if-then logic, not the LLM's reasoning.
This design choice is a pragmatic engineering solution to the current limitations of LLMs in managing long contexts, where studies have shown performance can drop by over 39%. When a conversation reaches 85% of the context limit, Hermes uses simple string replacement for compression rather than risking a faulty AI-generated summary. This conservative, rules-based approach ensures stability and predictability, which Hermes's creators bet is more valuable to users than the brittle performance of a fully autonomous but unreliable LLM. The strategy is to establish an ecosystem now and wait for model improvements to gradually raise the ceiling of what can be safely automated.
This article is for informational purposes only and does not constitute investment advice.