🇨🇳 中文 • 🇯🇵 日本語 • 🇰🇷 한국어 • 🇧🇷 Português • 🇪🇸 Español • 🇩🇪 Deutsch • 🇫🇷 Français • 🇷🇺 Русский • 🇮🇳 हिन्दी • 🇹🇷 Türkçe
Claude, Cursor, Copilot, Codex, and MiniMax only see 5% of your codebase. Entroly gives them a 2M-token brain for 90% less — a daemon that continuously self-evolves, compressing your context and dreaming up new skills with one obsession: saving more of your tokens and sharpening every answer. The first AI runtime whose learning is provably token-negative.
npm install entroly-wasm && npx entroly-wasm | Live demo →
| Without Entroly | With Entroly | |
|---|---|---|
| Files the AI sees | 5–10 | Your entire repo |
| Tokens per request | ~186,000 | 9,300 – 55,000 |
| Cost per 1K requests | ~$560 | $28 – $168 |
| Effective context window | 200K | ~2M (via variable-resolution compression) |
| Learning cost over time | Grows (tokens) | $0 — provably token-negative |
| Setup | Hours of prompt hacks | 30 seconds |
Critical files go in full. Supporting files as signatures. The rest as references. Your AI gets the whole picture. You pay for almost none of it.
Not a roadmap. This trace is from this repo's vault, right now:
[detect] gap observed → entity="auth", miss_count=3
[synthesize] StructuralSynthesizer ($0, deterministic, no LLM)
[benchmark] skill=ddb2e2969bb0 → fitness 1.0 (1 pass / 0 fail, 338 ms)
[promote] status: draft → promoted
[registry] .entroly/vault/evolution/registry.md updated
[spend] $0.0000 — invariant C_spent ≤ τ·S(t) holds
Every other self-improving agent burns tokens to learn. Entroly's evolution ledger stays at $0 because the synthesizer reads your code graph, not an LLM.
1. Token Economy — A ValueTracker measures lifetime savings S(t). The evolution budget is strictly capped:
C_spent(t) ≤ τ · S(t) (τ = 5%)
The runtime is mathematically incapable of costing more to improve than it saves.
2. Structural Induction ($0) — Before any token is touched, a deterministic synthesizer reads the AST, dependency edges, and entropy gradient of your code and emits a working tool. No LLM. No embeddings. No cloud.
3. Dreaming Loop — When idle for >60 s, the system generates synthetic queries, perturbs its scoring weights, and self-plays against benchmarks. Strict improvements are kept; regressions are discarded. You open your laptop in the morning to a smarter runtime.
npm install entroly-wasm && npx entroly-wasm
# or
pip install entroly && entroly goThat's it. It detects your IDE, wires itself into Claude/Cursor/Copilot/Codex/MiniMax, and starts compressing. Both runtimes have full parity — budget invariant, agentskills.io export, the three chat gateways, and a shared on-disk vault so skills promoted by one runtime are visible to the other.
Node:
const { VaultObserver, TelegramGateway, ValueTracker, exportAgentSkills } = require('entroly-wasm');
const obs = new VaultObserver('.entroly/vault');
new TelegramGateway({ token, chatId }).attach(obs).start();Python:
from entroly.evolution_daemon import EvolutionDaemon
from entroly.integrations.telegram_gateway import TelegramGateway
daemon.start()
TelegramGateway(token, chat_id).attach(daemon).start()The daemon is useful silently — but seeing it move is what makes it real. Three chat gateways ship in the box — Telegram, Discord, Slack. Zero extra dependencies on either runtime.
# 1. Set one (or all three) of these
export ENTROLY_TG_TOKEN=... # from @BotFather
export ENTROLY_TG_CHAT_ID=...
export ENTROLY_DISCORD_WEBHOOK=... # Discord channel → Integrations → Webhooks
export ENTROLY_SLACK_WEBHOOK=... # Slack app → Incoming WebhooksNode (native fetch, no deps):
node node_modules/entroly-wasm/js/gateways.jsPython (stdlib urllib, no deps):
python -m entroly.integrations.telegram_gateway
python -m entroly.integrations.discord_gateway
python -m entroly.integrations.slack_gatewayEvery gap detection, synthesis, promotion, and dream-cycle win streams to your chat. Telegram is 2-way — /status, /skills, /gaps, /dream.
Promoted skills aren't locked in Entroly. Export to the open agentskills.io v0.1 spec and any compatible runtime can consume them:
# Node
node node_modules/entroly-wasm/js/agentskills_export.js ./dist/agentskills
# Python
python -m entroly.integrations.agentskills ./dist/agentskillsEvery exported skill carries origin.token_cost: 0.0 — the zero-token provenance is portable too.
Claude Code • Cursor • Copilot • Codex CLI • MiniMax • Windsurf • Cody • OpenAI API • Anthropic API • LangChain • LlamaIndex • MCP-native
Architecture, benchmarks, PRISM RL internals, 3-resolution compression, provenance guarantees, RAG comparison, full API → docs/DETAILS.md
Stop paying for tokens your AI wastes. Start running an AI that teaches itself.
npm install entroly-wasm && npx entroly-wasm
Discussions • Issues • MIT License
