BotBeat
...
← Back

> ▌

EntrolyEntroly
PRODUCT LAUNCHEntroly2026-04-14

Entroly Launches Token-Negative AI Runtime: Compresses 586K Tokens to 5K With 100% Accuracy

Key Takeaways

  • ▸Entroly reduces AI context token consumption by 90%, shrinking per-request tokens from ~186,000 to 9,300–55,000 while maintaining full codebase visibility
  • ▸The system is provably 'token-negative'—learning improvements cost $0 through deterministic AST-based synthesis, violating no token budget invariants
  • ▸Setup takes 30 seconds (npm/pip install); the daemon self-evolves during idle periods, promoting skills across multiple runtimes (Claude, Cursor, Copilot, MiniMax, Codex)
Source:
Hacker Newshttps://github.com/juyterman1000/entroly↗

Summary

Entroly has launched a new AI runtime daemon that dramatically reduces token consumption for AI coding assistants like Claude, Cursor, Copilot, and others. The system compresses entire codebases from ~186,000 tokens per request down to 9,300–55,000 tokens while maintaining full context awareness, reducing costs by up to 90% and expanding effective context windows from 200K to ~2M tokens.

The innovation relies on three core pillars: a token economy with strict mathematical caps on learning costs, deterministic structural induction that synthesizes solutions without LLM calls, and a self-improvement dreaming loop that runs idle. The system is mathematically guaranteed to be "token-negative"—it saves more tokens than it consumes during learning, with improvement costs strictly bounded at 5% of cumulative savings.

Entroly is available now via npm and pip, integrating seamlessly with existing AI workflows in 30 seconds. The runtime includes live monitoring dashboards via Telegram, Discord, and Slack, allowing developers to watch the daemon detect gaps, synthesize new skills, and self-evolve in real-time with zero additional learning costs.

  • Live monitoring via Telegram, Discord, and Slack shows real-time gap detection, skill synthesis, benchmarking, and promotion events as they occur

Editorial Opinion

Entroly presents a genuinely novel approach to the token-efficiency problem plaguing AI-assisted development: instead of relying on embeddings or LLM-based compression, it performs deterministic code-graph analysis to build working tools before touching a single token. The claim of mathematical impossibility for learning to exceed savings—enforced by a hard invariant—is a refreshing counterpoint to black-box fine-tuning models. If the self-evolution loop delivers on its promise of consistent improvement with zero cost, this could reshape how developers think about AI runtime efficiency.

Large Language Models (LLMs)AI AgentsMachine LearningMLOps & Infrastructure

More from Entroly

EntrolyEntroly
PRODUCT LAUNCH

Entroly Launches Context Engineering Engine for AI Coding Agents, Reduces Token Usage by 78%

2026-04-03

Comments

Suggested

OpenAIOpenAI
RESEARCH

OpenAI's GPT-5.4 Pro Solves Longstanding Erdős Math Problem, Reveals Novel Mathematical Connections

2026-04-17
AnthropicAnthropic
PARTNERSHIP

White House Pushes US Agencies to Adopt Anthropic's AI Technology

2026-04-17
AnthropicAnthropic
RESEARCH

AI Safety Convergence: Three Major Players Deploy Agent Governance Systems Within Weeks

2026-04-17
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us