Skip to main content
Glama

See It In Action

Run it yourself: pip install entroly && entroly demo

Open the interactive HTML demo for the full animated experience, or generate your own with python docs/generate_demo.py.


The Value

Every AI coding tool — Cursor, Copilot, Claude Code, Cody — stuffs tokens into the context window until it's full, then cuts. Your AI sees 5-10 files and the rest of your codebase is invisible.

Entroly fixes this. It compresses your entire codebase into the context window at variable resolution, removes duplicates and boilerplate, and learns which context produces better AI responses over time.

You install it once. It runs invisibly. Your AI gives better answers and you spend less on tokens.

Benefit

Details

78% fewer tokens per request

Duplicate code, boilerplate, and low-information content are stripped automatically

100% codebase visibility

Every file is represented — critical files in full, supporting files as signatures, peripheral files as references

AI responses improve over time

Reinforcement learning adjusts context selection weights from session outcomes — no manual tuning

Persistent savings tracking

Lifetime cost savings survive restarts — daily/weekly/monthly trends in the dashboard

Smart auto-config

entroly go detects your project size, language, and team patterns to pick the optimal quality preset

Built-in security scanning

55 SAST rules catch hardcoded secrets, SQL injection, command injection, and 5 more CWE categories

Codebase health grades

Clone detection, dead symbol finder, god file detection — get an A-F health grade for your project

< 10ms overhead

The Rust engine adds under 10ms per request. You won't notice it

Works with any AI tool

MCP server for Cursor/Claude Code, or transparent HTTP proxy for anything else

Runs on Linux, macOS, and Windows

Native support. No WSL required on Windows. Docker optional on all platforms


Context Engineering, Automated

"The LLM is the CPU, the context window is RAM."

Today, every AI coding tool fills that RAM manually — you craft system prompts, configure RAG, curate docs. Entroly automates the entire process.

Layer

What it solves

Documentation tools

Give your agent up-to-date API docs

Memory systems

Remember things across conversations

RAG / retrieval

Find relevant code chunks

Entroly (optimization)

Makes everything fit — optimally compresses your entire codebase + docs + memory into the token budget

These layers are complementary. Doc tools give you better docs. Memory gives you persistence. RAG retrieves relevant chunks. Entroly is the optimization layer that makes sure all of it actually fits in your context window without wasting tokens.

Entroly works standalone, or on top of any doc tool, memory system, RAG pipeline, or context source.


30-Second Quickstart

pip install entroly[full]
entroly go

That's it. entroly go auto-detects your project, configures your IDE, picks the optimal quality preset, starts the proxy and dashboard — all in one command.

Then point your AI tool's API base URL to http://localhost:9377/v1.

Or step by step

pip install entroly                # install
entroly init                       # auto-detect IDE + generate config
entroly proxy --quality balanced   # start the proxy

AI Tool

Setup

Cursor

entroly init (generates .cursor/mcp.json)

Claude Code

claude mcp add entroly -- entroly

VS Code / Windsurf

entroly init (auto-detected)

Any other tool

entroly proxy → point to localhost:9377/v1

Verify it's working

entroly status     # check if the server/proxy is running
entroly demo       # before/after comparison with dollar savings on YOUR project
entroly dashboard  # live metrics: savings trends, health grade, security, PRISM weights

Install options

pip install entroly           # Core — MCP server + Python fallback engine
pip install entroly[proxy]    # Add proxy mode (transparent HTTP interception)
pip install entroly[native]   # Add native Rust engine (50-100x faster)
pip install entroly[full]     # Everything

Docker

docker pull ghcr.io/juyterman1000/entroly:latest
docker run --rm -p 9377:9377 -p 9378:9378 -v .:/workspace:ro ghcr.io/juyterman1000/entroly:latest

Multi-arch: linux/amd64 and linux/arm64 (Apple Silicon, AWS Graviton).

Or with Docker Compose: docker compose up -d


Works With Everything

AI Tool

Setup

Method

Cursor

entroly init

MCP server

Claude Code

claude mcp add entroly -- entroly

MCP server

VS Code + Copilot

entroly init

MCP server

Windsurf

entroly init

MCP server

Cline

entroly init

MCP server

OpenClaw

See below

Context Engine

Cody

entroly proxy

HTTP proxy

Any LLM API

entroly proxy

HTTP proxy


OpenClaw Integration

OpenClaw users get the deepest integration. Entroly plugs in as a Context Engine that optimizes every agent type automatically:

Agent Type

What Entroly Does

Token Savings

Main agent

Full codebase visibility at variable resolution

~78%

Heartbeat

Only loads what changed since last check

~90%

Subagents

Parent context inherited + budget-split via Nash bargaining

~70% per agent

Cron jobs

Minimal context — just relevant memories + schedule

~85%

Group chat

Entropy-based message filtering — only high-signal kept

~60%

ACP sessions

Cross-agent context sharing without duplication

~75%

When OpenClaw spawns multiple agents, Entroly's multi-agent budget allocator splits your token budget optimally across all of them. No agent starves. No tokens wasted.

from entroly.context_bridge import MultiAgentContext

ctx = MultiAgentContext(workspace_path="~/.openclaw/workspace")
ctx.ingest_workspace()

# Spawn subagents with automatic budget splitting
sub = ctx.spawn_subagent("main", "researcher", "find auth bugs")

# Schedule background checks
ctx.schedule_cron("email_checker", "check inbox", interval_seconds=900)

# Every agent gets optimized context automatically

How It Works

  1. Ingest — Indexes your codebase, builds dependency graphs, fingerprints every fragment for instant dedup

  2. Score — Ranks fragments by information density — high-value code scores high, boilerplate scores low

  3. Select — Picks the mathematically optimal subset that fits your token budget, with diversity (auth + DB + API, not 3x auth files)

  4. Deliver — 3 resolution levels: critical files in full, supporting files as signatures, peripheral files as one-line references

  5. Learn — Tracks which context produced good AI responses, improves selection weights over time


Why Not Just RAG?

Most AI tools use embedding-based retrieval (RAG). Entroly takes a fundamentally different approach:

RAG (vector search)

Entroly

Picks context by

Cosine similarity to your query

Information-theoretic optimization

Codebase coverage

Top-K similar files only

100% — every file represented at some resolution

Handles duplicates

Sends the same code 3x

SimHash dedup catches copies in O(1)

Learns from usage

No

Yes — RL updates weights from AI response quality

Dependency-aware

No

Yes — includes auth_config.py when you include auth.py

Budget optimal

Approximate (top-K)

Mathematically optimal (knapsack solver)

Needs embeddings API

Yes (cost + latency)

No — runs locally in <10ms


Platform Support

Linux

macOS

Windows

Python 3.10+

Yes

Yes

Yes

Pre-built Rust wheel

Yes

Yes (Intel + Apple Silicon)

Yes

Docker

Optional

Optional (Docker Desktop)

Optional (Docker Desktop)

WSL required

N/A

N/A

No

Admin rights required

No

No

No


CLI Commands

Command

What it does

entroly go

One command to rule them all — auto-detect, init, proxy, dashboard, smart quality

entroly init

Detects your project and AI tool, generates config

entroly proxy

Starts the invisible proxy. Point your AI tool to localhost:9377

entroly demo

Before/after comparison with dollar savings and file-level breakdown

entroly dashboard

Live metrics: lifetime savings trends, health grade, PRISM weights, security, cache

entroly doctor

Runs 7 diagnostic checks — finds problems before you do

entroly health

Codebase health grade (A-F): clones, dead code, god files, architecture violations

entroly role

Weight presets for your workflow: frontend, backend, sre, data, fullstack

entroly autotune

Auto-optimizes engine parameters using mutation-based search

entroly digest

Weekly summary of value delivered — tokens saved, cost reduction, improvements

entroly status

Check if server/proxy/dashboard are running

entroly migrate

Upgrades config and checkpoints when you update Entroly

entroly clean

Clear cached state and start fresh

entroly benchmark

Run competitive benchmark: Entroly vs raw context vs top-K retrieval

entroly completions

Generate shell completions for bash, zsh, or fish


Production Ready

Entroly is built for real-world reliability, not demos.

  • Persistent value tracking — lifetime savings stored in ~/.entroly/value_tracker.json, survives restarts, feeds dashboard trend charts

  • IDE status bar integration/confidence endpoint delivers real-time optimization confidence for VS Code widgets

  • Rich response headersX-Entroly-Confidence, X-Entroly-Coverage-Pct, X-Entroly-Cost-Saved-Today on every response

  • Connection recovery — auto-reconnects dropped connections without restarting

  • Large file protection — 500 KB ceiling prevents out-of-memory on giant logs or vendor files

  • Binary file detection — 40+ file types (images, audio, video, archives, databases) are auto-skipped

  • Crash recovery — gzipped checkpoints restore state in under 100ms

  • Cross-platform file locking — safe to run multiple instances

  • Schema migrationentroly migrate handles config upgrades between versions

  • Fragment feedbackPOST /feedback lets your AI tool rate context quality, improving future selections

  • Explainable decisionsGET /explain shows exactly why each code fragment was included or excluded


Need Help?

Self-service:

entroly doctor    # runs 7 diagnostic checks automatically
entroly --help    # see all available commands

Get support:

If you run into any issue, email autobotbugfix@gmail.com with:

  1. The output of entroly doctor

  2. A screenshot of the error

  3. Your OS (Windows/macOS/Linux) and Python version

We respond within 24 hours.

Common issues:

Homebrew Python requires a virtual environment:

python3 -m venv ~/.venvs/entroly
source ~/.venvs/entroly/bin/activate
pip install entroly[full]
python -m pip install entroly
entroly proxy --port 9378

Entroly falls back to the Python engine automatically. For the Rust speedup:

pip install entroly[native]

If no pre-built wheel exists for your platform, install the Rust toolchain first.


Part of the Ebbiforge Ecosystem

Entroly integrates with hippocampus-sharp-memory for persistent cross-session memory and Ebbiforge for TF embeddings and RL weight learning. Both are optional.


Quality Presets

Control the speed vs. quality tradeoff:

entroly proxy --quality speed       # minimal optimization, lowest latency
entroly proxy --quality fast        # light optimization
entroly proxy --quality balanced    # recommended for most projects
entroly proxy --quality quality     # deeper analysis, more context diversity
entroly proxy --quality max         # full pipeline, best results
entroly proxy --quality 0.7         # or any float from 0.0 to 1.0

Environment Variables

Variable

Default

What it does

ENTROLY_QUALITY

0.5

Quality dial (0.0-1.0 or preset name)

ENTROLY_PROXY_PORT

9377

Proxy port

ENTROLY_MAX_FILES

5000

Max files to auto-index

ENTROLY_RATE_LIMIT

0

Max requests/min (0 = unlimited)

ENTROLY_NO_DOCKER

-

Skip Docker, run natively

ENTROLY_MCP_TRANSPORT

stdio

MCP transport (stdio or sse)


How Entroly Compares

Cody / Copilot

Entroly

Approach

Embedding similarity search

Information-theoretic compression + online RL

Coverage

5-10 files (the rest is invisible)

100% codebase at variable resolution

Selection

Top-K by cosine distance

KKT-optimal bisection with submodular diversity

Dedup

None

SimHash + LSH in O(1)

Learning

Static

REINFORCE with KKT-consistent baseline

Security

None

Built-in SAST (55 rules, taint-aware)

Temperature

User-set

Self-calibrating (no tuning needed)

Architecture

Hybrid Rust + Python. All math runs in Rust via PyO3 (50-100x faster). MCP protocol and orchestration run in Python.

+-----------------------------------------------------------+
|  IDE (Cursor / Claude Code / Cline / Copilot)             |
|                                                           |
|  +---- MCP mode ----+    +---- Proxy mode ----+          |
|  | entroly MCP server|    | localhost:9377     |          |
|  | (JSON-RPC stdio)  |    | (HTTP reverse proxy)|         |
|  +--------+----------+    +--------+-----------+          |
|           |                        |                      |
|  +--------v------------------------v-----------+          |
|  |          Entroly Engine (Python)             |          |
|  |  +-------------------------------------+    |          |
|  |  |  entroly-core (Rust via PyO3)       |    |          |
|  |  |  21 modules . 380 KB . 249 tests    |    |          |
|  |  +-------------------------------------+    |          |
|  +---------------------------------------------+          |
+-----------------------------------------------------------+

Rust Core (21 modules)

Module

What

How

hierarchical.rs

3-level codebase compression

Skeleton map + dep-graph expansion + knapsack-optimal fragments

knapsack.rs

Context subset selection

KKT dual bisection O(30N) or exact 0/1 DP

knapsack_sds.rs

Information-Optimal Selection

Submodular diversity + multi-resolution knapsack

prism.rs

Weight optimizer

Spectral natural gradient on 4x4 gradient covariance

entropy.rs

Information density scoring

Shannon entropy + boilerplate detection + redundancy

depgraph.rs

Dependency graph

Auto-linking imports, type refs, function calls

skeleton.rs

Code skeleton extraction

Preserves signatures, strips bodies (60-80% reduction)

dedup.rs

Duplicate detection

64-bit SimHash, Hamming threshold 3, LSH buckets

lsh.rs

Semantic recall index

12-table multi-probe LSH, ~3 us over 100K fragments

sast.rs

Security scanning

55 rules, 8 CWE categories, taint-flow analysis

health.rs

Codebase health

Clone detection, dead symbols, god files, arch violations

guardrails.rs

Safety-critical pinning

Criticality levels with task-aware budget multipliers

query.rs

Query analysis

Vagueness scoring, keyword extraction, intent classification

query_persona.rs

Query archetype discovery

RBF kernel + Pitman-Yor process + per-archetype weights

anomaly.rs

Entropy anomaly detection

MAD-based robust Z-scores, grouped by directory

semantic_dedup.rs

Semantic redundancy removal

Greedy marginal information gain, (1-1/e) optimal

utilization.rs

Response utilization scoring

Trigram + identifier overlap feedback loop

nkbe.rs

Multi-agent budget allocation

Arrow-Debreu KKT bisection + Nash bargaining + REINFORCE

cognitive_bus.rs

Event routing for agent swarms

ISA routing, Poisson rate models, Welford spike detection

fragment.rs

Core data structure

Content, metadata, scoring dimensions, SimHash fingerprint

lib.rs

PyO3 bridge

All modules exposed to Python, 249 tests

Python Layer

Module

What

proxy.py

Invisible HTTP reverse proxy with /confidence and /trends endpoints

proxy_transform.py

Request parsing, context formatting, temperature calibration

value_tracker.py

Persistent lifetime savings tracker — daily/weekly/monthly trends, atomic writes

server.py

MCP server with 10+ tools and Python fallbacks

auto_index.py

File-system crawler for automatic codebase indexing

checkpoint.py

Gzipped JSON state serialization

prefetch.py

Predictive context pre-loading

provenance.py

Hallucination risk detection

multimodal.py

Image OCR, diagram parsing, voice transcript extraction

context_bridge.py

Multi-agent orchestration for OpenClaw (LOD, HCC, AutoTune)

MCP Tools

Tool

Purpose

remember_fragment

Store context with auto-dedup, entropy scoring, dep linking

optimize_context

Select optimal context subset for a token budget

recall_relevant

Sub-linear semantic recall via multi-probe LSH

record_outcome

Feed the reinforcement learning loop

explain_context

Per-fragment scoring breakdown

checkpoint_state

Save full session state

resume_state

Restore from checkpoint

prefetch_related

Predict and pre-load likely-needed context

get_stats

Session statistics and cost savings

health_check

Clone detection, dead symbols, god files

Novel Algorithms

Entropic Context Compression (ECC) — 3-level hierarchical codebase representation. L1: skeleton map of all files (5% budget). L2: dependency cluster expansion (25%). L3: submodular diversified full fragments (70%).

IOS (Information-Optimal Selection) — Combines Submodular Diversity Selection with Multi-Resolution Knapsack in one greedy pass. (1-1/e) optimality guarantee.

KKT-REINFORCE — The dual variable from the forward budget constraint serves as a per-item REINFORCE baseline. Forward and backward use the same probability.

PRISM — Natural gradient preconditioning via exact Jacobi eigendecomposition of the 4x4 gradient covariance.

PSM (Persona Spectral Manifold) — RBF kernel mean embedding in RKHS for automatic query archetype discovery. Each archetype learns specialized selection weights via Pitman-Yor process.

ADGT — Duality gap as a self-regulating temperature signal. No decay constant needed.

PCNT — PRISM spectral condition number as a weight-uncertainty-aware temperature modulator.

NKBE (Nash-KKT Budgetary Equilibrium) — Game-theoretic multi-agent token allocation. Arrow-Debreu KKT bisection finds the dual price, Nash bargaining ensures fairness, REINFORCE gradient learns from outcomes.

ISA Cognitive Bus — Information-Surprise-Adaptive event routing for agent swarms. Poisson rate models compute KL divergence surprise. Welford accumulators detect anomalous spikes in real-time.

References

Shannon (1948), Charikar (2002), Ebbinghaus (1885), Nemhauser-Wolsey-Fisher (1978), Sviridenko (2004), Boyd & Vandenberghe (Convex Optimization), Williams (1992), Muandet-Fukumizu-Sriperumbudur (2017), LLMLingua (EMNLP 2023), RepoFormer (ICML 2024), FILM-7B (NeurIPS 2024), CodeSage (ICLR 2024).


License

MIT

-
security - not tested
A
license - permissive license
-
quality - not tested

Resources

Unclaimed servers have limited discoverability.

Looking for Admin?

If you are the server author, to access and configure the admin panel.

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/juyterman1000/entroly'

If you have feedback or need assistance with the MCP directory API, please join our Discord server