Skip to main content
Glama

A swarm intelligence engine that rehearses the future.

Feed it a document. Describe a scenario. Watch hundreds of AI agents with distinct personalities, memories, and social instincts interact — and return with a prediction.

License npm Docker Website


What It Does

DeepMiro extracts entities and relationships from any document — a policy draft, a market report, a chapter of a novel — and constructs a parallel digital world. Inside it, hundreds of autonomous agents form opinions, argue on simulated social platforms, shift allegiances, and produce emergent behavior that no single prompt could predict.

You get back a structured prediction report and a living world you can interrogate, agent by agent.

Input: A PDF and a question in plain language. Output: A detailed prediction report + an interactive simulation you can explore.

How It Works

Document ──► Entity Extraction ──► Agent Generation ──► Dual-Platform Simulation ──► Prediction Report
              (NER + GraphRAG)    (personas, memory,     (Twitter-like + Reddit-like     (ReportAgent with
                                   social networks)       parallel interaction)            deep analysis tools)

Phase

What happens

Graph Build

Extracts entities, relationships, and context from your documents. Builds a knowledge graph via GraphRAG.

Environment Setup

Generates agent personas with distinct personalities, beliefs, and social connections.

Simulation

Agents interact across dual platforms (Twitter-like and Reddit-like) in parallel. Dynamic memory updates each round.

Report Generation

A ReportAgent analyzes the post-simulation environment — sentiment shifts, faction formation, viral dynamics, outcome trajectories.

Deep Interaction

Chat with any agent to understand their reasoning. Query the ReportAgent for follow-up analysis.

Quick Start

1. Get an API key

Sign up at deepmiro.org → Dashboard → API Keys. Your key looks like dm_xxxxxxxxx.

2. Install

Claude Code (plugin — recommended) — one command gets you the /predict skill + the MCP server wired up:

claude plugin marketplace add kakarot-dev/deepmiro
claude plugin install deepmiro@deepmiro-marketplace
export DEEPMIRO_API_KEY=dm_your_key   # or set it in ~/.claude/settings.json

Then restart Claude Code and say /predict or predict how people will react to [scenario].

Other clients:

Client

Install

OpenAI Codex

codex plugin install kakarot-dev/deepmiro

Claude Desktop

Add to claude_desktop_config.json: "deepmiro": {"command": "npx", "args": ["-y", "deepmiro-mcp"], "env": {"DEEPMIRO_API_KEY": "dm_xxx"}}

ChatGPT Desktop

Settings → MCP Servers → Add → npx deepmiro-mcp with env DEEPMIRO_API_KEY

Cursor / Windsurf

Settings → MCP → Add → npx deepmiro-mcp with env DEEPMIRO_API_KEY

VS Code (Copilot)

Add to .vscode/mcp.json: "deepmiro": {"command": "npx", "args": ["-y", "deepmiro-mcp"], "env": {"DEEPMIRO_API_KEY": "dm_xxx"}}

Self-host

No API key needed. Run the engine locally and point the MCP server at it:

git clone https://github.com/kakarot-dev/deepmiro.git
cd deepmiro
cp .env.example .env    # add your LLM API key
docker compose -f docker/docker-compose.yml up -d

# Connect your AI client to the local engine
claude mcp add deepmiro -e MIROFISH_URL=http://localhost:5001 -- npx -y deepmiro-mcp
# Required in .env
LLM_API_KEY=your_key
LLM_BASE_URL=https://api.openai.com/v1
LLM_MODEL_NAME=gpt-4o-mini

SURREALDB_URL=ws://localhost:8000/rpc
SURREALDB_USER=root
SURREALDB_PASS=root

MCP Server

DeepMiro is an MCP server. MCP is the universal standard adopted by Claude, ChatGPT, Gemini, Cursor, VS Code, and every major AI client — one server, works everywhere.

npx deepmiro-mcp

Available tools: create_simulation, quick_predict, simulation_status, get_report, interview_agent, upload_document, list_simulations, search_simulations.

What's Different

DeepMiro is a performance-focused fork of the original MiroFish engine. Same OASIS simulation core, rebuilt infrastructure:

Component

MiroFish (original)

DeepMiro

Recommendation engine

Full LLM call every round (~200s/round)

Cached TWHIN-BERT embeddings (~15ms/round)

Entity extraction

Sequential NER

5-worker parallel NER via ThreadPoolExecutor

Graph build time

~5 minutes

~56 seconds

Graph database

Zep Cloud (proprietary)

SurrealDB (self-hosted, open-source)

Vector search

Cloud-dependent

Hybrid HNSW + BM25 (local, 768-dim cosine)

Embedding model

Tied to Zep

nomic-embed-text-v1.5 via Fireworks (swappable)

Document ingestion

Manual text input

Upload endpoint with magic-byte validation (PDF, MD, TXT)

LLM provider

Alibaba Qwen (hardcoded)

Any OpenAI-compatible API

Deployment

Docker only

Docker + Helm chart + k3s-ready

Persona Fidelity: How DeepMiro Keeps Agents In Character

Multi-agent LLM simulations have a dirty secret: personas drift. By round 20, Tucker Carlson starts quoting the ACLU. By round 45, Marco Rubio sounds like Bernie Sanders. Every distinct voice collapses into the same bland "helpful assistant" register.

This isn't a prompting problem — it's an attention decay problem. Kim et al. (COLM 2024) proved that LLM attention to system-prompt tokens decays geometrically over turns. LLaMA2-70B drifts significantly within 8 turns. Larger models drift more, not less. A 2KB persona cannot compete with 50KB of accumulated conversation history.

Every naive multi-agent simulation hits this wall. DeepMiro doesn't, because we copied what Stanford's Generative Agents (Park et al. 2023) did for their 25-agent Smallville simulation — with some practical shortcuts.

What we do

1. Structured personas with explicit negative examples. Every agent gets a structured profile alongside the prose bio:

  • ideology_anchor — a 2-5 word partisan tag ("conservative populist", "progressive labor")

  • core_beliefs — 3-5 first-person declarative statements, no hedging

  • verbal_tics — 3-5 literal phrases the person actually uses

  • never_say — 3-5 sentences the person would refuse to utter

  • speaking_style — register + rhetorical habits

The never_say block is the drift killer. Models drift toward the centroid of what they say. Explicit negative examples ("Tucker Carlson would never say 'I stand with the ACLU'") anchor the LLM against that collapse.

2. Dynamic persona regeneration per round. Instead of locking the persona in at the system-prompt level and watching attention decay from round 1, we rebuild system_message.content before every agent acts. Each round, the agent sees a fresh third-person character brief:

# Character Brief: Tucker Carlson

The agent in this conversation is Tucker Carlson.
You are simulating how Tucker Carlson would respond.

## What Tucker Carlson Would NEVER Say
- "I stand with the ACLU"
- "We need to find common ground with progressives"
...

## What Tucker Carlson Has Said Recently
- "Permanent Washington wants you to believe..."
- "Let's pause for a moment — they're not even hiding it"
...

## Task
What would Tucker Carlson actually do? React in his authentic voice.
Do not become a neutral assistant. Do not seek balance.

The persona never gets stale because it's built fresh from the same structured fields every turn.

3. Third-person framing. "You are Tucker Carlson" triggers RLHF helpful-assistant sycophancy — the model tries to be polite and balanced because that's how it was trained to respond to "you are X" instructions. Third-person framing ("the agent is Tucker Carlson", "what would Tucker Carlson do?") bypasses that trigger entirely. This single change is load-bearing.

4. Self-consistency anchor. Each round injects the agent's own 3 most recent posts as reference material. Tucker Carlson sees what he just said, which makes him more likely to say something consistent with it. This is cheap drift resistance — no extra LLM calls, just reading from the action log.

5. No accumulated chat history. Unlike naive multi-agent setups, DeepMiro does NOT feed each agent the rolling conversation history from previous rounds. Agents get their fresh persona + the current feed observations. Attention stays focused on character + present context, not on 50KB of stale noise.

What we don't do

  • We don't script reactions. Agents aren't told "mock liberal content" or "support conservative content" — that would script the outcome and destroy the simulation's predictive value. The emergent behavior is the whole point.

  • We don't filter feeds by ideology. Tucker Carlson sees AOC's posts. That's how he has something to push back against. Echo chambers are not simulations.

  • We don't fork OASIS. The entire fix is a runtime wrapper around CAMEL's agent pager. No upstream drift, no fork maintenance.

Research foundations

Technique

Source

Attention decay over system prompts

Kim et al. — Measuring and Controlling Persona Drift (COLM 2024)

Third-person framing bypasses RLHF sycophancy

Park et al. — Generative Agents (Stanford 2023)

Negative examples > positive instruction

Examining Identity Drift in LLM Agents (arXiv 2412.00804)

Dynamic persona summary per action

Park et al. — Generative Agents (Stanford 2023)

JSON personas collapse to neutral register

Persona-Aware Contrastive Learning (ACL 2025)

Benchmarks

15-agent quick simulation, enriched prompt, measured end-to-end:

Stage

Time

Graph build

~10s

Agent generation

~3 min

Simulation (110 Twitter + 26 Reddit actions)

~4 min

Total pipeline

~7 min (quick) / ~12 min (standard, 80 agents)

The biggest win is the recommendation system: TWHIN-BERT embeddings are computed once per user at setup, then only new posts are embedded incrementally each round. Cosine similarity via numpy replaces what was previously a full LLM inference call — 13,000x faster per round.

Monorepo Structure

deepmiro/
├── engine/              # Python Flask simulation backend
│   ├── app/
│   │   ├── api/         # REST endpoints (simulation, graph, documents, report)
│   │   ├── services/    # Graph builder, simulation runner, report agent
│   │   ├── storage/     # SurrealDB adapter, embedding service, NER
│   │   └── utils/       # LLM client, retry logic, logging
│   └── pyproject.toml
├── mcp-server/          # TypeScript MCP server (npm: deepmiro-mcp)
│   └── src/
├── .claude-plugin/      # Claude Code plugin + marketplace manifests
├── .codex-plugin/       # OpenAI Codex plugin manifest
├── .agents/             # Codex marketplace catalog
├── .mcp.json            # MCP config (auto-loaded when running `claude` here)
├── skills/predict/      # /predict skill (auto-setup, narration, interviews)
├── helm-chart/          # Kubernetes (k3s) deployment
├── docker/              # Dockerfiles + compose
├── docs/                # Landing page
└── locales/             # i18n (en, zh)

Use Cases

Domain

Example

Market analysis

Upload an earnings report. "How will retail investors react to this guidance revision?"

Policy testing

Upload a draft regulation. "What public backlash should we expect, and from which demographics?"

PR & comms

Upload a press release. "How will this announcement play on social media over 48 hours?"

Competitive analysis

Upload competitor product specs. "How will our user base respond to this feature gap?"

Creative exploration

Upload a novel's first 80 chapters. "What ending would emerge from these character dynamics?"

Crisis simulation

Upload an incident report. "How does public opinion evolve if we respond with X vs Y?"

Acknowledgments

DeepMiro is a fork of MiroFish, originally created by Guo Hangjiang and supported by Shanda Group. The simulation layer is powered by OASIS from the CAMEL-AI team.

License

AGPL-3.0


deepmiro.org · Built by Joel Libni

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/kakarot-dev/deepmiro'

If you have feedback or need assistance with the MCP directory API, please join our Discord server