Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@MCP4DRLShow the recommended action and explain why the agent chose it"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
MCP4DRL - Model Context Protocol for Deep Reinforcement Learning
MCP server that exposes a trained Deep Q-Network (DQN) agent for business process resource allocation through conversational interfaces. Makes "black box" RL systems transparent via natural language queries.
Features
Environment State Queries - View simulation state, waiting/active cases, resources
Q-Value Analysis - Inspect Q-values for all actions
Action Recommendations - Get agent's top choice with justification
Explainability - Detailed explanations of why actions are chosen
Heuristic Comparison - Compare with FIFO, SPT, EDF, LST baselines
Simulation Control - Step through episodes, reset, run full episodes
Installation
Requirements: Python 3.8+, TensorFlow 2.16+
Quick Start
Test locally
Run MCP server
Claude Desktop Integration
Add to claude_desktop_config.json:
Available MCP Tools
Tool | Description |
| Current simulation state |
| All possible actions with validity |
| Q-values for all actions |
| Agent's best action |
| Detailed action explanation |
| Compare with FIFO/SPT/EDF/LST |
| Execute one step |
| Reset to initial state |
| Run full episode with policy |
Project Structure
Configuration
Environment variables (optional):
MCP4DRL_MODEL_PATH- Path to trained model (.h5)MCP4DRL_EVENT_LOG- Path to XES event logMCP4DRL_TRANSPORT-stdio(default) orsse
Context
Part of doctoral dissertation on intelligent automation of business process management. Demonstrates that RL systems can be made transparent through conversational interfaces.
License
Research prototype.