Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@Registry Review MCP Serverinitialize a new review session for the Botany Farm 2023-2024 project"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
Registry Review MCP Server
MCP server that automates carbon credit project registration reviews through an eight-stage workflow.
Overview
The Registry Review MCP Server transforms a 6-8 hour manual document review into a guided workflow where AI handles document organization, data extraction, and consistency checking while humans provide expertise, judgment, and final approval.
Core Capabilities:
Document discovery and intelligent classification
Requirement mapping with semantic matching
Evidence extraction with page citations
Cross-document validation (dates, land tenure, project IDs)
Structured report generation (Markdown + JSON)
Quick Start
# Install dependencies
uv sync
# Run the MCP server
uv run python -m registry_review_mcp.server
# Run tests (expensive tests excluded by default)
uv run pytestClaude Desktop Integration
Add to your claude_desktop_config.json:
{
"mcpServers": {
"registry-review": {
"command": "uv",
"args": [
"--directory", "/path/to/regen-registry-review-mcp",
"run", "python", "-m", "registry_review_mcp.server"
]
}
}
}The Eight-Stage Workflow
Each stage produces artifacts for human verification before proceeding. The workflow follows a collaboration model where AI handles tedious document processing while humans provide expertise and final judgment.
Stage A: Initialize
Create a review session with project metadata and load the checklist template.
/A-initialize Botany Farm 2022-2023, /path/to/documentsOutput: Session ID, project metadata, loaded checklist (23 requirements for Soil Carbon v1.2.2)
Stage B: Document Discovery
Scan the documents directory, extract file metadata, and classify each document by type.
/B-document-discoveryAgent Actions:
Recursively scan for PDFs, shapefiles, GeoJSON, spreadsheets
Classify documents (project plan, baseline report, monitoring report, land tenure, etc.)
Generate document inventory with confidence scores
Human Actions: Review classifications, mark documents as in-scope/ignored/pinned
Output: Document inventory with normalized names, types, and source references
Stage C: Requirement Mapping
Connect discovered documents to specific checklist requirements using semantic matching.
/C-requirement-mappingAgent Actions:
Parse checklist into structured requirements with expected evidence types
Analyze documents and suggest requirement → document mappings
Flag requirements with no plausible matches
Human Actions: Confirm/reject suggested mappings, manually add missing mappings
Output: Mapping matrix linking each requirement to 0+ documents with confidence scores
Stage D: Evidence Extraction
Extract key data points and text snippets from mapped documents.
/D-evidence-extractionAgent Actions:
Parse document content (PDF text, tables, metadata)
Extract 0-3 evidence snippets per requirement with page citations
Extract structured data: dates, locations, ownership info, numerical values
Human Actions: Review snippets, delete irrelevant ones, add manual notes
Output: Evidence database with snippets, citations, and structured data points
Stage E: Cross-Validation
Verify consistency, completeness, and compliance across all extracted evidence.
/E-cross-validationValidation Checks:
Date Alignment: Sampling dates within ±120 days of imagery dates
Land Tenure: Owner names consistent across documents (fuzzy matching)
Project ID: Consistent project identifiers across all documents
Completeness: Each requirement has mapped documents with sufficient evidence
Output: Validation results with pass/warning/fail flags and coverage statistics
Stage F: Report Generation
Produce structured, auditable Registry Review Report.
/F-report-generationOutput Formats:
Markdown: Human-readable report with executive summary, per-requirement findings, citations
JSON: Machine-readable for audit trails and downstream systems
Report Contents: Project metadata, coverage statistics, requirement findings with evidence snippets, validation results, items requiring human review
Stage G: Human Review
Expert validation, annotation, and revision handling.
/G-human-reviewHuman Actions:
Review flagged items requiring judgment
Override agent assessments where expert knowledge differs
Request revisions from proponent if gaps identified
Make final determination: Approve / Conditional / Reject / On Hold
Output: Finalized report with human annotations and approval decision
Stage H: Completion
Finalize and archive the review.
/H-completionAgent Actions:
Lock finalized report
Generate archive package with audit trail
Prepare data for on-chain registration (if approved)
Output: Locked report, complete audit trail, archived session
Quick Example
/A-initialize Botany Farm 2022-2023, /home/user/projects/botany-farm
/B-document-discovery
/C-requirement-mapping
/D-evidence-extraction
/E-cross-validation
/F-report-generationEach stage auto-selects the most recent session, so you can run them in sequence without specifying session IDs.
Available Tools
Session Management:
create_session- Create new review sessionload_session/list_sessions/delete_session- Session lifecyclestart_review- Quick-start: create session + discover documentslist_example_projects- List available test projects
File Upload:
create_session_from_uploads- Create session from uploaded filesupload_additional_files- Add files to existing sessionstart_review_from_uploads- Full workflow from uploads
Document Processing:
discover_documents- Scan and classify project documentsadd_documents- Add document sources to sessionextract_pdf_text- Extract text from PDFsextract_gis_metadata- Extract GIS shapefile metadata
Requirement Mapping:
map_all_requirements- Semantic mapping to documentsconfirm_mapping/remove_mapping- Manual mapping adjustmentsget_mapping_status- View mapping statistics
Evidence & Validation:
extract_evidence- Extract evidence for all requirementsmap_requirement- Map and extract for single requirement
Configuration
Copy .env.example to .env and configure:
# Required for LLM-powered extraction
REGISTRY_REVIEW_ANTHROPIC_API_KEY=sk-ant-api03-...
REGISTRY_REVIEW_LLM_EXTRACTION_ENABLED=true
# Optional
REGISTRY_REVIEW_LLM_MODEL=claude-sonnet-4-5-20250929
REGISTRY_REVIEW_LOG_LEVEL=INFOSee .env.example for all configuration options including chunking, image processing, cost management, and validation settings.
Project Structure
regen-registry-review-mcp/
├── src/registry_review_mcp/
│ ├── server.py # MCP entry point
│ ├── config/ # Settings management
│ ├── extractors/ # PDF and LLM extraction
│ ├── models/ # Pydantic models
│ ├── prompts/ # A-H workflow prompts
│ ├── services/ # Document processing
│ ├── tools/ # MCP tool implementations
│ └── utils/ # State, cache, helpers
├── data/
│ ├── checklists/ # Methodology requirements (JSON)
│ ├── sessions/ # Active sessions (gitignored)
│ └── cache/ # Cached extractions (gitignored)
├── tests/ # Test suite
├── docs/
│ └── specs/ # Workflow specifications
└── examples/ # Test data (Botany Farm)Development
# Run tests (fast tests only - expensive tests excluded)
uv run pytest
# Format and lint
uv run black src/ tests/
uv run ruff check src/ tests/Test Markers:
smoke- Critical path tests (<1s)expensive- Tests with API costs (excluded by default)marker- PDF extraction tests (slow, 8GB+ RAM)accuracy- Ground truth validation tests
See pytest.ini for marker configuration.
Requirements
Python >= 3.10
uv package manager
4GB RAM minimum (8GB recommended for large PDFs)
License
Copyright 2025 Regen Network Development, Inc.
This server cannot be installed
Resources
Looking for Admin?
Admins can modify the Dockerfile, update the server description, and track usage metrics. If you are the server author, to access the admin panel.