Skip to main content
Glama
gaiaaiagent

Registry Review MCP Server

by gaiaaiagent

Registry Review MCP Server

MCP server that automates carbon credit project registration reviews through an eight-stage workflow.

Overview

The Registry Review MCP Server transforms a 6-8 hour manual document review into a guided workflow where AI handles document organization, data extraction, and consistency checking while humans provide expertise, judgment, and final approval.

Core Capabilities:

  • Document discovery and intelligent classification

  • Requirement mapping with semantic matching

  • Evidence extraction with page citations

  • Cross-document validation (dates, land tenure, project IDs)

  • Structured report generation (Markdown + JSON)

Quick Start

# Install dependencies uv sync # Run the MCP server uv run python -m registry_review_mcp.server # Run tests (expensive tests excluded by default) uv run pytest

Claude Desktop Integration

Add to your claude_desktop_config.json:

{ "mcpServers": { "registry-review": { "command": "uv", "args": [ "--directory", "/path/to/regen-registry-review-mcp", "run", "python", "-m", "registry_review_mcp.server" ] } } }

The Eight-Stage Workflow

Each stage produces artifacts for human verification before proceeding. The workflow follows a collaboration model where AI handles tedious document processing while humans provide expertise and final judgment.

Stage A: Initialize

Create a review session with project metadata and load the checklist template.

/A-initialize Botany Farm 2022-2023, /path/to/documents

Output: Session ID, project metadata, loaded checklist (23 requirements for Soil Carbon v1.2.2)

Stage B: Document Discovery

Scan the documents directory, extract file metadata, and classify each document by type.

/B-document-discovery

Agent Actions:

  • Recursively scan for PDFs, shapefiles, GeoJSON, spreadsheets

  • Classify documents (project plan, baseline report, monitoring report, land tenure, etc.)

  • Generate document inventory with confidence scores

Human Actions: Review classifications, mark documents as in-scope/ignored/pinned

Output: Document inventory with normalized names, types, and source references

Stage C: Requirement Mapping

Connect discovered documents to specific checklist requirements using semantic matching.

/C-requirement-mapping

Agent Actions:

  • Parse checklist into structured requirements with expected evidence types

  • Analyze documents and suggest requirement → document mappings

  • Flag requirements with no plausible matches

Human Actions: Confirm/reject suggested mappings, manually add missing mappings

Output: Mapping matrix linking each requirement to 0+ documents with confidence scores

Stage D: Evidence Extraction

Extract key data points and text snippets from mapped documents.

/D-evidence-extraction

Agent Actions:

  • Parse document content (PDF text, tables, metadata)

  • Extract 0-3 evidence snippets per requirement with page citations

  • Extract structured data: dates, locations, ownership info, numerical values

Human Actions: Review snippets, delete irrelevant ones, add manual notes

Output: Evidence database with snippets, citations, and structured data points

Stage E: Cross-Validation

Verify consistency, completeness, and compliance across all extracted evidence.

/E-cross-validation

Validation Checks:

  • Date Alignment: Sampling dates within ±120 days of imagery dates

  • Land Tenure: Owner names consistent across documents (fuzzy matching)

  • Project ID: Consistent project identifiers across all documents

  • Completeness: Each requirement has mapped documents with sufficient evidence

Output: Validation results with pass/warning/fail flags and coverage statistics

Stage F: Report Generation

Produce structured, auditable Registry Review Report.

/F-report-generation

Output Formats:

  • Markdown: Human-readable report with executive summary, per-requirement findings, citations

  • JSON: Machine-readable for audit trails and downstream systems

Report Contents: Project metadata, coverage statistics, requirement findings with evidence snippets, validation results, items requiring human review

Stage G: Human Review

Expert validation, annotation, and revision handling.

/G-human-review

Human Actions:

  • Review flagged items requiring judgment

  • Override agent assessments where expert knowledge differs

  • Request revisions from proponent if gaps identified

  • Make final determination: Approve / Conditional / Reject / On Hold

Output: Finalized report with human annotations and approval decision

Stage H: Completion

Finalize and archive the review.

/H-completion

Agent Actions:

  • Lock finalized report

  • Generate archive package with audit trail

  • Prepare data for on-chain registration (if approved)

Output: Locked report, complete audit trail, archived session

Quick Example

/A-initialize Botany Farm 2022-2023, /home/user/projects/botany-farm /B-document-discovery /C-requirement-mapping /D-evidence-extraction /E-cross-validation /F-report-generation

Each stage auto-selects the most recent session, so you can run them in sequence without specifying session IDs.

Available Tools

Session Management:

  • create_session - Create new review session

  • load_session / list_sessions / delete_session - Session lifecycle

  • start_review - Quick-start: create session + discover documents

  • list_example_projects - List available test projects

File Upload:

  • create_session_from_uploads - Create session from uploaded files

  • upload_additional_files - Add files to existing session

  • start_review_from_uploads - Full workflow from uploads

Document Processing:

  • discover_documents - Scan and classify project documents

  • add_documents - Add document sources to session

  • extract_pdf_text - Extract text from PDFs

  • extract_gis_metadata - Extract GIS shapefile metadata

Requirement Mapping:

  • map_all_requirements - Semantic mapping to documents

  • confirm_mapping / remove_mapping - Manual mapping adjustments

  • get_mapping_status - View mapping statistics

Evidence & Validation:

  • extract_evidence - Extract evidence for all requirements

  • map_requirement - Map and extract for single requirement

Configuration

Copy .env.example to .env and configure:

# Required for LLM-powered extraction REGISTRY_REVIEW_ANTHROPIC_API_KEY=sk-ant-api03-... REGISTRY_REVIEW_LLM_EXTRACTION_ENABLED=true # Optional REGISTRY_REVIEW_LLM_MODEL=claude-sonnet-4-5-20250929 REGISTRY_REVIEW_LOG_LEVEL=INFO

See .env.example for all configuration options including chunking, image processing, cost management, and validation settings.

Project Structure

regen-registry-review-mcp/ ├── src/registry_review_mcp/ │ ├── server.py # MCP entry point │ ├── config/ # Settings management │ ├── extractors/ # PDF and LLM extraction │ ├── models/ # Pydantic models │ ├── prompts/ # A-H workflow prompts │ ├── services/ # Document processing │ ├── tools/ # MCP tool implementations │ └── utils/ # State, cache, helpers ├── data/ │ ├── checklists/ # Methodology requirements (JSON) │ ├── sessions/ # Active sessions (gitignored) │ └── cache/ # Cached extractions (gitignored) ├── tests/ # Test suite ├── docs/ │ └── specs/ # Workflow specifications └── examples/ # Test data (Botany Farm)

Development

# Run tests (fast tests only - expensive tests excluded) uv run pytest # Format and lint uv run black src/ tests/ uv run ruff check src/ tests/

Test Markers:

  • smoke - Critical path tests (<1s)

  • expensive - Tests with API costs (excluded by default)

  • marker - PDF extraction tests (slow, 8GB+ RAM)

  • accuracy - Ground truth validation tests

See pytest.ini for marker configuration.

Requirements

  • Python >= 3.10

  • uv package manager

  • 4GB RAM minimum (8GB recommended for large PDFs)

License

Copyright 2025 Regen Network Development, Inc.

-
security - not tested
F
license - not found
-
quality - not tested

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/gaiaaiagent/regen-registry-review-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server