Skip to main content
Glama
fraim-ai

Fraim Context MCP

Official
by fraim-ai

Fraim Context MCP

Semantic search MCP server for project documentation.

Version: 5.1.0
Status: In Development

Overview

Fraim Context MCP exposes project documentation to LLMs via the Model Context Protocol (MCP). It supports:

  • Fast mode: Direct cache/search for immediate results

  • Deep mode: Multi-round synthesis for complex queries

  • Hybrid search: Vector similarity + full-text search with pgvector

  • Smart caching: Redis with corpus versioning for cache invalidation

Quick Start

# 1. Setup Doppler
doppler login
doppler setup  # Select: fraim-context → dev

# 2. Install dependencies
uv sync

# 3. Verify environment
doppler run -- uv run python scripts/verify_env.py

# 4. Run tests
doppler run -- uv run pytest tests/stage_0/ -v

Development

This project uses Test-Driven Development (TDD). See DNA/DEVELOPMENT_PLAN.md for stages.

# Run all tests
doppler run -- uv run pytest tests/ -v

# Run specific stage
doppler run -- uv run pytest tests/stage_0/ -v

# Lint
uv run ruff check src/ tests/

# Type check
uv run mypy src/fraim_mcp

Architecture

  • LLM Access: Pydantic AI Gateway (unified key for all providers)

  • Database: PostgreSQL + pgvector (1024-dim embeddings)

  • Cache: Redis 7.x (native asyncio)

  • Observability: Logfire (OpenTelemetry)

See DNA/specs/ARCHITECTURE.md for full details.

License

MIT

-
security - not tested
F
license - not found
-
quality - not tested

Resources

Unclaimed servers have limited discoverability.

Looking for Admin?

If you are the server author, to access and configure the admin panel.

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/fraim-ai/Fraim-Context-MCP'

If you have feedback or need assistance with the MCP directory API, please join our Discord server