Ingests code repositories for analysis, enabling repository structure mapping, dependency tracking, and knowledge graph construction from GitHub-hosted codebases.
Stores and queries code repositories, project knowledge, and documentation as a graph database with vector search capabilities for intelligent code analysis and knowledge management.
Provides LLM inference and text embeddings for code analysis, knowledge queries, and document processing through locally-hosted language models.
Provides LLM inference and text embeddings for code analysis, knowledge queries, and document processing through OpenAI's API.
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@Code Graph Knowledge Systemanalyze dependencies for the authentication module in our main repository"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
Code Graph Knowledge System
Enterprise knowledge management platform with Neo4j graph database, multi-interface architecture (MCP/Web/REST), and intelligent code analysis for modern software development teams.
Overview
Code Graph Knowledge System is a production-ready platform that transforms code repositories and development documentation into a queryable knowledge graph. Built on Neo4j's graph database technology and powered by large language models, the system provides three distinct interfaces for different use cases: MCP protocol for AI assistants, Web UI for human users, and REST API for programmatic access.
The platform combines vector search, graph traversal, and LLM-driven analysis to deliver intelligent code intelligence capabilities including repository analysis, dependency mapping, impact assessment, and automated documentation generation.
Core Capabilities
Multi-Interface Architecture
MCP Protocol (Port 8000) - Model Context Protocol server for AI assistant integration
Direct integration with Claude Desktop, Cursor, and other MCP-compatible tools
25+ specialized tools for code analysis and knowledge management
Real-time task monitoring via Server-Sent Events
Supports stdio and SSE transport modes
Web UI (Port 8080) - Browser-based interface for team collaboration
Real-time task monitoring dashboard
Repository ingestion and management
Metrics visualization with interactive charts
Built with React 18, TypeScript, and shadcn/ui components
REST API (Ports 8000, 8080) - HTTP endpoints for system integration
Document ingestion and knowledge querying
Task management and monitoring
Prometheus metrics export
OpenAPI/Swagger documentation
Knowledge Graph Engine
Code Intelligence - Graph-based code analysis without requiring LLMs
Repository structure mapping and dependency tracking
Function and class relationship analysis
Impact analysis for code changes
Context pack generation for AI assistants
Support for 15+ programming languages
Memory Store - Project knowledge tracking with temporal awareness
Fact, decision, pattern, and insight recording
Memory evolution with superseding relationships
Automatic extraction from conversations, commits, and code
Vector search with embedding-based retrieval
Knowledge RAG - Document processing with hybrid search
Multi-format document ingestion (Markdown, PDF, code files)
Neo4j native vector indexing
Hybrid search combining vector similarity and graph traversal
Configurable chunking and embedding strategies
SQL Schema Parser - Database schema analysis with business domain classification
Multi-dialect support (Oracle, MySQL, PostgreSQL, SQL Server)
Configurable business domain templates (Insurance, E-commerce, Banking, Healthcare)
Automated relationship detection and documentation generation
Integration with knowledge graph for cross-referencing
Technology Stack
Backend Infrastructure
FastAPI - High-performance async web framework
Neo4j 5.x - Graph database with native vector indexing
Python 3.13+ - Modern Python with type hints
Uvicorn - ASGI server with WebSocket support
AI and ML Integration
LlamaIndex - Document processing and retrieval pipeline
Multiple LLM providers (Ollama, OpenAI, Gemini, OpenRouter)
Flexible embedding models (HuggingFace, Ollama, OpenAI)
Model Context Protocol (MCP) for AI assistant integration
Frontend Technology
React 18 - Modern UI library with concurrent features
TypeScript - Type-safe development
TanStack Router - Type-safe routing
shadcn/ui - Accessible component library
Vite - Fast build tooling
Quick Start
Prerequisites
Python 3.13 or higher
Neo4j 5.0 or higher
Docker (optional, for containerized deployment)
Node.js 18+ (for frontend development)
Querying Knowledge
Installation
Clone the repository and install dependencies:
Configure environment variables:
Start Neo4j database:
Running the System
Complete System (MCP + Web UI + REST API)
Access points:
MCP SSE Service:
http://localhost:8000/sseWeb UI:
http://localhost:8080REST API Documentation:
http://localhost:8080/docsPrometheus Metrics:
http://localhost:8080/metrics
MCP Server Only
Docker Deployment
Three deployment modes available:
Minimal Mode - Code Graph only (no LLM required)
Standard Mode - Code Graph + Memory Store (embedding model required)
Full Mode - All features (LLM + embedding required)
Usage Examples
MCP Integration
Configure in Claude Desktop or compatible MCP client:
Available MCP tools include:
code_graph_ingest_repo- Ingest code repositorycode_graph_related- Find related code elementscode_graph_impact- Analyze change impactquery_knowledge- Query knowledge baseadd_memory- Store project knowledgeextract_from_conversation- Extract insights from chatwatch_task- Monitor task progress
REST API
Ingest a repository:
Query knowledge base:
Monitor tasks:
Web UI
Navigate to http://localhost:8080 to access:
Dashboard - System health and quick actions
Tasks - Real-time task monitoring with progress indicators
Repositories - Repository management and ingestion
Metrics - System performance and usage metrics
Configuration
Key environment variables:
For complete configuration options, see Configuration Guide.
Architecture
Dual-Server Design
The system employs a dual-server architecture optimized for different access patterns:
Port 8000 (Primary) - MCP SSE Service
Server-Sent Events endpoint for real-time communication
Optimized for AI assistant integration
Handles long-running task monitoring
WebSocket support for bidirectional communication
Port 8080 (Secondary) - Web UI + REST API
React-based monitoring interface
RESTful API for external integrations
Prometheus metrics endpoint
Static file serving for frontend
Both servers share the same backend services and Neo4j database, ensuring consistency across all interfaces.
Component Architecture
Development
Project Structure
Running Tests
Code Quality
Frontend Development
Deployment
Production Deployment
See Docker Deployment Guide for production deployment configurations including:
Multi-stage Docker builds
Environment-specific configurations
Scaling and load balancing
Security best practices
Monitoring and logging setup
System Requirements
Minimum Configuration
CPU: 2 cores
RAM: 4 GB
Storage: 10 GB
Recommended Configuration
CPU: 4+ cores
RAM: 8+ GB
Storage: 50+ GB SSD
Network: 100 Mbps+
Documentation
Complete documentation available at https://vantagecraft.dev/docs/code-graph
Key Documentation Sections
Quick Start Guide - Get up and running in 5 minutes
Architecture Overview - System design and components
MCP Integration - AI assistant integration
REST API Reference - Complete API documentation
Deployment Guide - Production deployment
Development Guide - Contributing and development
Community and Support
Documentation: Complete Documentation
Neo4j Guide: README_Neo4j.md
Issues: GitHub Issues
Discussions: GitHub Discussions
License
This project is licensed under the MIT License - see the LICENSE file for details.
Acknowledgments
Built with excellent open source technologies:
Neo4j - Graph database platform
LlamaIndex - Data framework for LLM applications
FastAPI - Modern web framework for Python
React - Library for building user interfaces
Model Context Protocol - AI assistant integration standard