.env.example•1.8 kB
# Neo4j Database Configuration
NEO4J_URI=bolt://localhost:7687
NEO4J_USER=neo4j
NEO4J_PASSWORD=your_password
# Project Configuration
PROJECT_ISOLATION_STRATEGY=shared_db
DEFAULT_PROJECT_ID=default
CROSS_PROJECT_ANALYSIS=false
MAX_PROJECTS_SHARED_DB=100
# Semantic Search Configuration
# Options: openai, ollama, disabled
SEMANTIC_SEARCH_PROVIDER=openai
# ============================================
# OPENAI PROVIDER CONFIGURATION
# ============================================
# Required if SEMANTIC_SEARCH_PROVIDER=openai
OPENAI_API_KEY=sk-your-openai-api-key-here
# Optional: Custom OpenAI API base URL (for LLM Studio, Azure OpenAI, etc.)
# OPENAI_BASE_URL=http://localhost:1234/v1
# ============================================
# OLLAMA PROVIDER CONFIGURATION
# ============================================
# Required if SEMANTIC_SEARCH_PROVIDER=ollama
# OLLAMA_BASE_URL=http://localhost:11434
# ============================================
# UNIFIED MODEL CONFIGURATION
# ============================================
# Model name - auto-detected defaults based on provider:
# OpenAI: text-embedding-3-small, text-embedding-3-large, text-embedding-ada-002
# Ollama: nomic-embed-text, mxbai-embed-large
EMBEDDING_MODEL=text-embedding-3-small
# Optional: Override default dimensions (auto-detected based on model and provider)
# OpenAI text-embedding-3-small: 1536, text-embedding-3-large: 3072
# Ollama nomic-embed-text: 768, mxbai-embed-large: 1024
# EMBEDDING_DIMENSIONS=1536
# ============================================
# PROCESSING CONFIGURATION
# ============================================
EMBEDDING_MAX_TOKENS=8000
EMBEDDING_BATCH_SIZE=100
SIMILARITY_THRESHOLD=0.7
# MCP Response Configuration
# Maximum tokens for MCP tool responses (default: 10000)
MCP_TOKEN_LIMIT=10000