Skip to main content
Glama

CodeRAG

by JonnoC
.env.example1.8 kB
# Neo4j Database Configuration NEO4J_URI=bolt://localhost:7687 NEO4J_USER=neo4j NEO4J_PASSWORD=your_password # Project Configuration PROJECT_ISOLATION_STRATEGY=shared_db DEFAULT_PROJECT_ID=default CROSS_PROJECT_ANALYSIS=false MAX_PROJECTS_SHARED_DB=100 # Semantic Search Configuration # Options: openai, ollama, disabled SEMANTIC_SEARCH_PROVIDER=openai # ============================================ # OPENAI PROVIDER CONFIGURATION # ============================================ # Required if SEMANTIC_SEARCH_PROVIDER=openai OPENAI_API_KEY=sk-your-openai-api-key-here # Optional: Custom OpenAI API base URL (for LLM Studio, Azure OpenAI, etc.) # OPENAI_BASE_URL=http://localhost:1234/v1 # ============================================ # OLLAMA PROVIDER CONFIGURATION # ============================================ # Required if SEMANTIC_SEARCH_PROVIDER=ollama # OLLAMA_BASE_URL=http://localhost:11434 # ============================================ # UNIFIED MODEL CONFIGURATION # ============================================ # Model name - auto-detected defaults based on provider: # OpenAI: text-embedding-3-small, text-embedding-3-large, text-embedding-ada-002 # Ollama: nomic-embed-text, mxbai-embed-large EMBEDDING_MODEL=text-embedding-3-small # Optional: Override default dimensions (auto-detected based on model and provider) # OpenAI text-embedding-3-small: 1536, text-embedding-3-large: 3072 # Ollama nomic-embed-text: 768, mxbai-embed-large: 1024 # EMBEDDING_DIMENSIONS=1536 # ============================================ # PROCESSING CONFIGURATION # ============================================ EMBEDDING_MAX_TOKENS=8000 EMBEDDING_BATCH_SIZE=100 SIMILARITY_THRESHOLD=0.7 # MCP Response Configuration # Maximum tokens for MCP tool responses (default: 10000) MCP_TOKEN_LIMIT=10000

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/JonnoC/CodeRAG'

If you have feedback or need assistance with the MCP directory API, please join our Discord server