Skip to main content
Glama
.env.exampleβ€’4.69 kB
# Kimi K2 Coding Assistant - Environment Configuration # Copy this file to .env and fill in your actual API keys # ========================================== # REQUIRED API KEYS # ========================================== # OpenRouter API Key (Required for this configuration) # Get your API key from: https://openrouter.ai/keys # Used for: Kimi K2 model access and GPT-4o-mini evaluations OPENROUTER_API_KEY=your_openrouter_api_key_here # Firecrawl API Key (Required) # Get your API key from: https://firecrawl.dev/app/api-keys # Used for: Advanced web scraping and documentation research FIRECRAWL_API_KEY=your_firecrawl_api_key_here # ========================================== # ADDITIONAL API PROVIDERS (Optional) # ========================================== # At least one API key is required (OpenRouter is used in this config) # OpenAI API Key (Optional fallback) # Used as fallback if OpenRouter is unavailable OPENAI_API_KEY=your_openai_api_key_here # Anthropic API Key (Optional fallback) # Used as fallback if OpenRouter is unavailable ANTHROPIC_API_KEY=your_anthropic_api_key_here # Ollama Configuration (optional - only if using Ollama) # OLLAMA_BASE_URL=http://localhost:11434/v1 # Optional: OpenAI base URL (for custom endpoints) # OPENAI_BASE_URL=https://api.openai.com/v1 # ========================================== # APPLICATION SETTINGS # ========================================== NODE_ENV=development CIPHER_LOG_LEVEL=info REDACT_SECRETS=true # ========================================== # STORAGE CONFIGURATION # ========================================== STORAGE_CACHE_TYPE=in-memory # STORAGE_CACHE_HOST=localhost # STORAGE_CACHE_PORT=6379 # STORAGE_CACHE_USERNAME= # STORAGE_CACHE_PASSWORD= # STORAGE_CACHE_DATABASE=0 STORAGE_DATABASE_TYPE=in-memory # STORAGE_DATABASE_NAME=cipher.db # STORAGE_DATABASE_PATH=./data # ========================================== # MEMORY SEARCH CONFIGURATION # ========================================== # Options: knowledge, reflection, both SEARCH_MEMORY_TYPE=both # ========================================== # VECTOR STORE CONFIGURATION # ========================================== # Available types: in-memory, qdrant, milvus VECTOR_STORE_TYPE=in-memory # Qdrant configuration (only used if VECTOR_STORE_TYPE=qdrant) # VECTOR_STORE_HOST=localhost # VECTOR_STORE_PORT=6333 # VECTOR_STORE_URL=http://localhost:6333 # VECTOR_STORE_API_KEY=your-qdrant-api-key # Milvus configuration (only used if VECTOR_STORE_TYPE=milvus) # VECTOR_STORE_HOST=localhost # VECTOR_STORE_PORT=19530 # VECTOR_STORE_URL=your-milvus-cluster-endpoint # VECTOR_STORE_API_KEY=your-milvus-api-key # VECTOR_STORE_USERNAME=your-milvus-zilliz-cloud-username # VECTOR_STORE_PASSWORD=your-milvus-zilliz-cloud-password # Vector collection settings VECTOR_STORE_COLLECTION=knowledge_memory # VECTOR_STORE_COLLECTION_NAME=your_collection_name VECTOR_STORE_DIMENSION=1536 VECTOR_STORE_DISTANCE=Cosine VECTOR_STORE_ON_DISK=false # VECTOR_STORE_MAX_VECTORS=10000 # ========================================== # REFLECTION MEMORY CONFIGURATION # ========================================== REFLECTION_VECTOR_STORE_COLLECTION=reflection_memory DISABLE_REFLECTION_MEMORY=true # ========================================== # MCP CONFIGURATION # ========================================== # MCP_GLOBAL_TIMEOUT=30000 # Firecrawl Advanced Configuration (Optional) FIRECRAWL_RETRY_MAX_ATTEMPTS=3 FIRECRAWL_RETRY_INITIAL_DELAY=1000 FIRECRAWL_RETRY_MAX_DELAY=10000 FIRECRAWL_RETRY_BACKOFF_FACTOR=2 FIRECRAWL_CREDIT_WARNING_THRESHOLD=100 FIRECRAWL_CREDIT_CRITICAL_THRESHOLD=20 # ========================================== # KNOWLEDGE GRAPH CONFIGURATION (Optional) # ========================================== # NEO4J LOCAL CONFIGURATION (for local development) # Local Neo4j configuration # KNOWLEDGE_GRAPH_HOST=localhost # KNOWLEDGE_GRAPH_PORT=7687 # KNOWLEDGE_GRAPH_URI=bolt://localhost:7687 # KNOWLEDGE_GRAPH_USERNAME=neo4j # KNOWLEDGE_GRAPH_PASSWORD=your-local-password # KNOWLEDGE_GRAPH_DATABASE=neo4j # NEO4J CLOUD CONFIGURATION (for cloud deployment) # KNOWLEDGE_GRAPH_URI=neo4j+s://your-instance.databases.neo4j.io:7687 # KNOWLEDGE_GRAPH_USERNAME=neo4j # KNOWLEDGE_GRAPH_PASSWORD=your-password # KNOWLEDGE_GRAPH_DATABASE=neo4j # ========================================== # TROUBLESHOOTING # ========================================== # If you encounter issues: # 1. Verify all required API keys are set correctly # 2. Check API key permissions and quotas # 3. Ensure Node.js 18+ is installed for MCP servers # 4. Test individual MCP servers: npx -y @modelcontextprotocol/server-filesystem --help # 5. Set CIPHER_LOG_LEVEL=debug for detailed logging

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/campfirein/cipher'

If you have feedback or need assistance with the MCP directory API, please join our Discord server