Skip to main content
Glama

cognee-mcp

.env.template10.7 kB
############################################################################### # NOTE: With default settings Cognee only needs an OpenAI LLM_API_KEY to be set. # The rest of the settings don't have to be set. # Default relational database: SQLite # Default vector database : LanceDB # Default graph database : Kuzu # # These default databases are all file-based, so no extra setup is needed # for local use. The data by default will be stored in your .venv ############################################################################### ################################################################################ # 🧠 LLM Settings ################################################################################ # Currently we support BAML and Instructor(using litellm) for structured outputs STRUCTURED_OUTPUT_FRAMEWORK="instructor" LLM_API_KEY="your_api_key" LLM_MODEL="openai/gpt-4o-mini" LLM_PROVIDER="openai" LLM_ENDPOINT="" LLM_API_VERSION="" LLM_MAX_TOKENS="16384" EMBEDDING_PROVIDER="openai" EMBEDDING_MODEL="openai/text-embedding-3-large" EMBEDDING_ENDPOINT="" EMBEDDING_API_VERSION="" EMBEDDING_DIMENSIONS=3072 EMBEDDING_MAX_TOKENS=8191 # If embedding key is not provided same key set for LLM_API_KEY will be used #EMBEDDING_API_KEY="your_api_key" # If using BAML structured output these env variables will be used BAML_LLM_PROVIDER=openai BAML_LLM_MODEL="gpt-4o-mini" BAML_LLM_ENDPOINT="" BAML_LLM_API_KEY="your_api_key" BAML_LLM_API_VERSION="" ################################################################################ # 📂 ROOT DIRECTORY FOR DATABASES ################################################################################ # Set up the Cognee system directory. Cognee will store system files and databases here. # Useful for setting root directory inside docker and also to avoid storing the databases in .venv # DATA_ROOT_DIRECTORY='/Users/<user>/Desktop/cognee/.cognee_data/' # SYSTEM_ROOT_DIRECTORY='/Users/<user>/Desktop/cognee/.cognee_system/' ################################################################################ # ☁️ Storage Backend Settings ################################################################################ # Configure storage backend (local filesystem or S3) # STORAGE_BACKEND="local" # Default: uses local filesystem # # -- To switch to S3 storage, uncomment and fill these: --------------------- # STORAGE_BACKEND="s3" # STORAGE_BUCKET_NAME="your-bucket-name" # AWS_REGION="us-east-1" # AWS_ACCESS_KEY_ID="your-access-key" # AWS_SECRET_ACCESS_KEY="your-secret-key" # # -- S3 Root Directories (optional) ----------------------------------------- # DATA_ROOT_DIRECTORY="s3://your-bucket/cognee/data" # SYSTEM_ROOT_DIRECTORY="s3://your-bucket/cognee/system" # # -- Cache Directory (auto-configured for S3) ------------------------------- # When STORAGE_BACKEND=s3, cache automatically uses S3: s3://BUCKET/cognee/cache # To override the automatic S3 cache location, uncomment: # CACHE_ROOT_DIRECTORY="s3://your-bucket/cognee/cache" ################################################################################ # 🗄️ Relational database settings ################################################################################ DB_PROVIDER="sqlite" DB_NAME=cognee_db # -- To switch to Postgres / PGVector, uncomment and fill these: ------------- #DB_PROVIDER=postgres #DB_NAME=cognee_db # To use Postgres with the Cognee backend in Docker compose use the following instead: DB_HOST=host.docker.internal #DB_HOST=127.0.0.1 #DB_PORT=5432 #DB_USERNAME=cognee #DB_PASSWORD=cognee ################################################################################ # 🕸️ Graph Database settings ################################################################################ # Default (local file-based) GRAPH_DATABASE_PROVIDER="kuzu" # -- To switch to Remote Kuzu uncomment and fill these: ------------------------------------------------------------- #GRAPH_DATABASE_PROVIDER="kuzu" #GRAPH_DATABASE_PROVIDER="kuzu-remote" #GRAPH_DATABASE_URL="http://localhost:8000" #GRAPH_DATABASE_USERNAME=XXX #GRAPH_DATABASE_PASSWORD=YYY # -- To switch to Neo4j uncomment and fill these: ------------------------------------------------------------------- #GRAPH_DATABASE_PROVIDER="neo4j" #GRAPH_DATABASE_URL=bolt://localhost:7687 #GRAPH_DATABASE_NAME="neo4j" #GRAPH_DATABASE_USERNAME=neo4j #GRAPH_DATABASE_PASSWORD=localneo4j ################################################################################ # 📐 Vector Database settings ################################################################################ # Supported providers: pgvector | qdrant | weaviate | milvus | lancedb | chromadb VECTOR_DB_PROVIDER="lancedb" # Not needed if a cloud vector database is not used VECTOR_DB_URL= VECTOR_DB_KEY= ################################################################################ # 🧩 Ontology resolver settings ################################################################################ # -- Ontology resolver params -------------------------------------- # ONTOLOGY_RESOLVER=rdflib # Default: uses rdflib and owl file to read ontology structures # MATCHING_STRATEGY=fuzzy # Default: uses fuzzy matching with 80% similarity threshold # ONTOLOGY_FILE_PATH=YOUR_FULL_FULE_PATH # Default: empty # To add ontology resolvers, either set them as it is set in ontology_example or add full_path and settings as envs. ################################################################################ # 🔄 MIGRATION (RELATIONAL → GRAPH) SETTINGS ################################################################################ MIGRATION_DB_PATH="/path/to/migration/directory" MIGRATION_DB_NAME="migration_database.sqlite" MIGRATION_DB_PROVIDER="sqlite" # -- Postgres-specific migration params -------------------------------------- # MIGRATION_DB_USERNAME=cognee # MIGRATION_DB_PASSWORD=cognee # MIGRATION_DB_HOST="127.0.0.1" # MIGRATION_DB_PORT=5432 ################################################################################ # 🔒 Security Settings ################################################################################ # When set to false don't allow adding of local system files to Cognee. Should be set to False when Cognee is used as a backend. ACCEPT_LOCAL_FILE_PATH=True # When set to false don't allow HTTP requests to be sent from Cognee. # This protects against Server Side Request Forgery when proper infrastructure is not in place. ALLOW_HTTP_REQUESTS=True # When set to false don't allow cypher search to be used in Cognee. ALLOW_CYPHER_QUERY=True # When set to False errors during data processing will be returned as info but not raised to allow handling of faulty documents RAISE_INCREMENTAL_LOADING_ERRORS=True # When set to True, the Cognee backend will require authentication for requests to the API. # If you're disabling this, make sure to also disable ENABLE_BACKEND_ACCESS_CONTROL. REQUIRE_AUTHENTICATION=False # Set this variable to True to enforce usage of backend access control for Cognee # Note: This is only currently supported by the following databases: # Relational: SQLite, Postgres # Vector: LanceDB # Graph: KuzuDB # # It enforces LanceDB and KuzuDB use and uses them to create databases per Cognee user + dataset ENABLE_BACKEND_ACCESS_CONTROL=False ################################################################################ # ☁️ Cloud Sync Settings ################################################################################ # Cognee Cloud API settings for syncing data to/from cloud infrastructure COGNEE_CLOUD_API_URL="http://localhost:8001" COGNEE_CLOUD_AUTH_TOKEN="your-auth-token" ################################################################################ # 🛠️ DEV Settings ################################################################################ ENV="local" TOKENIZERS_PARALLELISM="false" # LITELLM Logging Level. Set to quiet down logging LITELLM_LOG="ERROR" # Set this environment variable to disable sending telemetry data # TELEMETRY_DISABLED=1 # Default User Configuration # DEFAULT_USER_EMAIL="" # DEFAULT_USER_PASSWORD="" ------------------------------- END OF POSSIBLE SETTINGS ------------------------------- ############################################################################### # 🧪 EXAMPLE OVERRIDES (commented out) ############################################################################### # The blocks below show how to configure alternative providers. # Uncomment + fill values to switch. ########## Azure OpenAI ####################################################### #LLM_MODEL="azure/gpt-5-mini" #LLM_ENDPOINT="https://DNS.azure.com/openai/deployments/gpt-5-mini" #LLM_API_KEY="<<TALK TO YOUR AZURE GUY" #LLM_API_VERSION="2024-12-01-preview" ## llm api version might not be relevant #LLM_MAX_TOKENS="16384" #EMBEDDING_MODEL="azure/text-embedding-3-large" #EMBEDDING_ENDPOINT="https://DNS.openai.azure.com/openai/deployments/text-embedding-3-large" #EMBEDDING_API_KEY="<<TALK TO YOUR AZURE GUY>" #EMBEDDING_API_VERSION="2024-12-01-preview" #EMBEDDING_DIMENSIONS=3072 #EMBEDDING_MAX_TOKENS=8191 ########## Local LLM via Ollama ############################################### #LLM_API_KEY ="ollama" #LLM_MODEL="llama3.1:8b" #LLM_PROVIDER="ollama" #LLM_ENDPOINT="http://localhost:11434/v1" #EMBEDDING_PROVIDER="ollama" #EMBEDDING_MODEL="avr/sfr-embedding-mistral:latest" #EMBEDDING_ENDPOINT="http://localhost:11434/api/embeddings" #EMBEDDING_DIMENSIONS=4096 #HUGGINGFACE_TOKENIZER="Salesforce/SFR-Embedding-Mistral" ########## OpenRouter (also free) ######################################################### #LLM_API_KEY="<<go-get-one-yourself" #LLM_PROVIDER="custom" #LLM_MODEL="openrouter/google/gemini-2.0-flash-lite-preview-02-05:free" #LLM_ENDPOINT="https://openrouter.ai/api/v1" ########## DeepInfra ########################################################## #LLM_API_KEY="<<>>" #LLM_PROVIDER="custom" #LLM_MODEL="deepinfra/meta-llama/Meta-Llama-3-8B-Instruct" #LLM_ENDPOINT="https://api.deepinfra.com/v1/openai" #EMBEDDING_PROVIDER="openai" #EMBEDDING_API_KEY="<<>>" #EMBEDDING_MODEL="deepinfra/BAAI/bge-base-en-v1.5" #EMBEDDING_ENDPOINT="" #EMBEDDING_API_VERSION="" #EMBEDDING_DIMENSIONS=3072 #EMBEDDING_MAX_TOKENS=8191 ########## Release Test ############################################### #LLM_API_KEY="..." #OPENAI_API_KEY="..." #MIGRATION_DB_PATH="~/Downloads/" #MIGRATION_DB_NAME="Chinook_Sqlite.sqlite" #MIGRATION_DB_PROVIDER="sqlite" #GRAPH_DATABASE_URL="bolt://54.246.89.112:7687" #GRAPH_DATABASE_USERNAME="neo4j" #GRAPH_DATABASE_PASSWORD="pleaseletmein"

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/topoteretes/cognee'

If you have feedback or need assistance with the MCP directory API, please join our Discord server