Skip to main content
Glama

cognee-mcp

weighted_edges_tests.yml5.13 kB
name: Weighted Edges Tests on: push: branches: [ main, weighted_edges ] paths: - 'cognee/modules/graph/utils/get_graph_from_model.py' - 'cognee/infrastructure/engine/models/Edge.py' - 'cognee/tests/unit/interfaces/graph/test_weighted_edges.py' - 'examples/python/weighted_edges_example.py' - '.github/workflows/weighted_edges_tests.yml' pull_request: branches: [ main ] paths: - 'cognee/modules/graph/utils/get_graph_from_model.py' - 'cognee/infrastructure/engine/models/Edge.py' - 'cognee/tests/unit/interfaces/graph/test_weighted_edges.py' - 'examples/python/weighted_edges_example.py' - '.github/workflows/weighted_edges_tests.yml' env: RUNTIME__LOG_LEVEL: ERROR ENV: 'dev' jobs: test-weighted-edges-functionality: name: Test Weighted Edges Core Functionality runs-on: ubuntu-22.04 strategy: matrix: python-version: ['3.11', '3.12'] env: LLM_PROVIDER: openai LLM_MODEL: gpt-5-mini LLM_API_KEY: ${{ secrets.LLM_API_KEY }} steps: - name: Check out repository uses: actions/checkout@v4 with: fetch-depth: 0 - name: Cognee Setup uses: ./.github/actions/cognee_setup with: python-version: ${{ matrix.python-version }} - name: Run Weighted Edges Unit Tests run: | uv run pytest cognee/tests/unit/interfaces/graph/test_weighted_edges.py -v --tb=short - name: Run Standard Graph Tests (Regression) run: | uv run pytest cognee/tests/unit/interfaces/graph/get_graph_from_model_unit_test.py -v --tb=short test-with-different-databases: name: Test Weighted Edges with Different Graph Databases runs-on: ubuntu-22.04 strategy: matrix: database: ['kuzu', 'neo4j'] include: - database: kuzu graph_db_provider: "kuzu" - database: neo4j graph_db_provider: "neo4j" env: LLM_PROVIDER: openai LLM_MODEL: gpt-5-mini LLM_ENDPOINT: https://api.openai.com/v1/ LLM_API_KEY: ${{ secrets.LLM_API_KEY }} LLM_API_VERSION: "2024-02-01" EMBEDDING_PROVIDER: openai EMBEDDING_MODEL: text-embedding-3-small EMBEDDING_ENDPOINT: https://api.openai.com/v1/ EMBEDDING_API_KEY: ${{ secrets.LLM_API_KEY }} EMBEDDING_API_VERSION: "2024-02-01" steps: - name: Check out repository uses: actions/checkout@v4 with: fetch-depth: 0 - name: Cognee Setup uses: ./.github/actions/cognee_setup with: python-version: '3.11' - name: Dependencies already installed run: echo "Dependencies already installed in setup" - name: Run Weighted Edges Tests env: GRAPH_DATABASE_PROVIDER: ${{ matrix.graph_db_provider }} run: | uv run pytest cognee/tests/unit/interfaces/graph/test_weighted_edges.py -v --tb=short test-examples: name: Test Weighted Edges Examples runs-on: ubuntu-22.04 env: LLM_PROVIDER: openai LLM_MODEL: gpt-5-mini LLM_ENDPOINT: https://api.openai.com/v1/ LLM_API_KEY: ${{ secrets.LLM_API_KEY }} LLM_API_VERSION: "2024-02-01" EMBEDDING_PROVIDER: openai EMBEDDING_MODEL: text-embedding-3-small EMBEDDING_ENDPOINT: https://api.openai.com/v1/ EMBEDDING_API_KEY: ${{ secrets.LLM_API_KEY }} EMBEDDING_API_VERSION: "2024-02-01" steps: - name: Check out repository uses: actions/checkout@v4 with: fetch-depth: 0 - name: Cognee Setup uses: ./.github/actions/cognee_setup with: python-version: '3.11' - name: Test Weighted Edges Example run: | uv run python examples/python/weighted_edges_example.py - name: Verify Visualization File Created run: | if [ -f "examples/python/weighted_graph_visualization.html" ]; then echo "✅ Visualization file created successfully" ls -la examples/python/weighted_graph_visualization.html else echo "❌ Visualization file not found" exit 1 fi code-quality: name: Code Quality for Weighted Edges runs-on: ubuntu-22.04 steps: - name: Check out repository uses: actions/checkout@v4 with: fetch-depth: 0 - name: Cognee Setup uses: ./.github/actions/cognee_setup with: python-version: '3.11' - name: Run Linting on Weighted Edges Files uses: astral-sh/ruff-action@v2 with: args: "check cognee/modules/graph/utils/get_graph_from_model.py cognee/tests/unit/interfaces/graph/test_weighted_edges.py examples/python/weighted_edges_example.py" - name: Run Formatting Check on Weighted Edges Files uses: astral-sh/ruff-action@v2 with: args: "format --check cognee/modules/graph/utils/get_graph_from_model.py cognee/tests/unit/interfaces/graph/test_weighted_edges.py examples/python/weighted_edges_example.py"

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/topoteretes/cognee'

If you have feedback or need assistance with the MCP directory API, please join our Discord server