We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/tenequm/llms-txt-mcp'
If you have feedback or need assistance with the MCP directory API, please join our Discord server
name: CI
on:
push:
branches: [ main ]
pull_request:
branches: [ main ]
jobs:
build:
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Set up Python
uses: actions/setup-python@v5
with:
python-version: '3.12'
- name: Install uv
run: |
curl -LsSf https://astral.sh/uv/install.sh | sh
echo "$HOME/.cargo/bin" >> $GITHUB_PATH
echo "$HOME/.local/bin" >> $GITHUB_PATH
- name: Install dependencies
run: |
uv venv
uv pip install -e .[dev]
- name: Lint (ruff)
run: uv run ruff check .
- name: Type check (mypy)
run: uv run mypy src/
- name: Test (pytest)
run: uv run pytest -q