Skip to main content
Glama
docker-compose.yaml2.34 kB
--- services: innitctl: container_name: innitctl image: systeminit/innitctl:stable command: - "--configs" - "/configs" - "--output" - "/output" environment: - SI_INNITCTL__HOST_ENVIRONMENT=${SI_HOSTENV} - SI_INNITCTL__SERVICE_NAME=${SI_SERVICE} - SI_INNITCTL__INSTANCE_ID=${SI_INSTANCE_ID} - SI_INNIT__CLIENT_CA_ARN=${SI_PRIVATE_CA_ARN} volumes: - config:/output - /etc/ssl/private:/etc/ssl/private otelcol: container_name: otelcol image: systeminit/otelcol:stable ports: - 4317:4317 - 55679:55679 - 9090:9090 volumes: - config:/etc/otelcol/ depends_on: innitctl: condition: service_completed_successfully node-exporter: container_name: node_exporter image: prom/node-exporter:v1.7.0 command: - --collector.systemd - --collector.systemd.unit-include=${SI_SERVICE}.service - --path.procfs=/host/proc - --path.sysfs=/host/sys - --path.rootfs=/rootfs - --collector.filesystem.ignored-mount-points=^/(sys|proc|dev|host|etc)($$|/) ports: - 9000:9100 volumes: - /proc:/host/proc:ro - /sys:/host/sys:ro - /:/rootfs:ro - /var/run/dbus/system_bus_socket:/var/run/dbus/system_bus_socket:ro vector: container_name: vector image: timberio/vector:0.45.X-debian ports: - 8686:8686 volumes: - /var/run/docker.sock:/var/run/docker.sock - /var/log/journal:/var/log/journal - config:/etc/vector/ depends_on: innitctl: condition: service_completed_successfully prometheus: container_name: prometheus image: prom/prometheus ports: - "9100:9090" volumes: - "config:/etc/prometheus/" depends_on: - otelcol health-monitor: container_name: health-monitor image: systeminit/hypochondriac:stable profiles: - rebaser - veritech - pinga - forklift - edda environment: - SI_SERVICE=${SI_SERVICE} - SI_INSTANCE_ID=${SI_INSTANCE_ID} - AWS_REGION=${AWS_REGION:-us-east-1} - HEALTH_ENDPOINT=http://host.docker.internal:8080/health - SLEEP_INTERVAL=30 extra_hosts: - "host.docker.internal:host-gateway" restart: unless-stopped volumes: config:

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/systeminit/si'

If you have feedback or need assistance with the MCP directory API, please join our Discord server