Search for:
Why this server?
This server allows Large Language Models like Claude to query New Relic logs and metrics using NRQL queries.
Why this server?
A specialized MCP server that helps analyze and debug Model Context Protocol logs by providing Claude with direct access to log files across multiple platforms.
Why this server?
Allows developers and AI Agents to query Google Cloud Logging using natural language, translating queries into Google Cloud Logging Query Language (LQL) with Vertex AI Gemini 2.5.
Why this server?
Enables interaction with Datadog's monitoring platform to search logs, search trace spans, and perform trace span aggregation for analysis.
Why this server?
A read-only MCP server for Kubernetes that allows querying cluster information and diagnosing issues through natural language interfaces like Claude.
Why this server?
A Model Context Protocol server that enables LLMs to safely execute make targets from a Makefile, allowing AI assistants to run tests, format code, and automate various development tasks through natural language interaction.
Why this server?
An MCP server that enables Large Language Models to retrieve, analyze, and query metric data from Prometheus databases through pre-defined routes.
Why this server?
A code review tool for Cursor IDE that uses OpenAI's GPT models to provide intelligent code analysis and suggestions.
Why this server?
A remote Model Context Protocol server acting as middleware to the Sentry API, allowing AI assistants like Claude to access Sentry data and functionality through natural language interfaces.
Why this server?
A Model Context Protocol server that enables Large Language Models to seamlessly interact with ClickHouse databases, supporting resource listing, schema retrieval, and query execution.