The Atla MCP Server enables standardized evaluation of LLM responses using Atla's evaluation models, providing scores and textual critiques based on specific criteria.
- Evaluate LLM responses: Score and critique responses against single or multiple evaluation criteria simultaneously
- Choose evaluation models: Use flagship (
atla-selene
) or compact (atla-selene-mini
) models - Customize evaluations: Optionally include original context and expected outputs
- Integration options: Connect via OpenAI Agents SDK, Claude Desktop, or Cursor
Provides compatibility with the OpenAI Agents SDK, allowing users to connect to the Atla MCP server for LLM evaluation services.
Atla MCP Server
An MCP server implementation providing a standardized interface for LLMs to interact with the Atla API for state-of-the-art LLMJ evaluation.
Learn more about Atla here. Learn more about the Model Context Protocol here.
Available Tools
evaluate_llm_response
: Evaluate an LLM's response to a prompt using a given evaluation criteria. This function uses an Atla evaluation model under the hood to return a dictionary containing a score for the model's response and a textual critique containing feedback on the model's response.evaluate_llm_response_on_multiple_criteria
: Evaluate an LLM's response to a prompt across multiple evaluation criteria. This function uses an Atla evaluation model under the hood to return a list of dictionaries, each containing an evaluation score and critique for a given criteria.
Usage
To use the MCP server, you will need an Atla API key. You can find your existing API key here or create a new one here.
Installation
We recommend using
uv
to manage the Python environment. See here for installation instructions.
Manually running the server
Once you have uv
installed and have your Atla API key, you can manually run the MCP server using uvx
(which is provided by uv
):
Connecting to the server
Having issues or need help connecting to another client? Feel free to open an issue or contact us!
OpenAI Agents SDK
For more details on using the OpenAI Agents SDK with MCP servers, refer to the official documentation.
- Install the OpenAI Agents SDK:
- Use the OpenAI Agents SDK to connect to the server:
Claude Desktop
For more details on configuring MCP servers in Claude Desktop, refer to the official MCP quickstart guide.
- Add the following to your
claude_desktop_config.json
file:
- Restart Claude Desktop to apply the changes.
You should now see options from atla-mcp-server
in the list of available MCP tools.
Cursor
For more details on configuring MCP servers in Cursor, refer to the official documentation.
- Add the following to your
.cursor/mcp.json
file:
You should now see atla-mcp-server
in the list of available MCP servers.
Contributing
Contributions are welcome! Please see the CONTRIBUTING.md file for details.
License
This project is licensed under the MIT License. See the LICENSE file for details.
You must be authenticated.
remote-capable server
The server can be hosted and run remotely because it primarily relies on remote services or has no dependency on the local environment.
The Atla MCP Server provides a standardized interface for LLMs to interact with the Atla API for state-of-the-art LLMJ evaluation.
Related MCP Servers
- AsecurityAlicenseAqualityAn MCP server that provides LLMs access to other LLMsLast updated -41412JavaScriptMIT License
- -securityAlicense-qualityAn MCP server that provides tools to load and fetch documentation from any llms.txt source, giving users full control over context retrieval for LLMs in IDE agents and applications.Last updated -177PythonMIT License
- -securityAlicense-qualityMCP server that enables LLMs to interact with Tripadvisor API, supporting location data, reviews, and photos through standardized MCP interfacesLast updated -PythonMIT License
- AsecurityFlicenseAqualityA lightweight MCP server that provides a unified interface to various LLM providers including OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, and Ollama.Last updated -6218Python