The Locust MCP Server enables you to run configurable load tests with seamless AI integration.
Execute load tests: Run tests with configurable parameters (users, spawn rate, runtime)
Flexible modes: Supports both headless and UI-based execution
Custom scenarios: Define custom task scenarios in your test scripts
Real-time output: Monitor test execution as it happens
Protocol support: Built-in HTTP/HTTPS protocol handling
API access: Easy-to-use API for programmatic test execution
AI integration: Seamlessly works with Model Context Protocol for AI-powered development environments
Enables running Locust load tests with configurable parameters (users, spawn rate, runtime) for HTTP/HTTPS performance testing through a simple API.
🚀 ⚡️ locust-mcp-server
A Model Context Protocol (MCP) server implementation for running Locust load tests. This server enables seamless integration of Locust load testing capabilities with AI-powered development environments.
✨ Features
Simple integration with Model Context Protocol framework
Support for headless and UI modes
Configurable test parameters (users, spawn rate, runtime)
Easy-to-use API for running Locust load tests
Real-time test execution output
HTTP/HTTPS protocol support out of the box
Custom task scenarios support

Related MCP server: Web-QA
🔧 Prerequisites
Before you begin, ensure you have the following installed:
Python 3.13 or higher
uv package manager (Installation guide)
📦 Installation
Clone the repository:
Install the required dependencies:
Set up environment variables (optional): Create a
.envfile in the project root:
🚀 Getting Started
Create a Locust test script (e.g.,
hello.py):
Configure the MCP server using the below specs in your favorite MCP client (Claude Desktop, Cursor, Windsurf and more):
Now ask the LLM to run the test e.g.
run locust test for hello.py. The Locust MCP server will use the following tool to start the test:
run_locust: Run a test with configurable options for headless mode, host, runtime, users, and spawn rate
📝 API Reference
Run Locust Test
Parameters:
test_file: Path to your Locust test scriptheadless: Run in headless mode (True) or with UI (False)host: Target host to load testruntime: Test duration (e.g., "30s", "1m", "5m")users: Number of concurrent users to simulatespawn_rate: Rate at which users are spawned
✨ Use Cases
LLM powered results analysis
Effective debugging with the help of LLM
🤝 Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
📄 License
This project is licensed under the MIT License - see the LICENSE file for details.