The Locust MCP Server enables you to run configurable load tests with seamless AI integration.
Execute load tests: Run tests with configurable parameters (users, spawn rate, runtime)
Flexible modes: Supports both headless and UI-based execution
Custom scenarios: Define custom task scenarios in your test scripts
Real-time output: Monitor test execution as it happens
Protocol support: Built-in HTTP/HTTPS protocol handling
API access: Easy-to-use API for programmatic test execution
AI integration: Seamlessly works with Model Context Protocol for AI-powered development environments
Enables running Locust load tests with configurable parameters (users, spawn rate, runtime) for HTTP/HTTPS performance testing through a simple API.
🚀 ⚡️ locust-mcp-server
A Model Context Protocol (MCP) server implementation for running Locust load tests. This server enables seamless integration of Locust load testing capabilities with AI-powered development environments.
✨ Features
Simple integration with Model Context Protocol framework
Support for headless and UI modes
Configurable test parameters (users, spawn rate, runtime)
Easy-to-use API for running Locust load tests
Real-time test execution output
HTTP/HTTPS protocol support out of the box
Custom task scenarios support

🔧 Prerequisites
Before you begin, ensure you have the following installed:
Python 3.13 or higher
uv package manager (Installation guide)
📦 Installation
Clone the repository:
Install the required dependencies:
Set up environment variables (optional): Create a
.envfile in the project root:
🚀 Getting Started
Create a Locust test script (e.g.,
hello.py):
Configure the MCP server using the below specs in your favorite MCP client (Claude Desktop, Cursor, Windsurf and more):
Now ask the LLM to run the test e.g.
run locust test for hello.py. The Locust MCP server will use the following tool to start the test:
run_locust: Run a test with configurable options for headless mode, host, runtime, users, and spawn rate
📝 API Reference
Run Locust Test
Parameters:
test_file: Path to your Locust test scriptheadless: Run in headless mode (True) or with UI (False)host: Target host to load testruntime: Test duration (e.g., "30s", "1m", "5m")users: Number of concurrent users to simulatespawn_rate: Rate at which users are spawned
✨ Use Cases
LLM powered results analysis
Effective debugging with the help of LLM
🤝 Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
📄 License
This project is licensed under the MIT License - see the LICENSE file for details.
remote-capable server
The server can be hosted and run remotely because it primarily relies on remote services or has no dependency on the local environment.
Tools
A server that integrates Locust load testing capabilities with AI-powered development environments, allowing users to run performance tests through natural language commands.
- ✨ Features
- 🔧 Prerequisites
- 📦 Installation
- 🚀 Getting Started
- 📝 API Reference
- ✨ Use Cases
- 🤝 Contributing
- 📄 License
Related Resources
Related MCP Servers
- Asecurity-licenseAqualityA Model Context Protocol (MCP) server implementation that allows AI assistants to run k6 load tests through natural language commands, supporting custom test durations and virtual users.Last updated -215MIT License
- -security-license-qualityAn AI-powered MCP server that automates web testing workflows by enabling recording, execution, and discovery of tests through natural language prompts.Last updated -59Apache 2.0
- -security-license-qualityAI-powered agent that streamlines web testing workflows by allowing developers to record, execute, and discover tests using natural language prompts in their AI coding assistants.Last updated -59Apache 2.0
- -security-license-qualityA lightweight server that enables AI agents to interact with the Windows operating system, allowing for file navigation, application control, UI interaction, and QA testing through various tools.Last updated -3,328MIT License