Integrations
Supports environment variable configuration through .env files, allowing users to securely store LLM API keys and other configuration parameters
Integrates with GitHub Copilot to enable AI-assisted web testing automation, allowing Copilot to record test flows, execute regression tests, and discover potential test steps through natural language prompts
Leverages Python for web testing automation, with support for Playwright browser control and LLM integration for test recording, execution, and discovery
VibeShift Web tester
This project provides an AI-powered agent designed to streamline web testing workflows, particularly for developers using AI coding assistants like GitHub Copilot, Cursor, Roo Code, etc. It integrates directly into these assistants via the MCP (Machine Command Protocol), allowing you to automate test recording, execution, and discovery using natural language prompts.
The Problem: Manually testing web applications after generating code with AI assistants is time-consuming and error-prone. Furthermore, AI-driven code changes can inadvertently introduce regressions in previously working features.
The Solution: This tool bridges the gap by enabling your AI coding assistant to:
- Record new test flows: Describe a user journey in natural language, and the agent will interact with the browser (using Playwright) under AI guidance to generate a reproducible test script (JSON format).
- Execute existing tests: Run previously recorded test scripts to perform regression testing, ensuring new code changes haven't broken existing functionality.
- Discover potential test steps: Crawl a website, analyze pages using vision and DOM structure, and ask an LLM to suggest relevant test steps for different pages.
This creates a tighter feedback loop, automating the testing process and allowing the AI assistant (and the developer) to quickly identify and fix issues or regressions.
Demo (Click to play these videos)
Features
- MCP Integration: Seamlessly integrates with Cursor/Windsurf/Github Copilot/Roo Code
- AI-Assisted Test Recording: Generate Playwright-based test scripts from natural language descriptions (in automated mode).
- Deterministic Test Execution: Run recorded JSON test files reliably using Playwright.
- AI-Powered Test Discovery: Crawl websites and leverage any LLM (in openai compliant format) to suggest test steps for discovered pages.
- Regression Testing: Easily run existing test suites to catch regressions.
- Automated Feedback Loop: Execution results (including failures, screenshots, console logs) are returned, providing direct feedback to the AI assistant.
- Self Healing: Existing tests self heal in case of code changes. No need to manually update.
- UI tests: UI tests which aren't supported by playwright directly are also supported. For example,
Check if the text is overflowing in the div
- Visual Regression Testing: Using traditional pixelmatch and vision LLM approach.
How it Works
- User: Prompts their AI coding assistant (e.g., "Record a test for the login flow", "Run the regression test 'test_login.json'").
- AI Coding Agent: Recognizes the intent and uses MCP to call the appropriate tool provided by the
MCP Server
. - MCP Server: Routes the request to the corresponding function (
record_test_flow
,run_regression_test
,discover_test_flows
,list_recorded_tests
). - Web Test Agent:
- Recording: The
WebAgent
(in automated mode) interacts with the LLM to plan steps, controls the browser viaBrowserController
(Playwright), processes HTML/Vision, and saves the resulting test steps to a JSON file in theoutput/
directory. - Execution: The
TestExecutor
loads the specified JSON test file, usesBrowserController
to interact with the browser according to the recorded steps, and captures results, screenshots, and console logs. - Discovery: The
CrawlerAgent
usesBrowserController
andLLMClient
to crawl pages and suggest test steps.
- Recording: The
- Browser: Playwright drives the actual browser interaction.
- Feedback: The results (success/failure, file paths, error messages, discovered steps) are returned through the MCP server to the AI coding assistant, which then presents them to the user.
Getting Started
Prerequisites
- Python 3.10+
- Access to any LLM (gemini 2.0 flash works best for free in my testing)
- MCP installed (
pip install mcp[cli]
) - Playwright browsers installed (
playwright install
)
Installation
- Clone the repository:Copy
- Create a virtual environment (recommended):Copy
- Install dependencies:Copy
- Install Playwright browsers:Copy
Configuration
- Rename the .env.example to .env file in the project root directory.
- Add your LLM API key and other necessary details:Copy
- Replace
YOUR_LLM_API_KEY
with your actual key.
- Replace
Adding the MCP Server
Add this to you mcp config:
Keep this server running while you interact with your AI coding assistant.
Usage
Interact with the agent through your MCP-enabled AI coding assistant using natural language.
Examples:
- Record a Test:
"Record a test: go to https://practicetestautomation.com/practice-test-login/, type 'student' into the username field, type 'Password123' into the password field, click the submit button, and verify the text 'Congratulations student' is visible."
- (The agent will perform these actions automatically and save a
test_....json
file inoutput/
)
- (The agent will perform these actions automatically and save a
- Execute a Test:
"Run the regression test
output/test_practice_test_login_20231105_103000.json
"- (The agent will execute the steps in the specified file and report PASS/FAIL status with errors and details.)
- Discover Test Steps:
"Discover potential test steps starting from https://practicetestautomation.com/practice/"
- (The agent will crawl the site, analyze pages, and return suggested test steps for each.)
- List Recorded Tests:
"List the available recorded web tests."
- (The agent will return a list of
.json
files found in theoutput/
directory.)
- (The agent will return a list of
Output:
- Recorded Tests: Saved as JSON files in the
output/
directory (seetest_schema.md
for format). - Execution Results: Returned as a JSON object summarizing the run (status, errors, evidence paths). Full results are also saved to
output/execution_result_....json
. - Discovery Results: Returned as a JSON object with discovered URLs and suggested steps. Full results saved to
output/discovery_results_....json
.
Inspiration
- Browser Use: The dom context tree generation is heavily inspired from them and is modified to accomodate static/dynamic/visual elements. Special thanks to them for their contribution to open source.
Contributing
Contributions are welcome! Please see CONTRIBUTING.md for details on how to get started, report issues, and submit pull requests.
License
This project is licensed under the APACHE-2.0.
This server cannot be installed
AI-powered agent that streamlines web testing workflows by allowing developers to record, execute, and discover tests using natural language prompts in their AI coding assistants.
Related MCP Servers
- -securityAlicense-qualityA Message Control Protocol server that runs PHP tests and static analysis tools automatically for developers, providing results directly to AI assistants in Cursor editor.Last updated -1TypeScriptMIT License
- -securityAlicense-qualityEmpowers AI agents to perform web browsing, automation, and scraping tasks with minimal supervision using natural language instructions and Selenium.Last updated -1PythonApache 2.0
- -securityAlicense-qualityAllows AI agents to control web browser sessions via Selenium WebDriver, enabling web automation tasks like scraping, testing, and form filling through the Model Context Protocol.Last updated -TypeScriptMIT License
- -securityFlicense-qualityAn open-source self-hosted browser agent that provides a dockerized browser environment for AI automation, allowing other AI apps and agents to perform human-like web browsing tasks through natural language instructions.Last updated -11916TypeScript