AWT (AI Watch Tester)
Provides specialized support for testing Flutter web applications, utilizing native CanvasKit and Semantics detection for accurate interaction.
Enables automated testing within CI/CD pipelines on GitHub Actions, allowing for regression testing and screenshot reporting on every commit.
Supports automated testing for Next.js applications with optimized speed modes for functional and regression testing.
Supports local AI execution for test generation and self-healing via Ollama, enabling private and free-to-run automated testing.
Leverages OpenAI's GPT-4o models to automatically scan web applications, generate test steps, and fix failing tests through a DevQA Loop.
Optimized for testing web applications built with React, featuring specific speed modes for rapid UI verification.
Integrates with Vercel for cloud-hosted testing environments and provides a dashboard for monitoring test results.
Why I built this
I was building a Flutter web app. Every time the UI changed, my Playwright tests broke. Fixing selectors, re-recording flows, updating assertions — it ate hours every week.
So I started hacking on a tool that could look at a page, write its own tests, and when those tests broke, figure out why and fix itself.
That became AWT.
It's not perfect. OCR-based matching can be flaky on certain fonts. The self-healing loop sometimes needs a nudge. But for my day-to-day QA work — especially on Canvas and Flutter Web apps that Cypress can't even touch — it works well enough that I stopped writing test code entirely.
I'm sharing it because I suspect I'm not the only one who got fed up. If you've ever spent an afternoon fixing broken selectors, this might save you some time.
What it does
You give AWT a URL. It opens a real Chrome window, reads the page, writes test steps, runs them, and reports what passed and what failed. If something breaks, the DevQA Loop kicks in — AI reads the error, patches the test, and retries automatically (up to 5 times).
No test code to write. No recording sessions. No manual selector updates.
Start in 5 Minutes
Option 1 — Agent Skill (works with Claude Code, Cursor, Copilot...)
# One-line install
npx skills add ksgisang/awt-skill --skill awt -g
# Then just ask your AI tool:
"Test the login flow on http://localhost:3000"
"Check if the signup form works"
"Run regression tests after my last commit"
# → AWT scans, generates steps, runs them, and reports backOption 2 — MCP Server (Claude Desktop, Cursor, Windsurf)
# Install
pip install aat-devqa
# Add to Claude Code
claude mcp add awt -- python mcp/server.py
# Tools available: aat_devqa, aat_run, aat_doctor, aat_list_scenarios,
# aat_validate, aat_generate_from_doc,
# aat_snapshot, aat_diff, aat_watchCloud & Local CLI versions are in active development and will be announced separately.
How It Works
You give AWT a URL
│
▼
🔍 SCAN — AWT opens Chrome and reads every button, input, and link
│
▼
📝 GENERATE — AI writes a step-by-step test plan (you review & approve)
│
▼
▶️ RUN — AWT clicks, types, and navigates like a real user
│
├── ✅ All passed → screenshot report saved
│
└── ❌ Something failed
│
▼
🔄 DEVQA LOOP — AI reads the failure,
fixes the test (or your code),
and tries again (up to 5 times)The DevQA Loop — AWT's Core Feature
Most testing tools stop when a test fails and wait for a human. AWT keeps going.
When a step fails, AWT:
Takes a screenshot of exactly what the browser shows
Reads the error message and the visible page content
Re-scans the page to check if anything moved or changed
Patches the specific failing step and retries
If the failure is a bug in your source code (not just a wrong selector), AWT can trace it — finding the route handler, component, or API endpoint that's misbehaving — and suggest or apply a fix.
# Watch the loop run live
aat devqa "checkout flow test" --url http://localhost:3000
# Or use it with your AI coding tool (Claude Code, Cursor, Copilot...)
# "Test the registration page" → AWT scans, generates, runs, fixesFour Ways to Use AWT
Cloud | Local CLI | Agent Skill | MCP Server | |
How to start | Sign up at ai-watch-tester.vercel.app |
|
|
|
Browser | Headless (server) | Real Chrome on your machine | Real Chrome on your machine | Real Chrome on your machine |
AI key needed | No (server-provided or BYOK) | Yes (your OpenAI / Anthropic / Ollama) | No — your AI tool is the brain | No |
Best for | Quick tests, PMs, planners | Developers, CI/CD | AI-assisted development | Claude Desktop, Cursor, Windsurf |
Price | Free (5/mo) · Pro $28.99 · Team $98.99 | Free forever (MIT) | Free forever | Free forever |
Agent Skill — Let your AI coding tool drive AWT
# One-line install
npx skills add ksgisang/awt-skill --skill awt -g
# Then ask your AI tool:
"Test the login flow on http://localhost:3000"
"Check if the signup form works"
"Run regression tests after my last commit"
# → AWT scans, generates test steps, runs them, and reports backMCP Server — Protocol-native
# Add to Claude Code
claude mcp add awt -- python mcp/server.py
# Tools available: aat_run, aat_doctor, aat_list_scenarios, aat_validate,
# aat_cost, aat_generate_from_doc,
# aat_snapshot, aat_diff, aat_watchUse AWT as a Quality Gate in Your AI Harness
If you use GSD (Get Shit Done), Cursor Rules, CLAUDE.md, or any AI coding workflow — AWT plugs in as the Verify step that catches what your AI missed.
GSD Workflow Integration
Discuss → Plan → Execute → Verify (AWT)
│
aat_generate_from_doc(PLAN.md)
│
→ scenarios generated
→ aat_run_skill_mode
→ pass/fail report# MCP: pass your GSD plan file to generate test scenarios
"Generate test scenarios from my PLAN.md and run them"
# → AWT reads the plan, creates E2E tests, runs in real Chrome
# CLI: same thing
aat generate --from PLAN.md
aat run scenarios/Add AWT to Your CLAUDE.md / Cursor Rules
## QA Verification (AWT)
- After writing code, verify with AWT before committing.
- Run: aat scan --url {URL} → review scenario → aat run
- All tests must pass before marking a task as done.This turns AWT into an automated quality gate — your AI writes code, AWT verifies it works, and the loop continues until everything passes.
Visual Regression — Catch UI Changes Automatically
AWT detects unintended UI changes by comparing screenshots before and after code changes. Zero AI tokens — pure Playwright + OpenCV.
# Step 1: Capture baselines (before code change)
aat snapshot scenarios/login.yaml --url http://localhost:3000
# Step 2: After code changes, compare
aat diff scenarios/login.yaml --url http://localhost:3000
# Result:
# step001 99.8% ✅ PASS
# step002 87.3% ❌ FAIL ← UI change detected
# step003 100.0% ✅ PASSResponsive — Test 3 Viewports at Once
# Capture mobile + tablet + desktop baselines
aat snapshot scenarios/ --url http://localhost:3000 --responsive
# Compare all 3 viewports
aat diff scenarios/ --url http://localhost:3000 --responsive
# Or test a single custom viewport
aat snapshot scenarios/ --viewport 414x896Viewport | Size | Use Case |
mobile | 375×812 | iPhone-class devices |
tablet | 768×1024 | iPad-class devices |
desktop | 1280×720 | Standard desktop |
Console Error Collection
# Capture baselines and check for JS errors
aat snapshot scenarios/ --url http://localhost:3000 --console
# Fail the run if console errors are found
aat snapshot scenarios/ --url http://localhost:3000 --console-failCatches hidden JavaScript errors that pass visual checks — TypeError, network 404s, uncaught exceptions.
Auto-Open Diff Images
# Open failed diff images in your system viewer
aat diff scenarios/ --url http://localhost:3000 --open
# → macOS: opens in Preview
# → Linux: opens with xdg-openWatch Mode — Auto-Test on File Changes
aat watch scenarios/ --url http://localhost:3000
# → Saves a file? Tests re-run automatically
# → Baselines exist? Visual diff included
# → Ctrl+C to stopScenario file changed → re-run that scenario only
Source file changed → re-run all scenarios (full regression)
Powered by watchfiles (Rust-based), with polling fallback
PR Comments (GitHub Action)
# Output GitHub-flavored markdown for PR comments
aat diff scenarios/ --format=github
# JSON output for other tools
aat diff scenarios/ --format=jsonA .github/workflows/visual-regression.yml template is included — drop it into your repo and every PR gets automatic visual regression comments.
What AWT Is Great At
Feature | Description | |
🤖 | Zero-code test generation | Point at a URL — AI generates complete test steps with real selectors |
🔄 | Self-healing DevQA Loop | Tests fail? AI fixes and retries automatically (up to 5 attempts) |
👁️ | Visual verification | Screenshots before/after every action — not just DOM checks |
🌐 | Real browser | Chrome with human-like mouse movement and typing speed |
📱 | Flutter support | Native CanvasKit + Semantics detection — tests Flutter web apps too |
📄 | Document-based generation | Feed a PDF/DOCX spec — AI generates tests from requirements |
⚡ | Speed modes |
|
📸 | Smart screenshots |
|
🔌 | Plugin architecture | Swap engines, matchers, AI providers via simple registries |
📊 | Visual regression | SSIM-based screenshot comparison — zero AI tokens, pure OpenCV |
📱 | Responsive testing | Test mobile/tablet/desktop viewports in one command ( |
🖥️ | Watch mode | Auto-run tests on file save — instant feedback loop |
🔍 | Console error capture | Catch hidden JS errors that pass visual checks ( |
Known Limitations (being honest)
AWT works well for me, but it has rough edges I haven't fully solved yet:
OCR matching can misfire on custom fonts or low-contrast UIs — fallback to DOM selectors when that happens
The self-healing loop occasionally fixates on the wrong element. Setting
--max-attempts 3helps avoid spinning wheelsComplex SPAs with heavy animation may need
--verbosity=detailedand a longer wait timeFirst-run scenario generation is only as good as the page's accessibility tree — poorly structured HTML gives poor results
If you hit something broken, please open an issue. I fix reported bugs fast.
AWT vs Other Tools
vs Playwright / Cypress
Playwright and Cypress are excellent — and AWT is built on top of Playwright. The difference is who writes the tests:
AWT | Playwright / Cypress | |
Who writes tests | AI (from your URL) | You (code) |
Maintenance when UI changes | AI auto-heals | You update selectors manually |
Learning curve | Zero — just paste a URL | Moderate (framework API + JS/TS) |
Flexibility | High (YAML scenarios) | Maximum (full code control) |
Use Playwright/Cypress when you want full programmatic control. Use AWT when you want tests without writing them.
vs testRigor
AWT | testRigor | |
Test authoring | AI generates from URL — you write nothing | Plain English (you write commands) |
Self-healing | DevQA Loop (AI re-generates automatically) | Built-in auto-maintenance |
Pricing | Free (MIT, self-host) | Enterprise (~$800+/mo) |
Open source | ✅ MIT License | ❌ |
vs Applitools
Applitools specializes in visual regression with AI-powered comparison. AWT now includes its own visual regression (SSIM-based, zero AI tokens) plus functional testing. For pixel-perfect comparison across browsers and viewports, Applitools is still more mature. But AWT covers both functional + visual in a single free tool.
Speed & Screenshot Modes
Control the trade-off between thoroughness and speed:
# CI/CD — fastest, minimal storage
aat run --verbosity=concise --screenshots=on-failure scenarios/
# Standard QA — balanced (recommended)
aat run --verbosity=concise --screenshots=before-after scenarios/
# Full audit — every step recorded
aat run --verbosity=detailed --screenshots=all scenarios/Mode | Steps | Screenshots | ~Time | Use For |
| 12–15 | 0–1 | ~1 min | CI/CD gates |
| 12–15 | 24 | ~2 min | Daily QA |
| 60–80 | 68 | ~5 min | Compliance / audit |
Supported AI Providers
Provider | Models | Cost | Setup |
OpenAI | gpt-4o, gpt-4o-mini | Pay-per-use |
|
Anthropic | Claude Sonnet 4 | Pay-per-use |
|
Ollama | codellama, llama3, mistral | Free (local) |
|
# aat.yaml
ai:
provider: openai # openai | anthropic | ollama
model: gpt-4o
api_key: ${OPENAI_API_KEY}Architecture
aat devqa / aat run / aat snapshot / aat diff / aat watch
│
▼
┌─────────────────────────────────────┐
│ CLI (Typer) │
├─────────────────────────────────────┤
│ Core Orchestrator │
│ Executor · Comparator · DevQALoop │
├────────────┬────────────┬───────────┤
│ Engine │ Matcher │ AI Adapter│
│ web/desktop│ ocr/cv/ai │openai/etc.│
├────────────┴────────────┴───────────┤
│ Visual Regression · Watch Mode │
│ Pydantic v2 Models · SQLite Learn │
└─────────────────────────────────────┘All modules follow a plugin registry pattern — add a new engine, matcher, or AI provider by implementing one base class and registering it in __init__.py.
Development
Prerequisites
Python 3.11+
Tesseract OCR:
brew install tesseract/apt install tesseract-ocr
Commands
Command | What it does |
| Install all dependencies + Playwright + pre-commit |
| Check code style (ruff) |
| Auto-fix formatting |
| Strict type checking (mypy) |
| Run all tests (pytest) |
| Tests + coverage report |
git clone https://github.com/ksgisang/AI-Watch-Tester.git
cd AI-Watch-Tester
python -m venv .venv && source .venv/bin/activate
make dev
make test # verify everything works
aat dashboard # launch at http://localhost:9500Contributing
Contributions, bug reports, and new plugins are welcome — see CONTRIBUTING.md.
If something doesn't work for your use case, please open an issue before spending time on a PR. Happy to help scope the right fix.
git checkout -b feat/my-feature
make format && make lint && make typecheck && make test
git commit -m "feat(scope): description"FAQ
No. The Cloud version at ai-watch-tester.vercel.app needs nothing — just a browser. The local CLI needs one terminal command to install.
The only thing AWT needs from you is a URL and (optionally) a description of what to test.
When a web app changes — a button moves, a label changes, a new form field appears — traditional tests break and stay broken until someone manually updates them.
AWT's DevQA Loop re-scans the page after a failure, finds the updated element, and patches the test step automatically. You don't have to touch the test files.
Cloud (no install): ai-watch-tester.vercel.app
Local:
pip install aat-devqa
playwright install chromium
aat dashboard # opens at http://localhost:9500From source:
git clone https://github.com/ksgisang/AI-Watch-Tester.git
cd AI-Watch-Tester
make dev && aat dashboard
|
| |
Starting point | Just a description + URL | Existing scenario file |
Test generation | Automatic (scans and writes) | Uses your file |
Failure fixing | Patches the test YAML | AI patches your source code |
Best for | First run, quick testing | Iterative dev with code fixes |
Use aat devqa when starting from scratch. Use aat loop when you want AWT to also fix your application code.
--verbosity — how many steps run:
detailed(default): all steps including wait/assert/screenshotconcise: core actions only (navigate, click, type) — faster
--screenshots — how many images are saved:
all(default): after every stepbefore-after: before + after each click/type/navigate (~70% fewer files)on-failure: only when a step fails (great for CI/CD)
# Recommended for daily QA
aat run --verbosity=concise --screenshots=before-after scenarios/
# For CI/CD pipelines
aat run --verbosity=concise --screenshots=on-failure scenarios/Provider | Models | Cost |
OpenAI | gpt-4o, gpt-4o-mini | Pay-per-use |
Anthropic | Claude Sonnet 4 | Pay-per-use |
Ollama | codellama, llama3, mistral | Free (local GPU) |
Cloud BYOK keys are encrypted at rest (Fernet/AES-128-CBC).
Plan | Price | Tests/month |
Free | $0 | 5 |
Pro | $28.99/mo | 100 |
Team | $98.99/mo | 500 |
The local CLI is free forever with no limits.
Visual regression compares screenshots taken before and after a code change. If the UI changed unexpectedly (layout shift, missing element, color change), AWT flags it.
aat snapshot scenarios/ --url http://localhost:3000 # before change
# ... make code changes ...
aat diff scenarios/ --url http://localhost:3000 # compareAWT uses SSIM (Structural Similarity Index) — no AI tokens needed. Add --responsive to test mobile, tablet, and desktop viewports at once.
Yes. For local runs, use the --screenshots=on-failure flag to keep output minimal. For cloud, the API accepts a POST request:
curl -X POST https://your-awt-server.com/api/v1/run \
-H "X-API-Key: awt_your_key" \
-H "Content-Type: application/json" \
-d '{"target_url": "https://staging.example.com"}'See the CI/CD Guide for GitHub Actions and GitLab CI examples.
All traffic encrypted via HTTPS/TLS
BYOK API keys: Fernet-encrypted (AES-128-CBC + HMAC-SHA256) at rest
Screenshots: auto-deleted after 7 days
Local mode: nothing leaves your machine
License
MIT — free for personal and commercial use.
This server cannot be installed
Maintenance
Latest Blog Posts
MCP directory API
We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/ksgisang/AI-Watch-Tester'
If you have feedback or need assistance with the MCP directory API, please join our Discord server