browser-use MCP server
local-only server
The server can only run on the clientβs local machine because it depends on local resources.
Integrations
Connects to Google's Gemini API using the Google API key for AI-driven browser automation tasks
Provides integration with Ollama's local LLM hosting service, supporting customizable context settings and model parameters for browser automation
Integrates with OpenAI's API for LLM functionality, enabling AI-powered browser control with customizable parameters
browser-use MCP server
Project Note: This MCP server implementation builds upon the browser-use/web-ui foundation. Core browser automation logic and configuration patterns are adapted from the original project.
AI-driven browser automation server implementing the Model Context Protocol (MCP) for natural language browser control.
Features
- π§ MCP Integration - Full protocol implementation for AI agent communication
- π Browser Automation - Page navigation, form filling, and element interaction
- ποΈ Visual Understanding - Screenshot analysis and vision-based interactions
- π State Persistence - Maintain browser sessions between tasks
- π Multi-LLM Support - OpenAI, Anthropic, Azure, DeepSeek integration
Quick Start
Prerequisites
- Python 3.11 or higher
- uv (fast Python package installer)
- Chrome/Chromium browser
Installation
Claude Desktop
On MacOS: ~/Library/Application\ Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
Local Development
Development
Troubleshooting
- Browser Conflicts: Close all Chrome instances before starting.
- API Errors: Verify API keys in environment variables match your LLM provider.
- Vision Support: Ensure
MCP_USE_VISION=true
for screenshot analysis.
Provider Configuration
The server supports multiple LLM providers through environment variables. Here are the available options for MCP_MODEL_PROVIDER
:
Provider | Value | Required Env Variables |
---|---|---|
Anthropic | anthropic | ANTHROPIC_API_KEY ANTHROPIC_ENDPOINT (optional) |
OpenAI | openai | OPENAI_API_KEY OPENAI_ENDPOINT (optional) |
Azure OpenAI | azure_openai | AZURE_OPENAI_API_KEY AZURE_OPENAI_ENDPOINT |
DeepSeek | deepseek | DEEPSEEK_API_KEY DEEPSEEK_ENDPOINT (optional) |
Gemini | gemini | GOOGLE_API_KEY |
Mistral | mistral | MISTRAL_API_KEY MISTRAL_ENDPOINT (optional) |
Ollama | ollama | OLLAMA_ENDPOINT (optional, defaults to localhost:11434) |
OpenRouter | openrouter | OPENROUTER_API_KEY OPENROUTER_ENDPOINT (optional) |
Notes:
- For endpoints marked as optional, default values will be used if not specified
- Temperature can be configured using
MCP_TEMPERATURE
(default: 0.3) - Model can be specified using
MCP_MODEL_NAME
- For Ollama models, additional context settings like
num_ctx
andnum_predict
are configurable
Credits
This project extends the browser-use/web-ui under MIT License. Special thanks to the original authors for their browser automation framework.
License
MIT - See LICENSE for details.
This server cannot be installed
AI-driven browser automation server that implements the Model Context Protocol to enable natural language control of web browsers for tasks like navigation, form filling, and visual interaction.