Enables web search capabilities through DuckDuckGo search engine as part of a multi-engine search approach with parallel execution and result deduplication
Provides web search functionality through Startpage search engine as part of a comprehensive multi-engine search system with source attribution
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@Web Search MCP Serverfind recent developments in quantum computing"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
WebSearch MCP Server
High-performance Model Context Protocol (MCP) server for web search and content extraction with intelligent fallback system.
β¨ Features
π Fast: Pure async with connection pooling and zero event loop overhead
π Connection Reuse: Global HTTP connection pool (30-50% faster requests)
β‘ Parallel Execution: asyncio.gather for 2-3x faster batch operations
π Multi-Engine: Google, Bing, DuckDuckGo, Startpage, Brave Search
π‘οΈ Intelligent Fallbacks: GoogleβStartpage, BingβDuckDuckGo, Brave (standalone)
π Content Extraction: Clean text extraction from web pages
πΎ Smart Caching: LRU cache with compression and deduplication
π API Integration: Google Custom Search, Brave Search APIs with quota management
π Auto-Rotation: Timestamped logs (weekly) and metrics (monthly) with auto-cleanup
β‘ Resilient: Automatic failover and comprehensive error handling
π¦ Installation
Quick Start (Recommended)
# Install uv
brew install uv
# Run directly - no setup needed
uvx --from git+https://github.com/vishalkg/web-search websearch-serverDevelopment
git clone https://github.com/vishalkg/web-search.git
cd web-search
uv pip install -e .βοΈ Configuration
API Keys (Optional but Recommended)
For best results, configure API keys for Google Custom Search and Brave Search. Without API keys, the server falls back to web scraping which is less reliable.
Get API Keys:
Google: Custom Search API
Brave: Brave Search API
Kiro CLI (formerly Amazon Q CLI)
# Add to Kiro CLI with API keys
kiro-cli mcp add --name websearch \
--command uvx \
--args "--from" --args "git+https://github.com/vishalkg/web-search" --args "websearch-server" \
--env "GOOGLE_CSE_API_KEY=your-google-api-key" \
--env "GOOGLE_CSE_ID=your-search-engine-id" \
--env "BRAVE_SEARCH_API_KEY=your-brave-api-key"
# Verify installation
kiro-cli mcp listConfiguration scopes:
--scope global: Available to all agents--scope workspace: Project-specific configuration--scope default: Default agent configuration (default)
Claude Code
Add to your Claude Code MCP settings (~/.claude/mcp_settings.json):
{
"mcpServers": {
"websearch": {
"command": "uvx",
"args": ["--from", "git+https://github.com/vishalkg/web-search", "websearch-server"],
"env": {
"GOOGLE_CSE_API_KEY": "your-google-api-key",
"GOOGLE_CSE_ID": "your-search-engine-id",
"BRAVE_SEARCH_API_KEY": "your-brave-api-key"
}
}
}
}Test the integration:
# Restart Claude Code after configuration
# Use search_web or fetch_page_content tools in your conversationsClaude Desktop
Add to your MCP settings file with API keys:
{
"mcpServers": {
"websearch": {
"command": "uvx",
"args": ["--from", "git+https://github.com/vishalkg/web-search", "websearch-server"],
"env": {
"GOOGLE_CSE_API_KEY": "your-google-api-key",
"GOOGLE_CSE_ID": "your-search-engine-id",
"BRAVE_SEARCH_API_KEY": "your-brave-api-key"
}
}
}
}ποΈ File Structure
The server automatically manages files in OS-appropriate locations:
macOS:
~/Library/Application Support/websearch/ # Data
~/Library/Logs/websearch/ # Logs
~/Library/Application Support/websearch/ # ConfigLinux:
~/.local/share/websearch/ # Data
~/.local/state/websearch/ # Logs
~/.config/websearch/ # ConfigFiles:
data/
βββ search-metrics.jsonl # Search analytics (auto-rotated)
βββ quota/
βββ quotas.json # API quota tracking
logs/
βββ web-search.log # Application logs (auto-rotated)
config/
βββ .env # Configuration file
βββ cache/ # Optional cachingEnvironment Variable Overrides
WEBSEARCH_HOME: Base directory (default:~/.websearch)WEBSEARCH_CONFIG_DIR: Config directory overrideWEBSEARCH_LOG_DIR: Log directory override
π§ Usage
The server provides two main tools with multiple search modes:
Search Web
# Standard 5-engine search (backward compatible)
search_web("quantum computing applications", num_results=10)
# New 3-engine fallback search (optimized)
search_web_fallback("machine learning tutorials", num_results=5)Search Engines:
Google Custom Search API (with Startpage fallback)
Bing (with DuckDuckGo fallback)
Brave Search API (standalone)
DuckDuckGo (scraping)
Startpage (scraping)
Fetch Page Content
# Extract clean text from URLs
fetch_page_content("https://example.com")
fetch_page_content(["https://site1.com", "https://site2.com"]) # Batch processingποΈ Architecture
websearch/
βββ core/
β βββ async_search.py # Async search orchestration
β βββ async_fallback_search.py # 3-engine fallback system
β βββ ranking.py # Quality-first result ranking
β βββ content.py # Content fetching with async
β βββ common.py # Shared utilities
βββ engines/
β βββ google_api.py # Google Custom Search API
β βββ brave_api.py # Brave Search API (native async)
β βββ async_search.py # Async engine implementations
β βββ parsers.py # HTML parsing utilities
βββ utils/
β βββ connection_pool.py # Global HTTP connection pooling
β βββ unified_quota.py # Unified API quota management
β βββ advanced_cache.py # LRU cache with compression
β βββ deduplication.py # Result deduplication
β βββ http.py # HTTP utilities
βββ server.py # FastMCP server (pure async)π§ Advanced Configuration
Environment Variables
# API Configuration
export GOOGLE_CSE_API_KEY=your_google_api_key
export GOOGLE_CSE_ID=your_google_cse_id
export BRAVE_SEARCH_API_KEY=your_brave_api_key
# Quota Management (Optional)
export GOOGLE_DAILY_QUOTA=100 # Default: 100 requests/day
export BRAVE_MONTHLY_QUOTA=2000 # Default: 2000 requests/month
# Performance Tuning
export WEBSEARCH_CACHE_SIZE=1000
export WEBSEARCH_TIMEOUT=10
export WEBSEARCH_LOG_LEVEL=INFOHow to Get API Keys
Google Custom Search API
API Key: Go to https://developers.google.com/custom-search/v1/introduction and click "Get a Key"
CSE ID: Go to https://cse.google.com/cse/ and follow prompts to create a search engine
Brave Search API
Go to Brave Search API
Sign up for a free account
Go to your dashboard
Copy the API key as
BRAVE_API_KEYFree tier: 2000 requests/month
Quota Management
Unified System: Single quota manager for all APIs
Google: Daily quota (default 100 requests/day)
Brave: Monthly quota (default 2000 requests/month)
Storage: Quota files stored in
~/.websearch/directoryAuto-reset: Quotas automatically reset at period boundaries
Fallback: Automatic fallback to scraping when quotas exhausted
Search Modes
Standard Mode: Uses all 5 engines for maximum coverage
Fallback Mode: Uses 3 engines with intelligent fallbacks for efficiency
API-First Mode: Prioritizes API calls over scraping when keys available
π Troubleshooting
Issue | Solution |
No results | Check internet connection and logs |
API quota exhausted | System automatically falls back to scraping |
Google API errors | Verify |
Brave API errors | Check |
Permission denied |
|
Import errors | Ensure Python 3.12+ and dependencies installed |
Circular import warnings | Fixed in v2.0+ (10.00/10 pylint score) |
Debug Mode
# Enable detailed logging
export WEBSEARCH_LOG_LEVEL=DEBUG
python -m websearch.serverAPI Status Check
# Test API connectivity
cd debug/
python test_brave_api.py # Test Brave API
python test_fallback.py # Test fallback systemπ Performance & Monitoring
Metrics
Pylint Score: 10.00/10 (perfect code quality)
Search Speed: ~1-1.5 seconds for 3-engine search (50-70% faster with optimizations)
Batch Operations: 2-3x faster with asyncio.gather vs threading
Connection Reuse: 90%+ reuse rate (200-500ms saved per request)
Memory Efficiency: ~4000x less memory for batch operations (async vs threads)
Cache Hit Rate: ~85% for repeated queries
API Quota Efficiency: Automatic fallback prevents service interruption
Monitoring
Logs are written to web-search.log with structured format:
tail -f web-search.log | grep "search completed"π Security
No hardcoded secrets: All API keys via environment variables
Clean git history: Secrets scrubbed from all commits
Input validation: Comprehensive sanitization of search queries
Rate limiting: Built-in quota management for API calls
Secure defaults: HTTPS-only requests, timeout protection
π Performance Tips
Connection pooling is automatic - connections are reused for 30-50% faster requests
Batch operations use asyncio.gather for optimal parallelism (no manual tuning needed)
Set API keys to reduce reliance on scraping (faster + more reliable)
Caching is enabled by default with LRU and compression
Pure async implementation eliminates event loop overhead automatically
π€ Contributing
Fork the repository
Create feature branch (
git checkout -b feature/amazing-feature)Run tests (
pytest)Commit changes (
git commit -m 'Add amazing feature')Push to branch (
git push origin feature/amazing-feature)Open Pull Request
π License
MIT License - see LICENSE file for details.