Research Powerpack MCP is an AI-powered research toolkit that automates web searching, Reddit mining, URL scraping, and information synthesis to provide structured context for large language models.
Core Tools:
Batch Web Search (
web_search): Execute 3-100 parallel Google searches with advanced operators (site:,-exclude,"exact phrase",filetype:), delivering 10 CTR-ranked results per keywordReddit Search (
search_reddit): Find Reddit discussions via Google with 10-50 parallel queries, supporting date filters and Reddit-specific operators (intitle:,r/subreddit)Reddit Post Fetching (
get_reddit_post): Retrieve 2-50 posts with smart comment allocation (1,000-comment budget distributed automatically), optional AI extraction, and upvote-sorted commentsUniversal Web Scraping (
scrape_links): Extract content from 1-50 URLs with automatic fallback (basic → JavaScript rendering → geo-targeting), AI-powered extraction, and custom instructions for handling SPAs, paywalls, and geo-restrictionsDeep Research (
deep_research): Process 2-10 complex questions in parallel with AI synthesis, web search, citations, and 32K token budget distribution, supporting file attachments for code-related queries
Key Features:
Modular Architecture: Tools activate independently based on API keys provided (Serper for search, Reddit OAuth for posts, Scrape.do for scraping, OpenRouter for AI)
Zero-Crash Design: Missing API keys return setup instructions instead of errors
Smart Optimization: Bounded concurrency, token-aware allocation, CTR-weighted ranking, and aggressive LLM guidance for optimal tool usage
Cross-Platform Support: Works with Claude Desktop, Claude Code, Cursor, and Windsurf MCP clients
Free Tier Access: 2,500 Google searches/month, unlimited Reddit access, 1,000 scraping credits/month
Ideal for: Technology decisions, competitive analysis, debugging obscure errors, and multi-angle research workflows.
Enables batch web search across up to 100 keywords in parallel using Google search via the Serper API, with support for search operators and CTR-weighted ranking to identify authoritative sources.
Supports OpenAI models through OpenRouter for AI-powered deep research synthesis and intelligent content extraction from scraped web pages.
Supports Perplexity's sonar-deep-research model through OpenRouter as the default model for conducting comprehensive AI-powered research with web search and citations.
Provides tools for searching Reddit discussions via Google, fetching posts with comments using the Reddit OAuth API, and smart comment allocation across multiple posts with upvote-based sorting.
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@Research Powerpack MCPresearch the best React state management libraries in 2024"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
🧭 Quick Navigation
⚡ Get Started • ✨ Key Features • 🎮 Usage & Examples • ⚙️ API Key Setup • 🆚 Why This Slaps
research-powerpack-mcp is the research assistant your AI wishes it had. Stop asking your LLM to guess about things it doesn't know. This MCP server acts like a senior researcher, searching the web, mining Reddit discussions, scraping documentation, and synthesizing everything into perfectly structured context so your AI can actually give you answers worth a damn.
How it slaps:
You: "What's the best database for my use case?"
AI + Powerpack: Searches Google, mines Reddit threads, scrapes docs, synthesizes findings.
You: Get an actually informed answer with real community opinions and citations.
Result: Ship better decisions. Skip the 47 browser tabs.
💥 Why This Slaps Other Methods
Manually researching is a vibe-killer. research-powerpack-mcp makes other methods look ancient.
We're not just fetching random pages. We're building high-signal, low-noise context with CTR-weighted ranking, smart comment allocation, and intelligent token distribution that prevents massive responses from breaking your LLM's context window.
🚀 Get Started in 60 Seconds
1. Install
2. Configure Your MCP Client
Client | Config File | Docs |
🖥️ Claude Desktop |
| |
⌨️ Claude Code |
| |
🎯 Cursor |
| |
🏄 Windsurf | MCP settings |
Claude Desktop
Add to your claude_desktop_config.json:
or quick install (for MacOS):
Claude Code (CLI)
One command to rule them all:
Or manually add to ~/.claude.json:
Cursor/Windsurf
Add to .cursor/mcp.json or equivalent:
✨ Zero Crash Promise: Missing API keys? No problem. The server always starts. Tools just return helpful setup instructions instead of exploding.
✨ Feature Breakdown: The Secret Sauce
Feature | What It Does | Why You Care |
🔍 Batch Search
| Search Google for up to 100 queries simultaneously | Cover every angle of a topic in one shot |
📊 CTR Ranking
| Identifies URLs that appear across multiple searches | Surfaces high-consensus authoritative sources |
💬 Reddit Mining
| Google-powered Reddit search + native API fetching | Get actual user experiences, not marketing fluff |
🎯 Smart Allocation
| 1,000 comment budget distributed across posts | Deep dive on 2 posts or quick scan on 50 |
🌐 Universal Scraping
| Auto-fallback: basic → JS render → geo-targeting | Handles SPAs, paywalls, and geo-restricted content |
🧠 Deep Research
| Batch research with web search and citations | Get comprehensive answers to complex questions |
🧩 Modular Design
| Each tool works independently | Pay only for the APIs you actually use |
🎮 Tool Reference
web_search
Batch web search using Google via Serper API. Search up to 100 keywords in parallel.
Parameter | Type | Required | Description |
|
| Yes | Search queries (1-100). Use distinct keywords for maximum coverage. |
Supports Google operators: site:, -exclusion, "exact phrase", filetype:
search_reddit
Search Reddit via Google with automatic site:reddit.com filtering.
Parameter | Type | Required | Description |
|
| Yes | Search queries (max 10) |
|
| No | Filter results after date (YYYY-MM-DD) |
Search operators: intitle:keyword, "exact phrase", OR, -exclude
get_reddit_post
Fetch Reddit posts with smart comment allocation (1,000 comment budget distributed automatically).
Parameter | Type | Required | Default | Description |
|
| Yes | — | Reddit post URLs (2-50) |
|
| No |
| Whether to fetch comments |
|
| No | auto | Override comment allocation |
Smart Allocation:
2 posts → ~500 comments/post (deep dive)
10 posts → ~100 comments/post
50 posts → ~20 comments/post (quick scan)
scrape_links
Universal URL content extraction with automatic fallback modes.
Parameter | Type | Required | Default | Description |
|
| Yes | — | URLs to scrape (3-50) |
|
| No |
| Timeout per URL (seconds) |
|
| No |
| Enable AI extraction |
|
| No | — | Extraction instructions for AI |
Automatic Fallback: Basic → JS rendering → JS + US geo-targeting
deep_research
AI-powered batch research with web search and citations.
Parameter | Type | Required | Description |
|
| Yes | Research questions (2-10) |
|
| Yes | The research question |
|
| No | Files to include as context |
Token Allocation: 32,000 tokens distributed across questions:
2 questions → 16,000 tokens/question (deep dive)
10 questions → 3,200 tokens/question (rapid multi-topic)
⚙️ Environment Variables & Tool Availability
Research Powerpack uses a modular architecture. Tools are automatically enabled based on which API keys you provide:
ENV Variable | Tools Enabled | Free Tier |
|
| 2,500 queries/mo |
|
| Unlimited |
|
| 1,000 credits/mo |
|
| Pay-as-you-go |
| Model for | Default: |
| Model for AI extraction in | Default: |
Configuration Examples
🔑 API Key Setup Guides
What you get
Fast Google search results via API
Enables
web_searchandsearch_reddittools
Setup Steps
Go to serper.dev
Click "Get API Key" (top right)
Sign up with email or Google
Copy your API key from the dashboard
Add to your config:
SERPER_API_KEY=your_key_here
Pricing
Free: 2,500 queries/month
Paid: $50/month for 50,000 queries
What you get
Full Reddit API access
Fetch posts and comments with upvote sorting
Enables
get_reddit_posttool
Setup Steps
Go to reddit.com/prefs/apps
Scroll down and click "create another app..."
Fill in:
Name:
research-powerpack(or any name)App type: Select "script" (important!)
Redirect URI:
http://localhost:8080
Click "create app"
Copy your credentials:
Client ID: The string under your app name
Client Secret: The "secret" field
Add to your config:
REDDIT_CLIENT_ID=your_client_id REDDIT_CLIENT_SECRET=your_client_secret
What you get
JavaScript rendering support
Geo-targeting and CAPTCHA handling
Enables
scrape_linkstool
Setup Steps
Go to scrape.do
Click "Start Free"
Sign up with email
Copy your API key from the dashboard
Add to your config:
SCRAPEDO_API_KEY=your_key_here
Credit Usage
Basic scrape: 1 credit
JavaScript rendering: 5 credits
Geo-targeting: +25 credits
What you get
Access to 100+ AI models via one API
Enables
deep_researchtoolEnables AI extraction in
scrape_links
Setup Steps
Go to openrouter.ai
Sign up with Google/GitHub/email
Go to openrouter.ai/keys
Click "Create Key"
Copy the key (starts with
sk-or-...)Add to your config:
OPENROUTER_API_KEY=sk-or-v1-xxxxx
Recommended Models for Deep Research
Recommended Models for AI Extraction (use_llm in scrape_links)
Note:
RESEARCH_MODELandLLM_EXTRACTION_MODELare independent. You can use a powerful model for deep research and a faster/cheaper model for content extraction, or vice versa.
🔥 Recommended Workflows
Research a Technology Decision
Competitive Analysis
Debug an Obscure Error
🔥 Enable Full Power Mode
For the best research experience, configure all four API keys:
This unlocks:
5 research tools working together
AI-powered content extraction in scrape_links
Deep research with web search and citations
Complete Reddit mining (search → fetch → analyze)
Total setup time: ~10 minutes. Total free tier value: ~$50/month equivalent.
🛠️ Development
🏗️ Architecture (v3.4.0+)
The codebase uses a YAML-driven configuration system with aggressive LLM optimization (v3.5.0+):
Core Architecture
Component | File | Purpose |
Tool Definitions |
| Single source of truth for all tool metadata |
Handler Registry |
| Declarative tool registration + |
YAML Loader |
| Parses YAML, generates MCP-compatible definitions (cached) |
Concurrency Utils |
| Bounded parallel execution ( |
Shared Utils |
| Common utility functions |
Adding a new tool:
Add tool definition to
tools.yamlCreate handler in
src/tools/Register in
src/tools/registry.ts
See docs/refactoring/04-migration-guide.md for detailed instructions.
Performance & Stability (v3.5.1+)
All parallel operations use bounded concurrency to prevent CPU spikes and API rate limits:
Operation | Before | After |
Reddit search queries | 50 concurrent | 8 concurrent |
Web scraping batches | 30 concurrent | 10 concurrent |
Deep research questions | Unbounded | 3 concurrent |
Reddit post fetching | 10 concurrent | 5 concurrent |
File attachments | Unbounded | 5 concurrent |
Additional optimizations:
YAML config cached in memory (no repeated disk reads)
Async file I/O (no event loop blocking)
Pre-compiled regex patterns for hot paths
Reddit auth token deduplication (prevents concurrent token requests)
LLM Optimization (v3.5.0+)
All tools include aggressive guidance to force LLMs to use them optimally:
Feature | Description |
Configurable Limits | All min/max values in YAML ( |
BAD vs GOOD Examples | Every tool shows anti-patterns and perfect usage |
Aggressive Phrasing | Changed from "you can" to "you MUST" |
Visual Formatting | Emoji headers, section dividers, icons for visual scanning |
Templates | Structured formats for questions, extractions, file descriptions |
Key Enhancements:
search_reddit: Minimum 10 queries (was 3), 10-category formuladeep_research: 7-section question template, file attachment requirementsscrape_links: Extraction template with OR statements, use_llm=true pushweb_search: Minimum 3 keywords, search operator examplesfile_attachments: Numbered 5-section description template
See docs/refactoring/07-llm-optimization-summary.md for full details.
🔥 Common Issues & Quick Fixes
Problem | Solution |
Tool returns "API key not configured" | Add the required ENV variable to your MCP config. The error message tells you exactly which key is missing. |
Reddit posts returning empty | Check your |
Scraping fails on JavaScript sites | This is expected for first attempt. The tool auto-retries with JS rendering. If still failing, the site may be blocking scrapers. |
Deep research taking too long | Use a faster model like |
Token limit errors | Reduce the number of URLs/questions per request. The tool distributes a fixed token budget. |
Built with 🔥 because manually researching for your AI is a soul-crushing waste of time.
MIT © Yiğit Konur