The Reddit Research MCP Server transforms Reddit into an AI-powered research tool with semantic discovery, efficient data retrieval, and automated content analysis. Key capabilities include:
• Semantic community discovery: Search across 20,000+ indexed Reddit communities to find relevant subreddits based on any topic, going beyond native search limitations • Intelligent batch operations: Fetch posts from up to 15 subreddits simultaneously and retrieve detailed comments, reducing API calls by up to 70% • Automated research & analysis: Built-in AI agent (like Claude Code) conducts end-to-end research, analyzing 100+ posts and comments on complex topics • Professional report generation: Produces comprehensive Markdown reports with full citations, sentiment analysis, and complete traceability back to Reddit sources • Comprehensive search functionality: Reddit-wide search with time filters, content limits, and comment analysis with sorting options • Guided operational flow: Three-layer architecture (Discovery, Requirements, Execution) provides structured guidance for AI models, ensuring correct parameter usage and error prevention
Provides comprehensive access to Reddit content through a three-layer architecture for discovering communities, fetching posts and comments, searching across subreddits, and conducting thorough research and analysis of Reddit discussions.
mcp-name: io.github.king-of-the-grackles/reddit-research-mcp
🔍 Reddit Research MCP Server
Turn Reddit's chaos into structured insights with full citations
Version: 0.4.0
What's New in 0.4.0:
✨ Added 5 feed management operations for persistent research
🏗️ Enhanced three-layer architecture with comprehensive schemas
🤖 Added reddit_research prompt for automated workflows
🔧 Improved error handling and recovery suggestions
📝 Complete terminology migration (watchlist → feed)
Your customers are on Reddit right now, comparing you to competitors, sharing pain points, requesting features. But finding those insights means hours of manual searching with no way to cite your sources.
This MCP server turns Reddit into a queryable research database that generates reports with links to every claim. Get comprehensive market research, competitive analysis, and customer insights in minutes instead of hours.
🚀 Quick Setup (60 Seconds)
No credentials or configuration needed! Connect to our hosted server:
Claude Code
Cursor
OpenAI Codex CLI
Gemini CLI
Direct MCP Server URL
For other AI assistants: https://reddit-research-mcp.fastmcp.app/mcp
Related MCP server: Academic Paper Search MCP Server
🎯 What You Can Do
Competitive Analysis
→ Get a comprehensive report comparing sentiment, feature requests, pain points, and migration experiences with links to every mentioned discussion.
Customer Discovery
→ Discover unmet needs, feature gaps, and pricing concerns directly from your target market with citations to real user feedback.
Market Research
→ Track adoption trends, concerns, success stories, and emerging use cases with temporal analysis showing how opinions evolved.
Product Validation
→ Identify pain points and validate your solution with evidence from actual discussions, not assumptions.
Long-Term Monitoring
→ Create a feed with relevant subreddits and periodically check for new insights without starting research from scratch each time.
✨ Why This Server?
Built for decision-makers who need evidence-based insights. Every report links back to actual Reddit posts and comments. When you say "users are complaining about X," you'll have the receipts to prove it. Check the /reports folder for examples of deep-research reports with full citation trails.
Zero-friction setup designed for non-technical users. Most MCP servers require cloning repos, managing Python environments, and hunting for API keys in developer dashboards. This one? Just paste the URL into Claude and start researching. Our hosted solution means no terminal commands, no credential management, no setup headaches.
Semantic search across 20,000+ active subreddits. Reddit's API caps at 250 search results - useless for comprehensive research. We pre-indexed every active subreddit (2k+ members, active in last 7 days) with vector embeddings. Now you search conceptually across all of Reddit, finding relevant communities you didn't even know existed. Built with the layered abstraction pattern for scalability.
Persistent research management. Save your subreddit lists, analyses, and configurations into feeds for ongoing monitoring. Track what matters without starting from scratch each time. Perfect for long-term competitive analysis, market research campaigns, and product validation projects.
📊 Server Statistics
MCP Tools: 3 (discover_operations, get_operation_schema, execute_operation)
Reddit Operations: 5 (discover, search, fetch_posts, fetch_multiple, fetch_comments)
Feed Operations: 5 (create, list, get, update, delete)
Indexed Subreddits: 20,000+ (2k+ members, active in last 7 days)
MCP Prompts: 1 (reddit_research for automated workflows)
Resources: 1 (reddit://server-info for comprehensive documentation)
📚 Specifications
Some of the AI-generated specs that were used to build this project with Claude Code:
📖 Architecture Overview - System design and component interaction
🤖 Research Agent Details - Agent implementation patterns
🔍 Deep Research Architecture - Research workflow and citation system
🗄️ ChromaDB Proxy Architecture - Vector search and authentication layer
Technical Details
This server follows the layered abstraction pattern for scalability and self-documentation:
Layer 1: Discovery
Purpose: See what operations are available and get workflow recommendations.
Returns: List of all operations (Reddit research + feed management) with descriptions and recommended workflows.
Layer 2: Schema Inspection
Purpose: Understand parameter requirements, validation rules, and see examples before executing.
Returns: Complete schema with parameter types, descriptions, ranges, and usage examples.
Layer 3: Execution
Purpose: Perform the actual operation with validated parameters.
Returns: Operation results wrapped in {"success": bool, "data": ...} format.
Why This Pattern?
Self-documenting: Operations describe their own requirements
Version-safe: Schema changes don't break existing clients
Extensible: Add new operations without changing core tools
Type-safe: Full validation before execution
Example Workflow
discover_subreddits
Find relevant communities using semantic vector search across 20,000+ indexed subreddits.
Returns: Communities with confidence scores (0-1), match tiers, subscribers, and descriptions.
Use for: Starting any research project, finding niche communities, validating topic coverage.
search_subreddit
Search for posts within a specific subreddit.
Returns: Posts matching the search query with scores, comments, and timestamps.
Use for: Deep-diving into specific communities, finding historical discussions.
fetch_posts
Get posts from a single subreddit by listing type (hot, new, top, rising).
Returns: Recent posts from the subreddit with scores, comments, and authors.
Use for: Monitoring specific communities, trend analysis, content curation.
fetch_multiple
70% more efficient - Batch fetch posts from multiple subreddits concurrently.
Returns: Posts from all specified subreddits, organized by community.
Use for: Comparative analysis, multi-community research, feed monitoring.
fetch_comments
Get complete comment tree for deep analysis of discussions.
Returns: Post details + nested comment tree with scores, authors, and timestamps.
Use for: Understanding community sentiment, identifying expert opinions, analyzing debates.
Feeds let you save research configurations for ongoing monitoring. Perfect for long-term projects, competitive analysis, and market research campaigns.
create_feed
Save discovered subreddits with analysis and metadata.
Returns: Created feed with UUID, timestamps, and metadata.
Use for: Starting long-term research projects, saving competitive analysis configurations.
list_feeds
View all your saved feeds with pagination.
Returns: Array of feeds with metadata, sorted by recently viewed.
Use for: Managing multiple research projects, reviewing saved configurations.
get_feed
Retrieve a specific feed by ID.
Returns: Complete feed details including subreddits, analysis, and metadata.
Use for: Resuming research projects, reviewing feed configurations.
update_feed
Modify feed name, subreddits, or analysis (partial updates supported).
Returns: Updated feed with new timestamps.
Use for: Refining research scope, adding newly discovered communities.
delete_feed
Remove a feed permanently.
Returns: Confirmation of deletion.
Use for: Cleaning up completed projects, removing outdated configurations.
Feed Workflow Example
This server uses Descope OAuth2 for secure authentication:
Type: OAuth2 with Descope provider
Scope: Read-only access to public Reddit content
Setup: No Reddit credentials needed - server handles authentication
Token: Automatically managed by your MCP client
Privacy: Only accesses public Reddit data, no personal information collected
Your AI assistant will prompt for authentication on first use. The process takes ~30 seconds and only needs to be done once.
The server provides intelligent recovery suggestions for common errors:
404 Not Found
Cause: Subreddit doesn't exist or name is misspelled
Recovery: Verify the subreddit name or use discover_subreddits to find communities
429 Rate Limited
Cause: Too many requests to Reddit API (60 requests/minute limit)
Recovery: Reduce the limit parameter or wait 30 seconds before retrying
403 Private/Forbidden
Cause: Subreddit is private, quarantined, or banned Recovery: Try other communities from your discovery results
422 Validation Error
Cause: Parameters don't match schema requirements
Recovery: Use get_operation_schema() to check parameter types and validation rules
401 Authentication Required
Cause: Descope token expired or invalid Recovery: Re-authenticate when prompted by your AI assistant
This project uses:
Python 3.11+ with type hints
FastMCP for the server framework
Vector search via authenticated proxy (Render.com)
ChromaDB for semantic search
PRAW for Reddit API interaction
httpx for HTTP client requests (feed operations)
Stop guessing. Start knowing what your market actually thinks.