Skip to main content
Glama

๐Ÿค– Second Opinion MCP

License Python MCP

๐ŸŽฏ Get instant second opinions from 17 AI platforms and 800,000+ models

OpenAI โ€ข Gemini โ€ข Grok โ€ข Claude โ€ข HuggingFace โ€ข DeepSeek โ€ข OpenRouter โ€ข Mistral โ€ข Together AI โ€ข Cohere โ€ข Groq โ€ข Perplexity โ€ข Replicate โ€ข AI21 Labs โ€ข Stability AI โ€ข Fireworks AI โ€ข Anyscale


๐Ÿš€ What it does

This MCP server allows Claude to consult other AI models for different perspectives on:

  • Coding problems - Compare approaches across models

  • Creative writing - Get diverse style feedback

  • Problem solving - Validate logic and reasoning

  • Cross-model analysis - See how different AIs tackle the same task

  • Group discussions - Host AI debates with multiple models

  • Custom model access - Use any HuggingFace model via Inference API

โœจ Version 5.0 Features & Improvements

๐ŸŽญ NEW: AI Personality System

  • 5 Distinct Personalities: honest, freind, coach, wise, creative

  • Intelligent Model Matching: Each personality uses models best suited for their character

  • Always Available: Works with any configured AI provider

๐Ÿง  NEW: Intelligent Model Selection

  • Quality-Based Ranking: 34+ models ranked by capability (Grok-4 โ†’ Gemini Pro โ†’ GPT-4.1)

  • Smart Defaults: Automatically selects the best available model

  • Personality Optimization: Different models for different personality types

๐Ÿ—๏ธ NEW: Modular Architecture

  • 5 Clean Files: Replaced 51k+ token monolith with maintainable modules

  • Professional Structure: client_manager.py, ai_providers.py, conversation_manager.py, mcp_server.py, main.py

  • JSON Configuration: Easy model priority updates via model_priority.json

๐Ÿš€ Major Platform Integrations

  • ๐ŸŽญ Replicate: Access to open-source models including Llama 2, CodeLlama, Mistral, and more

  • ๐ŸŒŸ AI21 Labs: Jamba 1.5 models with advanced reasoning capabilities

  • ๐ŸŽจ Stability AI: StableLM models including code-specialized variants

  • ๐Ÿ”ฅ Fireworks AI: Ultra-fast inference for popular open-source models

  • ๐Ÿš€ Anyscale: Ray-powered LLM serving with enterprise-grade reliability

๐Ÿ†• Enhanced Existing Platform Support

  • ๐Ÿค– Mistral AI: Direct access to Mistral's latest models including mistral-large-latest and codestral-latest

  • ๐Ÿ”— Together AI: Access to 200+ open-source models with fast inference

  • ๐Ÿง  Cohere: Enterprise-grade language models with Command R+ and Command R

  • โšก Groq Fast: Ultra-fast inference API for lightning-quick responses

  • ๐Ÿ” Perplexity AI: Web-connected AI with real-time search capabilities

๐Ÿ”ง Previous Bug Fixes (v3.0)

  • Fixed HuggingFace Models: Completely rebuilt HuggingFace integration with advanced retry logic, better model format detection, and comprehensive error handling

  • Fixed Gemini Blank Responses: Enhanced Gemini conversation handling to prevent empty responses in long chats with smart fallback and retry mechanisms

  • Improved Error Handling: Better error messages with helpful suggestions for troubleshooting

๐Ÿค– HuggingFace Integration (Enhanced)

Access any of the 800,000+ models on HuggingFace Hub via their Inference API with improved reliability:

  • meta-llama/Llama-3.1-8B-Instruct - Fast and reliable

  • meta-llama/Llama-3.1-70B-Instruct - Powerful reasoning

  • mistralai/Mistral-7B-Instruct-v0.3 - Efficient French-developed model

  • Qwen/Qwen2.5-7B-Instruct - Alibaba's latest model

๐Ÿง  DeepSeek Models

Get opinions from DeepSeek's powerful reasoning models:

  • deepseek-chat (DeepSeek-V3) - Fast and efficient

  • deepseek-reasoner (DeepSeek-R1) - Advanced reasoning

๐Ÿค” Grok 4 Thinking

Access xAI's thinking models with enhanced reasoning:

  • grok-4 - Latest flagship model

  • grok-3-thinking - Step-by-step reasoning model, last gen

  • grok-3-mini - Lightweight thinking model with reasoning_effort control, last ge

๐ŸŽญ Group Discussions (Enhanced)

Start multi-AI discussions where models can see and respond to each other's input:

> "Start a group discussion about the future of AI with GPT-4.1, Claude-4, Mistral, and Perplexity"

๐Ÿ”ง Enhanced Performance

  • Longer responses: Increased max_tokens (4000 default) to prevent cut-off responses

  • Better error handling: More robust API interactions with exponential backoff

  • Conversation persistence: Enhanced memory management with better context handling

  • Smart retry logic: Automatic retries with progressive delays for better reliability

๐Ÿ“‹ Requirements

Python Claude Desktop API Keys

  • Python 3.8+ - Programming language runtime

  • Claude Desktop or Claude Code - Choose your preferred MCP integration

  • API Keys - For any combination of the 16 supported AI platforms

๐Ÿ“ฆ Installation Guide

๐Ÿš€ Quick Start

  1. ๐Ÿ“ฅ Clone the repository

    git clone https://github.com/ProCreations-Official/second-opinion.git
    cd second-opinion
  2. โš™๏ธ Install dependencies

    pip install -r requirements.txt
  3. ๐Ÿ”‘ Get API Keys

    Platform

    Link

    Required

    OpenAI

    platform.openai.com

    โญ Popular

    Gemini

    aistudio.google.com

    โญ Popular

    Grok

    x.ai

    ๐Ÿ”ฅ Best benchmarks

    Claude

    anthropic.com

    ๐Ÿง  Advanced

    HuggingFace

    huggingface.co

    ๐Ÿค— 800k+ Models

    DeepSeek

    deepseek.com

    ๐Ÿ”ฌ Reasoning

    OpenRouter

    openrouter.ai

    ๐ŸŒ 200+ Models via One API

    Mistral

    console.mistral.ai

    ๐Ÿ‡ซ๐Ÿ‡ท European, fast

    Together AI

    api.together.xyz

    ๐Ÿ”— 200+ Models

    Cohere

    dashboard.cohere.com

    ๐Ÿข Enterprise

    Groq

    console.groq.com

    โšก Ultra-Fast

    Perplexity

    perplexity.ai

    ๐Ÿ” Web Search

    Replicate

    replicate.com

    ๐ŸŽญ Open Source

    AI21 Labs

    studio.ai21.com

    ๐Ÿงฌ Jamba Models

    Stability AI

    platform.stability.ai

    ๐ŸŽจ StableLM

    Fireworks AI

    fireworks.ai

    ๐Ÿ”ฅ Fast Inference

    Anyscale

    console.anyscale.com

    ๐Ÿš€ Ray Serving

  4. ๐Ÿ”ง Choose Your Integration Method

    Select the method that matches your Claude setup:

For Claude Desktop Users

Add this to your Claude Desktop MCP configuration:

{
  "mcpServers": {
    "second-opinion": {
      "command": "python3",
      "args": ["/path/to/your/main.py"],
      "env": {
        "OPENAI_API_KEY": "your_openai_key_here",
        "GEMINI_API_KEY": "your_gemini_key_here",
        "GROK_API_KEY": "your_grok_key_here",
        "CLAUDE_API_KEY": "your_claude_key_here",
        "HUGGINGFACE_API_KEY": "your_huggingface_key_here",
        "DEEPSEEK_API_KEY": "your_deepseek_key_here",
        "OPENROUTER_API_KEY": "your_openrouter_key_here",
        "MISTRAL_API_KEY": "your_mistral_key_here",
        "TOGETHER_API_KEY": "your_together_key_here",
        "COHERE_API_KEY": "your_cohere_key_here",
        "GROQ_FAST_API_KEY": "your_groq_key_here",
        "PERPLEXITY_API_KEY": "your_perplexity_key_here",
        "REPLICATE_API_TOKEN": "your_replicate_key_here",
        "AI21_API_KEY": "your_ai21_key_here",
        "STABILITY_API_KEY": "your_stability_key_here",
        "FIREWORKS_API_KEY": "your_fireworks_key_here",
        "ANYSCALE_API_KEY": "your_anyscale_key_here"
      }
    }
  }
}

๐Ÿ’ก Note: You only need to add API keys for the services you want to use. Missing keys will simply disable those specific features.

๐Ÿ”„ Restart Claude Desktop after configuration.

๐ŸŽฏ For Claude Code CLI Users

Claude Code MCP

๐Ÿ“ฆ Prerequisites

First, ensure Claude Code CLI is installed globally:

npm install -g @anthropic-ai/claude-code

๐Ÿš€ Installation Methods

Use the claude mcp add command to add the Second Opinion server:

# Navigate to your second-opinion directory
cd /path/to/your/second-opinion

# Add the MCP server with environment variables (use -e for each API key)
claude mcp add second-opinion -s user \
  -e OPENAI_API_KEY=your_openai_key_here \
  -e GEMINI_API_KEY=your_gemini_key_here \
  -e GROK_API_KEY=your_grok_key_here \
  -e CLAUDE_API_KEY=your_claude_key_here \
  -e HUGGINGFACE_API_KEY=your_huggingface_key_here \
  -e DEEPSEEK_API_KEY=your_deepseek_key_here \
  -e OPENROUTER_API_KEY=your_openrouter_key_here \
  -e MISTRAL_API_KEY=your_mistral_key_here \
  -e TOGETHER_API_KEY=your_together_key_here \
  -e COHERE_API_KEY=your_cohere_key_here \
  -e GROQ_FAST_API_KEY=your_groq_key_here \
  -e PERPLEXITY_API_KEY=your_perplexity_key_here \
  -e REPLICATE_API_TOKEN=your_replicate_key_here \
  -e AI21_API_KEY=your_ai21_key_here \
  -e STABILITY_API_KEY=your_stability_key_here \
  -e FIREWORKS_API_KEY=your_fireworks_key_here \
  -e ANYSCALE_API_KEY=your_anyscale_key_here \
  -- /path/to/your/second-opinion/run.sh

๐Ÿ’ก Quick Setup: You only need to include -e flags for the API keys you have. For example, if you only have OpenAI and Gemini keys:

claude mcp add second-opinion -s user \
  -e OPENAI_API_KEY=your_openai_key_here \
  -e GEMINI_API_KEY=your_gemini_key_here \
  -- /path/to/your/second-opinion/run.sh

Alternatively, you can manually add the server to your .claude.json file:

{
  "mcpServers": {
    "second-opinion": {
      "type": "stdio",
      "command": "/path/to/your/second-opinion/run.sh",
      "env": {
        "OPENAI_API_KEY": "your_openai_key_here",
        "GEMINI_API_KEY": "your_gemini_key_here",
        "GROK_API_KEY": "your_grok_key_here",
        "CLAUDE_API_KEY": "your_claude_key_here",
        "HUGGINGFACE_API_KEY": "your_huggingface_key_here",
        "DEEPSEEK_API_KEY": "your_deepseek_key_here",
        "OPENROUTER_API_KEY": "your_openrouter_key_here",
        "MISTRAL_API_KEY": "your_mistral_key_here",
        "TOGETHER_API_KEY": "your_together_key_here",
        "COHERE_API_KEY": "your_cohere_key_here",
        "GROQ_FAST_API_KEY": "your_groq_key_here",
        "PERPLEXITY_API_KEY": "your_perplexity_key_here",
        "REPLICATE_API_TOKEN": "your_replicate_key_here",
        "AI21_API_KEY": "your_ai21_key_here",
        "STABILITY_API_KEY": "your_stability_key_here",
        "FIREWORKS_API_KEY": "your_fireworks_key_here",
        "ANYSCALE_API_KEY": "your_anyscale_key_here"
      }
    }
  }
}

๐Ÿ”ง Why Use run.sh?

Feature

Benefit

๐Ÿ“ฆ Dependency Management

Automatically installs/updates requirements

๐Ÿ›ก๏ธ Error Handling

Checks for python3 availability and required files

๐Ÿ”„ Cross-platform

Works better than direct Python execution

โšก Reliability

Ensures consistent execution regardless of system

โœ… Verification

Check that your MCP server is properly installed:

claude mcp list

You should see second-opinion in the list of available MCP servers.

๐Ÿ”‘ Environment Variables: You only need to add API keys for the services you want to use. Missing keys will simply disable those specific AI platforms. The server will work with any combination of available API keys.

Total Models Platforms Updated

๐Ÿš€ Premium AI Platforms

Model

Description

Best For

o4-mini

Fast reasoning model

โšก Quick reasoning

gpt-4.1

Latest flagship non-reasoning model

๐ŸŽฏ General tasks

gpt-4o

Multimodal powerhouse

๐Ÿ–ผ๏ธ Vision + text

gpt-4o-mini

Lightweight GPT-4o

๐Ÿ’ฐ Cost-effective

Model

Description

Best For

gemini-2.5-flash-lite-preview-06-17

Lightweight and fast

โšก Quick responses

gemini-2.5-flash

Advanced reasoning and efficiency

๐Ÿงฎ Complex analysis

gemini-2.5-pro

Most capable Gemini model

๐Ÿง  Advanced tasks

Model

Description

Best For

grok-4

Latest flagship model

๐ŸŽฏ General excellence, best

grok-3-thinking

Step-by-step reasoning

๐Ÿค” Deep thinking (use grok 4)

grok-3-mini

Lightweight model

๐Ÿ’ก Quick insights

Model

Description

Best For

claude-4-opus-20250522

Most advanced Claude

๐Ÿง  Complex reasoning

claude-4-sonnet-20250522

Versatile general tasks

โš–๏ธ Balanced performance

claude-3-7-sonnet-20250224

Stable and reliable

๐Ÿ›ก๏ธ Production use

claude-3-5-sonnet-20241022

Efficient, lighter model

๐Ÿ’จ Fast responses

Featured Models:

Model

Description

Best For

meta-llama/Llama-3.1-8B-Instruct

Fast Meta model

โšก Speed

meta-llama/Llama-3.1-70B-Instruct

Powerful reasoning

๐Ÿง  Complex tasks

mistralai/Mistral-7B-Instruct-v0.3

French-developed

๐Ÿ‡ซ๐Ÿ‡ท European AI

Qwen/Qwen2.5-7B-Instruct

Alibaba's latest

๐Ÿข Enterprise

๐ŸŒŸ Special: Access to any model on HuggingFace Hub that supports text generation

๐Ÿ”ฌ Specialized AI Platforms

Model

Description

Best For

deepseek-chat

DeepSeek-V3 general tasks

๐Ÿ’ฌ Conversations

deepseek-reasoner

DeepSeek-R1 advanced reasoning

๐Ÿง  Complex logic

Model

Description

Best For

anthropic/claude-3-5-sonnet-20241022

OpenRouter access to Claude 3.5 Sonnet

๐ŸŽฏ Balanced excellence

openai/gpt-4-turbo

OpenRouter access to GPT-4 Turbo

๐Ÿง  Advanced reasoning

google/gemini-pro-1.5

OpenRouter access to Gemini Pro 1.5

๐Ÿ” Long context

meta-llama/llama-3.1-405b-instruct

OpenRouter access to largest Llama

๐Ÿฆฃ Massive scale

mistralai/mistral-large

OpenRouter access to Mistral Large

๐Ÿ‡ซ๐Ÿ‡ท European excellence

perplexity/llama-3.1-sonar-huge-128k-online

Web-connected via OpenRouter

๐ŸŒ Current information

๐ŸŒŸ Special: Access to 200+ models from multiple providers through a single OpenRouter API

Model

Description

Best For

mistral-large-latest

Most powerful Mistral

๐ŸŽฏ Top performance

mistral-small-latest

Fast and cost-effective

๐Ÿ’ฐ Budget-friendly

mistral-medium-latest

Balanced performance

โš–๏ธ General use

codestral-latest

Code generation specialist

๐Ÿ’ป Programming

Model

Description

Best For

meta-llama/Llama-3.1-8B-Instruct-Turbo

Fast Llama turbo

โšก Speed

meta-llama/Llama-3.1-70B-Instruct-Turbo

Powerful Llama turbo

๐Ÿš€ Performance

meta-llama/Llama-3.1-405B-Instruct-Turbo

Largest Llama model

๐Ÿฆฃ Massive scale

mistralai/Mixtral-8x7B-Instruct-v0.1

Mixture of experts

๐ŸŽญ Specialized tasks

Qwen/Qwen2.5-72B-Instruct-Turbo

Alibaba's fast model

๐Ÿข Enterprise

Cohere (Enterprise-grade)

Model

Description

Best For

command-r-plus

Most capable Cohere

๐ŸŽฏ Enterprise

command-r

Balanced performance

โš–๏ธ General business

command

Standard command model

๐Ÿ’ผ Basic tasks

Groq (Ultra-fast inference)

Model

Description

Best For

llama-3.1-70b-versatile

Fast 70B Llama

โšก Quick power

llama-3.1-8b-instant

Lightning-fast 8B

๐Ÿƒ Instant responses

mixtral-8x7b-32768

Fast Mixtral variant

๐ŸŽญ Quick specialization

gemma2-9b-it

Google's Gemma model

๐Ÿ” Search-optimized

Perplexity AI (Web-connected)

Model

Description

Best For

llama-3.1-sonar-large-128k-online

Web search + large context

๐ŸŒ Research

llama-3.1-sonar-small-128k-online

Web search + fast responses

๐Ÿ” Quick search

llama-3.1-sonar-large-128k-chat

Pure chat without web

๐Ÿ’ฌ Conversations

llama-3.1-sonar-small-128k-chat

Fast chat model

โšก Quick chat

Replicate (Open-source hosting)

Model

Description

Best For

meta/llama-2-70b-chat

Large Llama 2 chat

๐Ÿฆฃ Powerful chat

meta/llama-2-13b-chat

Medium Llama 2 chat

โš–๏ธ Balanced

meta/codellama-34b-instruct

Code-specialized Llama

๐Ÿ’ป Programming

microsoft/wizardcoder-34b

Microsoft's coding model

๐Ÿง™ Code magic

AI21 Labs (Advanced reasoning)

Model

Description

Best For

jamba-1.5-large

State-space capabilities

๐Ÿงฌ Complex reasoning

jamba-1.5-mini

Compact Jamba model

๐Ÿ’Ž Efficient reasoning

j2-ultra

Jurassic-2 Ultra model

๐Ÿฆ• Powerful

j2-mid

Jurassic-2 Mid model

โš–๏ธ Balanced

Stability AI (StableLM family)

Model

Description

Best For

stablelm-2-zephyr-1_6b

Efficient 1.6B parameter

โšก Lightweight

stable-code-instruct-3b

Code-specialized 3B

๐Ÿ’ป Programming

japanese-stablelm-instruct-beta-70b

Japanese language

๐Ÿ‡ฏ๐Ÿ‡ต Japanese tasks

stablelm-zephyr-3b

Balanced 3B parameter

โš–๏ธ General use

Fireworks AI (Ultra-fast inference)

Model

Description

Best For

accounts/fireworks/models/llama-v3p1-70b-instruct

Fast Llama 3.1 70B

๐Ÿ”ฅ Speed + power

accounts/fireworks/models/llama-v3p1-8b-instruct

Fast Llama 3.1 8B

โšก Quick responses

accounts/fireworks/models/mixtral-8x7b-instruct

Fast Mixtral model

๐ŸŽญ Fast specialization

accounts/fireworks/models/deepseek-coder-v2-lite-instruct

Code-specialized

๐Ÿ’ป Fast coding

Anyscale (Ray-powered serving)

Model

Description

Best For

meta-llama/Llama-2-70b-chat-hf

Enterprise Llama 2 70B

๐Ÿข Enterprise chat

meta-llama/Llama-2-13b-chat-hf

Enterprise Llama 2 13B

๐Ÿ’ผ Business tasks

codellama/CodeLlama-34b-Instruct-hf

Enterprise CodeLlama

๐Ÿ’ป Enterprise coding

mistralai/Mistral-7B-Instruct-v0.1

Enterprise Mistral

๐Ÿ‡ซ๐Ÿ‡ท Enterprise EU


โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆ

๐Ÿ’ก Usage Examples


๐ŸŽญ NEW: Personality Examples

๐Ÿ˜ค "Give me an honest opinion about this code" (brutally frank feedback)

๐Ÿ’• "I need some encouragement with this project" (supportive girlfriend mode)

๐Ÿ† "Help me stay motivated to finish this task" (motivational coach)

๐Ÿง™ "What's the deeper meaning behind this design pattern?" (ancient wisdom)

๐ŸŽจ "Think of a creative solution to this problem" (innovative thinking)

๐Ÿค– "Just give me the best available opinion" (automatic smart selection)

๐Ÿ”ฅ Popular Requests

๐Ÿ’ฌ "Get a second opinion from GPT-4.1 on this coding approach"

๐Ÿค” "What would Grok-4 think about this solution?" (NEW: Latest model)

โš–๏ธ "Compare how Claude-4-opus and gemini-2.5-flash would solve this problem"

๐Ÿค— "Get an opinion from meta-llama/Llama-3.1-70B-Instruct on HuggingFace"

๐ŸŒ "Get an OpenRouter opinion from anthropic/claude-3-5-sonnet-20241022"

๐Ÿง  "What does DeepSeek-reasoner think about this math problem?"

๐Ÿ‡ซ๐Ÿ‡ท "Ask Mistral-large-latest to review my code architecture"

โšก "Get a fast response from Groq's llama-3.1-8b-instant model"

๐ŸŒ "Use Perplexity's web search to research the latest AI developments"

๐Ÿข "What does Cohere's command-r-plus think about this business strategy?"

๐Ÿ”— "Get Together AI's Llama-405B opinion on this complex problem"

๐ŸŽญ Advanced Features

๐Ÿ—ฃ๏ธ "Start a group discussion about AI ethics with GPT-4.1, Claude-4, Mistral, and Perplexity"

๐Ÿ“Š "Cross-platform comparison of this algorithm across all 16 available platforms"

๐ŸŽญ "Get a Replicate opinion from meta/llama-2-70b-chat on this open-source approach"

๐Ÿงฌ "What does AI21's Jamba-1.5-large think about this reasoning problem?"

๐ŸŽจ "Ask Stability AI's StableLM about this code optimization"

๐Ÿ”ฅ "Get a super-fast response from Fireworks AI's Llama model"

๐Ÿš€ "Use Anyscale's enterprise-grade Llama serving for this complex task"

๐Ÿ”ง Available Tools

Single Model Opinions

  • get_openai_opinion - Get opinion from any OpenAI model

  • get_gemini_opinion - Get opinion from any Gemini model (enhanced with better conversation handling)

  • get_grok_opinion - Get opinion from any Grok model (includes thinking models)

  • get_claude_opinion - Get opinion from any Claude model

  • get_huggingface_opinion - Get opinion from any HuggingFace model (enhanced with better reliability)

  • get_deepseek_opinion - Get opinion from DeepSeek models

  • get_openrouter_opinion - Get opinion from 200+ models via OpenRouter (NEW)

  • get_mistral_opinion - Get opinion from Mistral AI models (NEW)

  • get_together_opinion - Get opinion from Together AI's 200+ models (NEW)

  • get_cohere_opinion - Get opinion from Cohere enterprise models (NEW)

  • get_groq_fast_opinion - Get ultra-fast responses from Groq (NEW)

  • get_perplexity_opinion - Get web-connected AI responses

  • get_replicate_opinion - Get opinion from Replicate's open-source models (NEW)

  • get_ai21_opinion - Get opinion from AI21 Labs' Jamba models (NEW)

  • get_stability_opinion - Get opinion from Stability AI's StableLM models (NEW)

  • get_fireworks_opinion - Get ultra-fast responses from Fireworks AI (NEW)

  • get_anyscale_opinion - Get enterprise-grade responses from Anyscale (NEW)

Model Comparisons

  • compare_openai_models - Compare multiple OpenAI models

  • compare_gemini_models - Compare multiple Gemini models

  • compare_grok_models - Compare multiple Grok models

  • compare_claude_models - Compare multiple Claude models

๐ŸŽญ NEW: Personality & Smart Default Tools

  • get_personality_opinion - Get AI responses with specific personality (honest, gf, coach, wise, creative)

  • get_default_opinion - Automatically uses the best available model (Grok-4 โ†’ Gemini Pro โ†’ GPT-4.1)

  • list_personalities - See all available AI personalities and their descriptions

Cross-Platform Features

  • cross_platform_comparison - Compare across all 17 AI platforms: OpenAI, Gemini, Grok, Claude, HuggingFace, DeepSeek, OpenRouter, Mistral, Together AI, Cohere, Groq Fast, Perplexity, Replicate, AI21 Labs, Stability AI, Fireworks AI & Anyscale

  • group_discussion - Multi-round discussions between AI models with shared context (supports all platforms)

Conversation Management

  • list_conversation_histories - See active conversation threads

  • clear_conversation_history - Reset conversation memory for specific models

๐Ÿง  Advanced Features

GPT 5 Thinking Models

For deeper reasoning, use thinking models:

> "Get a GPT 5 thinking opinion on this complex math problem with high reasoning effort"

This will use GPT 5 with reasoning effort set to high.

Group Discussions

Create AI debates and collaborative problem-solving:

> "Start a group discussion about renewable energy solutions with 3 rounds between GPT-4.1, Claude-4, Gemini, and DeepSeek"

Each AI can see previous responses and build on the discussion.

HuggingFace Model Access

Access cutting-edge open source models:

> "Get an opinion from Qwen/Qwen3-30B-A3B-Instruct-2507 about chatbot design patterns"

may be broken Perfect for testing specialized models or comparing open source vs proprietary AI.

๐Ÿ”’ Security

Your API keys stay private on your machine. The MCP server only sends model responses to the client, never your credentials.

๐Ÿ›Ÿ Troubleshooting

Import errors: Ensure you've installed all dependencies with pip install -r requirements.txt

API errors: Check that your API keys are correct and active

Server not connecting: Verify the file path in your MCP configuration

Cut-off responses: The new version uses 4000 max_tokens by default to prevent truncation

HuggingFace timeouts: Some models may take time to load. Try again after a few moments.

Model not available: Check if the HuggingFace model supports text generation or chat completion

๐Ÿš€ What's New in v5.0

  • โœ… ๐ŸŽญ AI Personality System: 5 distinct personalities (honest, gf, coach, wise, creative) with optimized model selection

  • โœ… ๐Ÿง  Intelligent Model Selection: Quality-ranked models with Grok-4 as top priority, smart defaults

  • โœ… ๐Ÿ—๏ธ Modular Architecture: Refactored from 51k+ token monolith into 5 maintainable modules

  • โœ… ๐Ÿ“Š JSON Configuration: Easy model priority updates via configuration files

  • โœ… ๐Ÿค– Always-Available Tools: Personality and default tools work with any provider setup

  • โœ… โšก Enhanced Performance: Optimized model selection and better error handling

  • โœ… ๐Ÿ”„ Future-Proof: Easy to add new models and update priorities

Previous Updates (v4.0)

  • โœ… 5 New Major AI Platforms: Replicate, AI21 Labs, Stability AI, Fireworks AI, Anyscale

  • โœ… 16 Total Platforms: Now supporting 16 different AI platforms and 800,000+ models

  • โœ… Advanced Reasoning: AI21 Labs' Jamba models with state-space architecture

  • โœ… Ultra-Fast Inference: Fireworks AI for blazing-fast open model serving

Previous Improvements (v3.0)

  • โœ… Major Bug Fixes: Fixed HuggingFace empty responses and Gemini blank chat issues

  • โœ… Enhanced HuggingFace: Completely rebuilt with advanced retry logic and better error handling

  • โœ… Improved Gemini: Smart conversation handling prevents blank responses in long chats

  • โœ… Web-Connected AI: Perplexity AI with real-time search capabilities

  • โœ… Enterprise Models: Cohere's command models for business use cases

๐Ÿค Contributing

Issues and pull requests welcome! This is an open-source project for the AI community.


โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“

๐Ÿง โœจ Built for developers who want access to the entire AI ecosystem at their fingertips

GitHub stars GitHub forks GitHub watchers

โญ Star us on GitHub โ€ข ๐Ÿด Fork the project โ€ข ๐Ÿ’– Contribute to the future of AI

โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“โ–“
-
security - not tested
F
license - not found
-
quality - not tested

Resources

Unclaimed servers have limited discoverability.

Looking for Admin?

If you are the server author, to access and configure the admin panel.

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/ProCreations-Official/second-opinion'

If you have feedback or need assistance with the MCP directory API, please join our Discord server