Why this server?
This server allows integration with various OpenAI SDK Compatible Chat Completion APIs, which could include methods for connecting to LM Studio, providing a path for Claude to use it.
AlicenseCqualityDmaintenanceIntegrate Claude with Any OpenAI SDK Compatible Chat Completion API - OpenAI, Perplexity, Groq, xAI, PyroPrompts and more.Last updated1121151MITWhy this server?
This server acts as a bridge to Ollama, enabling users to manage and run AI models locally, which could involve connecting Claude to a local LM studio instance configured with Ollama.
AlicenseBqualityDmaintenanceA bridge that enables seamless integration of Ollama's local LLM capabilities into MCP-powered applications, allowing users to manage and run AI models locally with full API coverage.Last updated101,00974AGPL 3.0Why this server?
This server directly integrates Ollama models and allows you to interact with them, enabling you to use local models with Claude Desktop.
AlicenseBqualityCmaintenanceMCP Ollama server integrates Ollama models with MCP clients, allowing users to list models, get detailed information, and interact with them through questions.Last updated332MITWhy this server?
While it doesn't directly mention LM Studio, this server enables bypassing plan limitations and potentially allowing custom integrations with local LLMs.
FlicenseBqualityDmaintenanceA server that enables Claude Desktop users to access the Claude API directly, allowing them to bypass Professional Plan limitations and use advanced features like custom system prompts and conversation management.Last updated134Why this server?
This server is a lightweight bridge that wraps OpenAI's built-in tools as MCP servers, which could be adapted to expose local LLMs as tools usable by Claude.
AlicenseBqualityFmaintenanceA lightweight bridge that wraps OpenAI's built-in tools (like web search and code interpreter) as Model Context Protocol servers, enabling their use with Claude and other MCP-compatible models.Last updated412MITWhy this server?
Enables text generation using the Qwen Max language model with seamless integration with Claude Desktop via the Model Context Protocol (MCP). This may be an option to use a different local LLM.
AlicenseBqualityDmaintenanceEnables text generation using the Qwen Max language model with configurable parameters and seamless integration with Claude Desktop via the Model Context Protocol (MCP).Last updated124MIT