Ollama MCP Server
local-only server
The server can only run on the clientβs local machine because it depends on local resources.
Integrations
Provides complete integration with Ollama, allowing users to pull, push, list, create, copy, and run local LLM models. Includes model management, execution of models with customizable prompts, and an OpenAI-compatible chat completion API.
Offers an OpenAI-compatible chat completion API that serves as a drop-in replacement, enabling the use of local Ollama models with the familiar OpenAI chat interface and message structure.
Ollama MCP Server
π A powerful bridge between Ollama and the Model Context Protocol (MCP), enabling seamless integration of Ollama's local LLM capabilities into your MCP-powered applications.
π Features
Complete Ollama Integration
- Full API Coverage: Access all essential Ollama functionality through a clean MCP interface
- OpenAI-Compatible Chat: Drop-in replacement for OpenAI's chat completion API
- Local LLM Power: Run AI models locally with full control and privacy
Core Capabilities
- π Model Management
- Pull models from registries
- Push models to registries
- List available models
- Create custom models from Modelfiles
- Copy and remove models
- π€ Model Execution
- Run models with customizable prompts
- Chat completion API with system/user/assistant roles
- Configurable parameters (temperature, timeout)
- Raw mode support for direct responses
- π Server Control
- Start and manage Ollama server
- View detailed model information
- Error handling and timeout management
π Getting Started
Prerequisites
- Ollama installed on your system
- Node.js and npm/pnpm
Installation
- Install dependencies:
- Build the server:
Configuration
Add the server to your MCP configuration:
For Claude Desktop:
MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
Windows: %APPDATA%/Claude/claude_desktop_config.json
π Usage Examples
Pull and Run a Model
Chat Completion (OpenAI-compatible)
Create Custom Model
π§ Advanced Configuration
OLLAMA_HOST
: Configure custom Ollama API endpoint (default: http://127.0.0.1:11434)- Timeout settings for model execution (default: 60 seconds)
- Temperature control for response randomness (0-2 range)
π€ Contributing
Contributions are welcome! Feel free to:
- Report bugs
- Suggest new features
- Submit pull requests
π License
MIT License - feel free to use in your own projects!
Built with β€οΈ for the MCP ecosystem
You must be authenticated.
A bridge that enables seamless integration of Ollama's local LLM capabilities into MCP-powered applications, allowing users to manage and run AI models locally with full API coverage.
- π Features
- π Getting Started
- π Usage Examples
- π§ Advanced Configuration
- π€ Contributing
- π License