Ollama MCP Server

local-only server

The server can only run on the client’s local machine because it depends on local resources.

Integrations

  • Provides complete integration with Ollama, allowing users to pull, push, list, create, copy, and run local LLM models. Includes model management, execution of models with customizable prompts, and an OpenAI-compatible chat completion API.

  • Offers an OpenAI-compatible chat completion API that serves as a drop-in replacement, enabling the use of local Ollama models with the familiar OpenAI chat interface and message structure.

Ollama MCP Server

πŸš€ A powerful bridge between Ollama and the Model Context Protocol (MCP), enabling seamless integration of Ollama's local LLM capabilities into your MCP-powered applications.

🌟 Features

Complete Ollama Integration

  • Full API Coverage: Access all essential Ollama functionality through a clean MCP interface
  • OpenAI-Compatible Chat: Drop-in replacement for OpenAI's chat completion API
  • Local LLM Power: Run AI models locally with full control and privacy

Core Capabilities

  • πŸ”„ Model Management
    • Pull models from registries
    • Push models to registries
    • List available models
    • Create custom models from Modelfiles
    • Copy and remove models
  • πŸ€– Model Execution
    • Run models with customizable prompts
    • Chat completion API with system/user/assistant roles
    • Configurable parameters (temperature, timeout)
    • Raw mode support for direct responses
  • πŸ›  Server Control
    • Start and manage Ollama server
    • View detailed model information
    • Error handling and timeout management

πŸš€ Getting Started

Prerequisites

  • Ollama installed on your system
  • Node.js and npm/pnpm

Installation

  1. Install dependencies:
pnpm install
  1. Build the server:
pnpm run build

Configuration

Add the server to your MCP configuration:

For Claude Desktop:

MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json Windows: %APPDATA%/Claude/claude_desktop_config.json

{ "mcpServers": { "ollama": { "command": "node", "args": ["/path/to/ollama-server/build/index.js"], "env": { "OLLAMA_HOST": "http://127.0.0.1:11434" // Optional: customize Ollama API endpoint } } } }

πŸ›  Usage Examples

Pull and Run a Model

// Pull a model await mcp.use_mcp_tool({ server_name: "ollama", tool_name: "pull", arguments: { name: "llama2" } }); // Run the model await mcp.use_mcp_tool({ server_name: "ollama", tool_name: "run", arguments: { name: "llama2", prompt: "Explain quantum computing in simple terms" } });

Chat Completion (OpenAI-compatible)

await mcp.use_mcp_tool({ server_name: "ollama", tool_name: "chat_completion", arguments: { model: "llama2", messages: [ { role: "system", content: "You are a helpful assistant." }, { role: "user", content: "What is the meaning of life?" } ], temperature: 0.7 } });

Create Custom Model

await mcp.use_mcp_tool({ server_name: "ollama", tool_name: "create", arguments: { name: "custom-model", modelfile: "./path/to/Modelfile" } });

πŸ”§ Advanced Configuration

  • OLLAMA_HOST: Configure custom Ollama API endpoint (default: http://127.0.0.1:11434)
  • Timeout settings for model execution (default: 60 seconds)
  • Temperature control for response randomness (0-2 range)

🀝 Contributing

Contributions are welcome! Feel free to:

  • Report bugs
  • Suggest new features
  • Submit pull requests

πŸ“ License

MIT License - feel free to use in your own projects!


Built with ❀️ for the MCP ecosystem

You must be authenticated.

A
security – no known vulnerabilities
F
license - not found
A
quality - confirmed to work

A bridge that enables seamless integration of Ollama's local LLM capabilities into MCP-powered applications, allowing users to manage and run AI models locally with full API coverage.

  1. 🌟 Features
    1. Complete Ollama Integration
      1. Core Capabilities
      2. πŸš€ Getting Started
        1. Prerequisites
          1. Installation
            1. Configuration
              1. For Claude Desktop:
            2. πŸ›  Usage Examples
              1. Pull and Run a Model
                1. Chat Completion (OpenAI-compatible)
                  1. Create Custom Model
                  2. πŸ”§ Advanced Configuration
                    1. 🀝 Contributing
                      1. πŸ“ License