Allows integration with Ollama, enabling use of Ollama models through the MCP interface. Provides capabilities to list models, get model details, and ask questions to Ollama models.
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@MCP Ollama Serverlist all my available models"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
MCP Ollama
A Model Context Protocol (MCP) server for integrating Ollama with Claude Desktop or other MCP clients.
Requirements
Python 3.10 or higher
Ollama installed and running (https://ollama.com/download)
At least one model pulled with Ollama (e.g.,
ollama pull llama2)
Configure Claude Desktop
Add to your Claude Desktop configuration (~/Library/Application Support/Claude/claude_desktop_config.json on macOS, %APPDATA%\Claude\claude_desktop_config.json on Windows):
Development
Install in development mode:
Test with MCP Inspector:
Related MCP server: Ollama MCP Server
Features
The server provides four main tools:
list_models- List all downloaded Ollama modelsshow_model- Get detailed information about a specific modelask_model- Ask a question to a specified model
License
MIT