Skip to main content
Glama

Ollama MCP Server

by etnlbck
README.md2.29 kB
# Ollama MCP Server An MCP (Model Context Protocol) server that provides tools to interact with Ollama models running on your local machine. ## Features - **List Models**: Get all available Ollama models - **Chat**: Interactive chat with conversation history - **Generate**: Single prompt generation - **Pull Models**: Download new models from Ollama registry - **Delete Models**: Remove models from local installation ## Prerequisites - [Ollama](https://ollama.ai/) installed and running locally - Node.js 18+ and npm ## Installation 1. Clone or download this repository 2. Install dependencies: ```bash npm install ``` 1. Build the project: ```bash npm run build ``` ## Usage ### Development Mode ```bash npm run dev ``` ### Production Mode ```bash npm run build npm start ``` ### Using with Claude Desktop Add this server to your Claude Desktop configuration: ```json { "mcpServers": { "ollama": { "command": "node", "args": ["/path/to/ollama-mcp/dist/index.js"] } } } ``` ### Using with Cursor If you're using [Cursor](https://cursor.com/), add the server to your MCP configuration file at `~/.cursor/mcp/config.json`: ```json { "mcpServers": { "ollama": { "command": "node", "args": ["/path/to/ollama-mcp/dist/index.js"] } } } ``` Alternatively, you can copy the ready-made config shipped with this repo: ```bash mkdir -p ~/.cursor/mcp cp /path/to/ollama-mcp/mcp.config.json ~/.cursor/mcp/config.json ``` ## Available Tools ### `ollama_list_models` Lists all available Ollama models on your system. ### `ollama_chat` Chat with a model using conversation history. - `model`: Name of the Ollama model - `messages`: Array of message objects with `role` ('system', 'user', 'assistant') and `content` ### `ollama_generate` Generate a response from a single prompt. - `model`: Name of the Ollama model - `prompt`: The input prompt ### `ollama_pull_model` Download a model from the Ollama registry. - `model`: Name of the model to download ### `ollama_delete_model` Remove a model from your local installation. - `model`: Name of the model to delete ## Configuration Set the `OLLAMA_BASE_URL` environment variable to change the Ollama server URL (default: `http://localhost:11434`). ## License MIT

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/etnlbck/ollama-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server