MCP Servers for Ollama

Ollama is an open-source project that allows you to run large language models (LLMs) locally on your own hardware, providing a way to use AI capabilities privately without sending data to external services.

View all MCP Servers

  • Why this server?

    Enables running, managing, and interacting with Ollama models including execution, information retrieval, downloading, listing, deletion, and chat completion

    A
    security
    A
    license
    A
    quality
    A Model Context Protocol (MCP) server that connects GraphDB's SPARQL endpoints and Ollama models to Claude, enabling Claude to query and manipulate ontology data while leveraging various AI models.
    Last updated -
    28
    TypeScript
    MIT License
    • Apple
    • Linux
  • Why this server?

    Provides integration with Ollama for local LLM support and embedding generation

    A
    security
    A
    license
    A
    quality
    A flexible memory system for AI applications that supports multiple LLM providers and can be used either as an MCP server or as a direct library integration, enabling autonomous memory management without explicit commands.
    Last updated -
    3
    131
    36
    JavaScript
    MIT License
  • Why this server?

    Integration with Ollama for local language model inference to power browser automation

    A
    security
    A
    license
    A
    quality
    Facilitates browser automation with custom capabilities and agent-based interactions, integrated through the browser-use library.
    Last updated -
    1
    367
    Python
    MIT License
    • Apple
  • Why this server?

    Supports local models through Ollama, allowing integration with locally hosted LLMs alongside cloud-based options.

    A
    security
    A
    license
    A
    quality
    An AI router that connects applications to multiple LLM providers (OpenAI, Anthropic, Google, DeepSeek, Ollama, etc.) with smart model orchestration capabilities, enabling dynamic switching between models for different reasoning tasks.
    Last updated -
    3
    1
    TypeScript
    MIT License
    • Linux
    • Apple
  • Why this server?

    Connects to locally running Ollama models to query multiple AI perspectives on a single question, with support for customizing system prompts and assigning different roles to each model.

    A
    security
    A
    license
    A
    quality
    An MCP server that queries multiple Ollama models and combines their responses, providing diverse AI perspectives on a single question for more comprehensive answers.
    Last updated -
    2
    17
    TypeScript
    MIT License
    • Apple
  • Why this server?

    Allows access to LLMs hosted through Ollama via the LLM_MODEL_PROVIDER environment variable

    A
    security
    A
    license
    A
    quality
    An MCP server that provides LLMs access to other LLMs
    Last updated -
    4
    14
    12
    JavaScript
    MIT License
  • Why this server?

    Enables research capabilities using any local LLM hosted by Ollama, supporting models like deepseek-r1 and llama3.2

    A
    security
    A
    license
    A
    quality
    This is a Model Context Protocol (MCP) server adaptation of LangChain Ollama Deep Researcher. It provides the deep research capabilities as MCP tools that can be used within the model context protocol ecosystem, allowing AI assistants to perform in-depth research on topics (locally) via Ollama
    Last updated -
    3
    5
    Python
    MIT License
    • Apple
    • Linux
  • Why this server?

    Allows integration with Ollama, enabling use of Ollama models through the MCP interface. Provides capabilities to list models, get model details, and ask questions to Ollama models.

    A
    security
    A
    license
    A
    quality
    MCP Ollama server integrates Ollama models with MCP clients, allowing users to list models, get detailed information, and interact with them through questions.
    Last updated -
    3
    12
    Python
    MIT License
    • Apple
  • Why this server?

    Provides integration with Ollama's LLM server, allowing interactive chat with Ollama models while using the Bybit tools to access cryptocurrency data.

    A
    security
    A
    license
    A
    quality
    A Model Context Protocol server that provides read-only access to Bybit's cryptocurrency exchange API, allowing users to query real-time cryptocurrency data using natural language.
    Last updated -
    9
    5
    TypeScript
    MIT License
  • Why this server?

    Provides integration with Ollama to run open source models locally, supporting models like Phi-3.

    -
    security
    A
    license
    -
    quality
    GenAIScript is a JavaScript runtime dedicated to build relaible, automatable LLM scripts. Every GenAIScript can be exposed as a MCP server automatically.
    Last updated -
    2,574
    TypeScript
    MIT License
    • Linux
    • Apple
  • Why this server?

    Supports integration with Ollama for local execution of Large Language Models, providing an alternative to cloud-based AI providers.

    A
    security
    F
    license
    A
    quality
    Enables AI agents to interact with web browsers using natural language, featuring automated browsing, form filling, vision-based element detection, and structured JSON responses for systematic browser control.
    Last updated -
    1
    38
    Python
    • Linux
    • Apple
  • Why this server?

    Supports Ollama as an LLM provider through API key integration

    A
    security
    F
    license
    A
    quality
    Enables browser automation using Python scripts, offering operations like taking webpage screenshots, retrieving HTML content, and executing JavaScript.
    Last updated -
    4
    18
    Python
    • Linux
  • Why this server?

    Uses Ollama models to generate text embeddings from Parquet file columns

    A
    security
    F
    license
    A
    quality
    A powerful MCP (Model Control Protocol) server that provides tools for manipulating and analyzing Parquet files. This server is designed to work with Claude Desktop and offers four main functionalities:
    Last updated -
    2
    2
    Python
    • Linux
    • Apple
  • Why this server?

    Allows interaction with locally-hosted Ollama models through a consistent API, supporting models like Llama 3.1.

    A
    security
    F
    license
    A
    quality
    A lightweight MCP server that provides a unified interface to various LLM providers including OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, and Ollama.
    Last updated -
    6
    84
    Python
  • Why this server?

    Provides access to Deepseek reasoning content through a local Ollama server

    A
    security
    F
    license
    A
    quality
    Provides reasoning content to MCP-enabled AI clients by interfacing with Deepseek's API or a local Ollama server, enabling focused reasoning and thought process visualization.
    Last updated -
    1
    54
    24
    JavaScript
  • Why this server?

    Provides complete integration with Ollama, allowing users to pull, push, list, create, copy, and run local LLM models. Includes model management, execution of models with customizable prompts, and an OpenAI-compatible chat completion API.

    A
    security
    F
    license
    A
    quality
    A bridge that enables seamless integration of Ollama's local LLM capabilities into MCP-powered applications, allowing users to manage and run AI models locally with full API coverage.
    Last updated -
    10
    33
    JavaScript
    • Apple
  • Why this server?

    Integration with Ollama's open-source AI models to create blockchain agents for Starknet operations.

    -
    security
    A
    license
    -
    quality
    Build powerful and secure AI agents powered by Starknet.
    Last updated -
    13
    87
    TypeScript
    MIT License
  • Why this server?

    Planned future integration to enable running embedding models locally for complete privacy and control

    -
    security
    A
    license
    -
    quality
    Web crawling and RAG implementation that enables AI agents to scrape websites and perform semantic search over the crawled content, storing everything in Supabase for persistent knowledge retrieval.
    Last updated -
    115
    Python
    MIT License
    • Linux
    • Apple
  • Why this server?

    Integrates with Ollama LLMs to enable natural language hardware control through OllamaHardwareAgent

    -
    security
    A
    license
    -
    quality
    A Python framework that enables secure hardware control through the Model Context Protocol, allowing AI agents and automation systems to interact with physical devices across multiple platforms.
    Last updated -
    Python
    Apache 2.0
    • Linux
    • Apple
  • Why this server?

    Provides integration with free/open-weights models through Ollama, enabling code analysis and editing without proprietary LLM APIs

    -
    security
    A
    license
    -
    quality
    A fully featured coding agent that uses symbolic operations (enabled by language servers) and works well even in large code bases. Essentially a free to use alternative to Cursor and Windsurf Agents, Cline, Roo Code and others.
    Last updated -
    764
    Python
    MIT License
    • Apple
  • Why this server?

    Enables integration with Ollama's local models to run MCTS analysis, allowing model selection, comparison between different Ollama models, and storing results organized by model name.

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol (MCP) server that enables Claude to use Monte Carlo Tree Search algorithms for deep, explorative analysis of topics, questions, or text inputs.
    Last updated -
    2
    Python
    MIT License
    • Apple
    • Linux
  • Why this server?

    Provides support for local LLM capabilities, allowing the server to use locally hosted models for embeddings and text generation rather than cloud-based solutions.

    -
    security
    A
    license
    -
    quality
    Provides a standardized protocol for tool invocation, enabling an AI system to search the web, retrieve information, and provide relevant answers through integration with LangChain, RAG, and Ollama.
    Last updated -
    Python
    MIT License
  • Why this server?

    Integrates with Ollama to allow running local LLMs as an alternative to cloud-based models for processing API interactions.

    -
    security
    A
    license
    -
    quality
    swagger-mcp is a tool designed to scrape Swagger UI by extracting the swagger.json file and dynamically generating well-defined mcp tools at runtime. These tools can be utilized by the MCP client for further tool selection.
    Last updated -
    25
    Go
    MIT License
  • Why this server?

    Leverages Ollama's local AI models (nomic-embed-text, phi4, clip) for document processing, metadata extraction, and vector embeddings of construction documents.

    -
    security
    A
    license
    -
    quality
    AI-powered MCP server that enables Claude and other LLMs to interact directly with construction documents, drawings, and specifications through advanced RAG and hybrid search capabilities.
    Last updated -
    1
    TypeScript
    MIT License
  • Why this server?

    Integrates with Ollama AI models for enhanced code analysis capabilities

    -
    security
    A
    license
    -
    quality
    Converts code into UML diagrams and flowcharts through static analysis, enabling visualization of code structure and explanation of functionality.
    Last updated -
    17
    JavaScript
    MIT License
    • Linux
    • Apple
  • Why this server?

    Enables using local Ollama models for security validation in the MCP gateway through the OLLAMA_HOST configuration

    -
    security
    A
    license
    -
    quality
    A prompt-validation service that checks incoming prompts against security rules, blocking those matching high-severity security or jailbreak patterns while authorizing safe prompts.
    Last updated -
    Python
    MIT License
    • Linux
    • Apple
  • Why this server?

    Provides access to Ollama's local LLM models through a Model Context Protocol server, allowing listing, pulling, and chatting with Ollama models

    -
    security
    A
    license
    -
    quality
    Enables seamless integration between Ollama's local LLM models and MCP-compatible applications, supporting model management and chat interactions.
    Last updated -
    165
    47
    TypeScript
    AGPL 3.0
  • Why this server?

    Offers alternative LLM provider integration for task management functions, allowing use of locally deployed Ollama models for PRD parsing and task suggestions

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server providing comprehensive task management capabilities with support for project organization, task tracking, and automatic PRD parsing into actionable items.
    Last updated -
    4
    Python
    MIT License
    • Linux
    • Apple
  • Why this server?

    Allows using Ollama's local language models as an alternative provider for generating embeddings and handling memory operations

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server that provides AI agents with persistent memory capabilities through Mem0, allowing them to store, retrieve, and semantically search memories.
    Last updated -
    2
    Python
    MIT License
  • Why this server?

    Provides local embeddings generation using Ollama's nomic-embed-text model as an alternative to cloud-based embedding services.

    -
    security
    A
    license
    -
    quality
    An MCP server implementation that provides tools for retrieving and processing documentation through vector search, enabling AI assistants to augment their responses with relevant documentation context. Uses Ollama or OpenAI to generate embeddings. Docker files included
    Last updated -
    19
    14
    TypeScript
    MIT License
    • Apple
    • Linux
  • Why this server?

    Integrates with Ollama for local embedding models, supporting document embedding and semantic search functionality.

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server that enables semantic search capabilities by providing tools to manage Qdrant vector database collections, process and embed documents using various embedding services, and perform semantic searches across vector embeddings.
    Last updated -
    89
    TypeScript
    MIT License
  • Why this server?

    Uses Ollama with nomic-embed-text to generate vector embeddings for documents, enabling semantic search capabilities in Solr collections.

    -
    security
    A
    license
    -
    quality
    A Python server that enables AI assistants to perform hybrid search queries against Apache Solr indexes through the Model Context Protocol, combining keyword precision with vector-based semantic understanding.
    Last updated -
    1
    Python
    MIT License
    • Linux
    • Apple
  • Why this server?

    Enables communication between Unity and local Large Language Models (LLMs) running through Ollama, allowing developers to automate Unity workflows, manipulate assets, and control the Unity Editor programmatically without cloud-based LLMs.

    -
    security
    A
    license
    -
    quality
    A server that connects Unity with local large language models through Ollama, enabling developers to automate workflows, manipulate assets, and control the Unity Editor programmatically without relying on cloud-based LLMs.
    Last updated -
    Python
    MIT License
    • Apple
    • Linux
  • Why this server?

    Mentioned as a planned feature for local embeddings generation as an alternative to OpenAI embeddings

    -
    security
    A
    license
    -
    quality
    Provides tools for retrieving and processing documentation through vector search, enabling AI assistants to augment their responses with relevant documentation context.
    Last updated -
    62
    TypeScript
    MIT License
    • Apple
  • Why this server?

    Supports integration with Ollama through MCPHost as a free alternative to Claude, enabling LLMs to interact with the MCP server

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server that enables AI agents to query Erick Wendel's talks, blog posts, and videos across different platforms using natural language.
    Last updated -
    55
    TypeScript
    MIT License
  • Why this server?

    Provides a standardized interface for interacting with Ollama's API, supporting model listing, chat functionality, text generation, embedding generation, and querying running models and model details.

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server that provides standardized interfaces for interacting with Ollama API, offering JSON responses, error handling, and intelligent guidance for LLM-based API calls.
    Last updated -
    Python
    MIT License
    • Linux
    • Apple
  • Why this server?

    Provides integration with Ollama's local LLM hosting service, supporting customizable context settings and model parameters for browser automation

    -
    security
    A
    license
    -
    quality
    AI-driven browser automation server that implements the Model Context Protocol to enable natural language control of web browsers for tasks like navigation, form filling, and visual interaction.
    Last updated -
    1
    Python
    MIT License
    • Apple
  • Why this server?

    Optionally connects to an Ollama server for prompt generation using LLMs hosted on Ollama

    -
    security
    A
    license
    -
    quality
    The Comfy MCP Server uses the FastMCP framework to generate images from prompts by interacting with a remote Comfy server, allowing automated image creation based on workflow configurations.
    Last updated -
    7
    Python
    MIT License
  • Why this server?

    Uses Ollama's embedding models (particularly nomic-embed-text) for creating vector embeddings for documentation search

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol (MCP) server that enables semantic search and retrieval of documentation using a vector database (Qdrant). This server allows you to add documentation from URLs or local files and then search through them using natural language queries.
    Last updated -
    14
    74
    JavaScript
    Apache 2.0
    • Apple
  • Why this server?

    Used for the default summarization and embedding models required by the server, specifically the snowflake-arctic-embed2 and llama3.1:8b models.

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol (MCP) server that enables LLMs to interact directly the documents that they have on-disk through agentic RAG and hybrid search in LanceDB. Ask LLMs questions about the dataset as a whole or about specific documents.
    Last updated -
    12
    31
    TypeScript
    MIT License
    • Apple
  • Why this server?

    Provides free embeddings for vector representation of documents

    -
    security
    A
    license
    -
    quality
    Provides RAG capabilities for semantic document search using Qdrant vector database and Ollama/OpenAI embeddings, allowing users to add, search, list, and delete documentation with metadata support.
    Last updated -
    5
    4
    TypeScript
    Apache 2.0
  • Why this server?

    Provides integration with Ollama using the Mistral model, allowing AI agents to interact with and leverage the model's capabilities through the MCP protocol

    -
    security
    F
    license
    -
    quality
    A TypeScript implementation of a Model Context Protocol server and client that enables interaction with language models (specifically Mistral running on Ollama).
    Last updated -
    TypeScript
  • Why this server?

    Utilizes Ollama server for embedding generation to support the similarity search functionality.

    -
    security
    F
    license
    -
    quality
    A Model Control Protocol server that provides web search capabilities and similarity search functionality for Claude Desktop, allowing users to perform web searches and extract relevant information from previous search results.
    Last updated -
    2
    Python
    • Apple
    • Linux
  • Why this server?

    Mentioned as a potential OpenAI API-compatible endpoint that can be used with the MCP server through the openai_compatible provider setting.

  • Why this server?

    Uses locally running Ollama models to process natural language commands, with the ability to switch between different models like llama3.2 or Gemma3, and query available models from the Ollama server.

    -
    security
    F
    license
    -
    quality
    A server that integrates Blender with local AI models via the Model Context Protocol, allowing users to control Blender using natural language prompts for 3D modeling tasks.
    Last updated -
    13
    Python
    • Linux
    • Apple
  • Why this server?

    Enables integration with local large language models like Mistral, allowing the MCP server to process conversational AI requests without relying on cloud services.

    -
    security
    F
    license
    -
    quality
    A conversational application server that integrates LLM capabilities via Ollama with vector memory context, supporting multiple users, sessions, automatic history summarization, and a plugin system for executing real actions.
    Last updated -
    Python
  • Why this server?

    Leverages Ollama's embedding model (nomic-embed-text) to create custom embedding functions for converting text into vector representations that can be searched.

    -
    security
    F
    license
    -
    quality
    A Node.js implementation for vector search using LanceDB and Ollama's embedding model.
    Last updated -
    JavaScript
  • Why this server?

    Enables running LLaMA 3.2 3B locally, allowing the MCP to integrate Yahoo Finance data with LLaMA's capabilities

    -
    security
    F
    license
    -
    quality
    Integrates Yahoo Finance API with LLaMA 3.2 3B, allowing users to retrieve real-time stock prices, company information, historical data, and market news while maintaining the model's original capabilities.
    Last updated -
    Python
  • Why this server?

    Provides services for generating embeddings and text with Ollama, allowing AI-powered applications to perform embedding generation and text generation operations locally.

    -
    security
    F
    license
    -
    quality
    A generic Model Context Protocol framework for building AI-powered applications that provides standardized ways to create MCP servers and clients for integrating LLMs with support for Ollama and Supabase.
    Last updated -
    TypeScript
  • Why this server?

    Provides integration with Ollama for AI-powered code reviews using local models, allowing the MCP server to utilize Ollama's capabilities to generate expert code reviews based on different programming principles.

    -
    security
    F
    license
    -
    quality
    A Python-based system that provides AI-powered code reviews through simulated expert personas like Martin Fowler and Robert C. Martin, using the Model Context Protocol (MCP).
    Last updated -
    Python
    • Apple
    • Linux
  • Why this server?

    Supports exporting fine-tuned models to Ollama format for local deployment and inference.

    -
    security
    F
    license
    -
    quality
    Provides tools for optimizing, fine-tuning, and deploying large language models with Unsloth, enabling 2x faster training with 80% less memory through model loading, fine-tuning, text generation, and model export capabilities.
    Last updated -
    1
    JavaScript
  • Why this server?

    Generates vector embeddings for emails using models like nomic-embed-text for enhanced semantic search capabilities

    -
    security
    F
    license
    -
    quality
    Processes emails from Outlook with date filtering, storing them in SQLite databases while generating vector embeddings for semantic search capabilities in MongoDB.
    Last updated -
    5
    Python
    • Apple
    • Linux
  • Why this server?

    Integrates with Ollama as a local LLM provider for context-aware querying. Allows users to send prompts to Ollama models with context from local files.

    -
    security
    F
    license
    -
    quality
    This server provides an API to query Large Language Models using context from local files, supporting various models and file types for context-aware responses.
    Last updated -
    1
    TypeScript
  • Why this server?

    Enables seamless communication with local Ollama LLM instances, providing capabilities for task decomposition, result evaluation, and direct model execution with configurable parameters.

    -
    security
    F
    license
    -
    quality
    A server that enables seamless integration between local Ollama LLM instances and MCP-compatible applications, providing advanced task decomposition, evaluation, and workflow management capabilities.
    Last updated -
    1
    Python
    • Apple
  • Why this server?

    Uses Ollama for efficient embedding generation, requiring it to be installed and running for vector operations

    -
    security
    F
    license
    -
    quality
    Provides a project memory bank and RAG context provider for enhanced code understanding and management through vector embeddings, integrated with RooCode and Cline.
    Last updated -
    9
    Python
    • Apple
  • Why this server?

    Provides integration with Ollama for local AI model usage and processing

    -
    security
    F
    license
    -
    quality
    Facilitates enhanced interaction with large language models (LLMs) by providing intelligent context management, tool integration, and multi-provider AI model coordination for efficient AI-driven workflows.
    Last updated -
    Python
  • Why this server?

    Uses Ollama as a Large Language Model provider to determine user intent and route requests

    -
    security
    F
    license
    -
    quality
    Facilitates executing system commands and retrieving web data using the Brave Search API by interpreting user intents via a Large Language Model (LLM).
    Last updated -
    1
    Python
  • Why this server?

    Uses Ollama as the default embedding provider for local embeddings generation, supporting semantic documentation search and vector storage.

    -
    security
    F
    license
    -
    quality
    Enables AI assistants to enhance their responses with relevant documentation through a semantic vector search, offering tools for managing and processing documentation efficiently.
    Last updated -
    62
    13
    TypeScript
  • Why this server?

    Integrates with Ollama to use the Deepseek model for AI capabilities through the MCP protocol

    -
    security
    F
    license
    -
    quality
    Enables seamless AI integration via Ollama's Deepseek model, providing protocol compliance and automatic configuration for clean AI-driven interactions.
    Last updated -
    1
    Python
  • Why this server?

    Allows querying Ollama models directly from Claude with performance tracking, supporting selection of different models and providing context for queries.

    -
    security
    F
    license
    -
    quality
    Facilitates initiating Ollama queries via Claude and manages a simple note storage system with capabilities to add, summarize, and access notes using custom URIs.
    Last updated -
    1
    2
    Python
    • Apple
  • Why this server?

    Leverages Ollama's LLM capabilities to interpret natural language questions, generate SQL queries, and provide AI-powered responses based on database results.

    -
    security
    F
    license
    -
    quality
    An interactive chat interface that combines Ollama's LLM capabilities with PostgreSQL database access through the Model Context Protocol (MCP). Ask questions about your data in natural language and get AI-powered responses backed by real SQL queries.
    Last updated -
    28
    TypeScript
  • Why this server?

    Allows communication with locally available Ollama models (like llama2, codellama) while maintaining persistent conversation history.

    -
    security
    F
    license
    -
    quality
    A TypeScript-based server that provides a memory system for Large Language Models (LLMs), allowing users to interact with multiple LLM providers while maintaining conversation history and offering tools for managing providers and model configurations.
    Last updated -
    20
    JavaScript
    • Apple