MCP Servers for Ollama

Ollama is an open-source project that allows you to run large language models (LLMs) locally on your own hardware, providing a way to use AI capabilities privately without sending data to external services.

View all MCP Servers

  • Why this server?

    Integration with Ollama for local language model inference to power browser automation

    A
    security
    A
    license
    A
    quality
    Facilitates browser automation with custom capabilities and agent-based interactions, integrated through the browser-use library.
    1
    367
    Python
    MIT License
    • Apple
  • Why this server?

    Supports local models through Ollama, allowing integration with locally hosted LLMs alongside cloud-based options.

    A
    security
    A
    license
    A
    quality
    An AI router that connects applications to multiple LLM providers (OpenAI, Anthropic, Google, DeepSeek, Ollama, etc.) with smart model orchestration capabilities, enabling dynamic switching between models for different reasoning tasks.
    3
    1
    TypeScript
    MIT License
    • Linux
    • Apple
  • Why this server?

    Provides integration with Ollama for local LLM support and embedding generation

    A
    security
    A
    license
    A
    quality
    A flexible memory system for AI applications that supports multiple LLM providers and can be used either as an MCP server or as a direct library integration, enabling autonomous memory management without explicit commands.
    3
    746
    32
    JavaScript
    MIT License
  • Why this server?

    Connects to locally running Ollama models to query multiple AI perspectives on a single question, with support for customizing system prompts and assigning different roles to each model.

    A
    security
    A
    license
    A
    quality
    An MCP server that queries multiple Ollama models and combines their responses, providing diverse AI perspectives on a single question for more comprehensive answers.
    2
    17
    TypeScript
    MIT License
    • Apple
  • Why this server?

    Allows access to LLMs hosted through Ollama via the LLM_MODEL_PROVIDER environment variable

    A
    security
    A
    license
    A
    quality
    An MCP server that provides LLMs access to other LLMs
    4
    14
    12
    JavaScript
    MIT License
  • Why this server?

    Enables research capabilities using any local LLM hosted by Ollama, supporting models like deepseek-r1 and llama3.2

    A
    security
    A
    license
    A
    quality
    This is a Model Context Protocol (MCP) server adaptation of LangChain Ollama Deep Researcher. It provides the deep research capabilities as MCP tools that can be used within the model context protocol ecosystem, allowing AI assistants to perform in-depth research on topics (locally) via Ollama
    3
    5
    Python
    MIT License
    • Apple
    • Linux
  • Why this server?

    Allows integration with Ollama, enabling use of Ollama models through the MCP interface. Provides capabilities to list models, get model details, and ask questions to Ollama models.

    A
    security
    A
    license
    A
    quality
    MCP Ollama server integrates Ollama models with MCP clients, allowing users to list models, get detailed information, and interact with them through questions.
    3
    12
    Python
    MIT License
    • Apple
  • Why this server?

    Provides integration with Ollama's LLM server, allowing interactive chat with Ollama models while using the Bybit tools to access cryptocurrency data.

    A
    security
    A
    license
    A
    quality
    A Model Context Protocol server that provides read-only access to Bybit's cryptocurrency exchange API, allowing users to query real-time cryptocurrency data using natural language.
    9
    5
    TypeScript
    MIT License
  • Why this server?

    Supports integration with Ollama for local execution of Large Language Models, providing an alternative to cloud-based AI providers.

    A
    security
    F
    license
    A
    quality
    Enables AI agents to interact with web browsers using natural language, featuring automated browsing, form filling, vision-based element detection, and structured JSON responses for systematic browser control.
    1
    23
    Python
    • Linux
    • Apple
  • Why this server?

    Provides access to Deepseek reasoning content through a local Ollama server

    A
    security
    F
    license
    A
    quality
    Provides reasoning content to MCP-enabled AI clients by interfacing with Deepseek's API or a local Ollama server, enabling focused reasoning and thought process visualization.
    1
    54
    24
    JavaScript
  • Why this server?

    Supports Ollama as an LLM provider through API key integration

    A
    security
    F
    license
    A
    quality
    Enables browser automation using Python scripts, offering operations like taking webpage screenshots, retrieving HTML content, and executing JavaScript.
    4
    15
    Python
    • Linux
  • Why this server?

    Provides complete integration with Ollama, allowing users to pull, push, list, create, copy, and run local LLM models. Includes model management, execution of models with customizable prompts, and an OpenAI-compatible chat completion API.

    A
    security
    F
    license
    A
    quality
    A bridge that enables seamless integration of Ollama's local LLM capabilities into MCP-powered applications, allowing users to manage and run AI models locally with full API coverage.
    10
    33
    JavaScript
    • Apple
  • Why this server?

    Integration with Ollama's open-source AI models to create blockchain agents for Starknet operations.

    -
    security
    A
    license
    -
    quality
    Build powerful and secure AI agents powered by Starknet.
    84
    84
    TypeScript
    MIT License
  • Why this server?

    Offers alternative LLM provider integration for task management functions, allowing use of locally deployed Ollama models for PRD parsing and task suggestions

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server providing comprehensive task management capabilities with support for project organization, task tracking, and automatic PRD parsing into actionable items.
    2
    Python
    MIT License
  • Why this server?

    Provides integration with free/open-weights models through Ollama, enabling code analysis and editing without proprietary LLM APIs

    oraios
    Verified
    -
    security
    A
    license
    -
    quality
    A fully featured coding agent that uses symbolic operations (enabled by language servers) and works well even in large code bases. Essentially a free to use alternative to Cursor and Windsurf Agents, Cline, Roo Code and others.
    542
    Python
    GPL 2.0
    • Apple
  • Why this server?

    Allows using Ollama's local language models as an alternative provider for generating embeddings and handling memory operations

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server that provides AI agents with persistent memory capabilities through Mem0, allowing them to store, retrieve, and semantically search memories.
    2
    Python
    MIT License
  • Why this server?

    Provides local embeddings generation using Ollama's nomic-embed-text model as an alternative to cloud-based embedding services.

    -
    security
    A
    license
    -
    quality
    An MCP server implementation that provides tools for retrieving and processing documentation through vector search, enabling AI assistants to augment their responses with relevant documentation context. Uses Ollama or OpenAI to generate embeddings. Docker files included
    19
    14
    TypeScript
    MIT License
    • Apple
    • Linux
  • Why this server?

    Integrates with Ollama AI models for enhanced code analysis capabilities

    -
    security
    A
    license
    -
    quality
    Converts code into UML diagrams and flowcharts through static analysis, enabling visualization of code structure and explanation of functionality.
    5
    JavaScript
    MIT License
    • Linux
    • Apple
  • Why this server?

    Integrates with Ollama for local embedding models, supporting document embedding and semantic search functionality.

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server that enables semantic search capabilities by providing tools to manage Qdrant vector database collections, process and embed documents using various embedding services, and perform semantic searches across vector embeddings.
    89
    TypeScript
    MIT License
  • Why this server?

    Uses Ollama with nomic-embed-text to generate vector embeddings for documents, enabling semantic search capabilities in Solr collections.

    -
    security
    A
    license
    -
    quality
    A Python server that enables AI assistants to perform hybrid search queries against Apache Solr indexes through the Model Context Protocol, combining keyword precision with vector-based semantic understanding.
    1
    Python
    MIT License
    • Linux
    • Apple
  • Why this server?

    Enables communication between Unity and local Large Language Models (LLMs) running through Ollama, allowing developers to automate Unity workflows, manipulate assets, and control the Unity Editor programmatically without cloud-based LLMs.

    -
    security
    A
    license
    -
    quality
    A server that connects Unity with local large language models through Ollama, enabling developers to automate workflows, manipulate assets, and control the Unity Editor programmatically without relying on cloud-based LLMs.
    Python
    MIT License
    • Apple
    • Linux
  • Why this server?

    Mentioned as a planned feature for local embeddings generation as an alternative to OpenAI embeddings

    -
    security
    A
    license
    -
    quality
    Provides tools for retrieving and processing documentation through vector search, enabling AI assistants to augment their responses with relevant documentation context.
    62
    TypeScript
    MIT License
    • Apple
  • Why this server?

    Supports integration with Ollama through MCPHost as a free alternative to Claude, enabling LLMs to interact with the MCP server

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server that enables AI agents to query Erick Wendel's talks, blog posts, and videos across different platforms using natural language.
    55
    TypeScript
    MIT License
  • Why this server?

    Leverages Ollama's local AI models (nomic-embed-text, phi4, clip) for document processing, metadata extraction, and vector embeddings of construction documents.

    -
    security
    A
    license
    -
    quality
    AI-powered MCP server that enables Claude and other LLMs to interact directly with construction documents, drawings, and specifications through advanced RAG and hybrid search capabilities.
    1
    TypeScript
    MIT License
  • Why this server?

    Provides a standardized interface for interacting with Ollama's API, supporting model listing, chat functionality, text generation, embedding generation, and querying running models and model details.

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol server that provides standardized interfaces for interacting with Ollama API, offering JSON responses, error handling, and intelligent guidance for LLM-based API calls.
    Python
    MIT License
    • Linux
    • Apple
  • Why this server?

    Provides integration with Ollama's local LLM hosting service, supporting customizable context settings and model parameters for browser automation

    -
    security
    A
    license
    -
    quality
    AI-driven browser automation server that implements the Model Context Protocol to enable natural language control of web browsers for tasks like navigation, form filling, and visual interaction.
    1
    Python
    MIT License
    • Apple
  • Why this server?

    Optionally connects to an Ollama server for prompt generation using LLMs hosted on Ollama

    -
    security
    A
    license
    -
    quality
    The Comfy MCP Server uses the FastMCP framework to generate images from prompts by interacting with a remote Comfy server, allowing automated image creation based on workflow configurations.
    7
    Python
    MIT License
  • Why this server?

    Uses Ollama's embedding models (particularly nomic-embed-text) for creating vector embeddings for documentation search

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol (MCP) server that enables semantic search and retrieval of documentation using a vector database (Qdrant). This server allows you to add documentation from URLs or local files and then search through them using natural language queries.
    14
    74
    JavaScript
    Apache 2.0
    • Apple
  • Why this server?

    Provides access to Ollama's local LLM models through a Model Context Protocol server, allowing listing, pulling, and chatting with Ollama models

    -
    security
    A
    license
    -
    quality
    Enables seamless integration between Ollama's local LLM models and MCP-compatible applications, supporting model management and chat interactions.
    50
    13
    TypeScript
    MIT License
  • Why this server?

    Used for the default summarization and embedding models required by the server, specifically the snowflake-arctic-embed2 and llama3.1:8b models.

    -
    security
    A
    license
    -
    quality
    A Model Context Protocol (MCP) server that enables LLMs to interact directly the documents that they have on-disk through agentic RAG and hybrid search in LanceDB. Ask LLMs questions about the dataset as a whole or about specific documents.
    12
    31
    TypeScript
    MIT License
    • Apple
  • Why this server?

    Provides free embeddings for vector representation of documents

    -
    security
    A
    license
    -
    quality
    Provides RAG capabilities for semantic document search using Qdrant vector database and Ollama/OpenAI embeddings, allowing users to add, search, list, and delete documentation with metadata support.
    5
    4
    TypeScript
    Apache 2.0
  • Why this server?

    Provides integration with Ollama using the Mistral model, allowing AI agents to interact with and leverage the model's capabilities through the MCP protocol

    -
    security
    F
    license
    -
    quality
    A TypeScript implementation of a Model Context Protocol server and client that enables interaction with language models (specifically Mistral running on Ollama).
    TypeScript
  • Why this server?

    Uses locally running Ollama models to process natural language commands, with the ability to switch between different models like llama3.2 or Gemma3, and query available models from the Ollama server.

    -
    security
    F
    license
    -
    quality
    A server that integrates Blender with local AI models via the Model Context Protocol, allowing users to control Blender using natural language prompts for 3D modeling tasks.
    13
    Python
    • Linux
    • Apple
  • Why this server?

    Enables integration with local large language models like Mistral, allowing the MCP server to process conversational AI requests without relying on cloud services.

    -
    security
    F
    license
    -
    quality
    A conversational application server that integrates LLM capabilities via Ollama with vector memory context, supporting multiple users, sessions, automatic history summarization, and a plugin system for executing real actions.
    Python
  • Why this server?

    Allows interaction with locally-hosted Ollama models through a consistent API, supporting models like Llama 3.1.

    -
    security
    F
    license
    -
    quality
    A lightweight MCP server that provides a unified interface to various LLM providers including OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, and Ollama.
    6
    Python
  • Why this server?

    Leverages Ollama's embedding model (nomic-embed-text) to create custom embedding functions for converting text into vector representations that can be searched.

    -
    security
    F
    license
    -
    quality
    A Node.js implementation for vector search using LanceDB and Ollama's embedding model.
    JavaScript
  • Why this server?

    Enables running LLaMA 3.2 3B locally, allowing the MCP to integrate Yahoo Finance data with LLaMA's capabilities

    -
    security
    F
    license
    -
    quality
    Integrates Yahoo Finance API with LLaMA 3.2 3B, allowing users to retrieve real-time stock prices, company information, historical data, and market news while maintaining the model's original capabilities.
    Python
  • Why this server?

    Provides services for generating embeddings and text with Ollama, allowing AI-powered applications to perform embedding generation and text generation operations locally.

    -
    security
    F
    license
    -
    quality
    A generic Model Context Protocol framework for building AI-powered applications that provides standardized ways to create MCP servers and clients for integrating LLMs with support for Ollama and Supabase.
    TypeScript
  • Why this server?

    Provides integration with Ollama for AI-powered code reviews using local models, allowing the MCP server to utilize Ollama's capabilities to generate expert code reviews based on different programming principles.

    -
    security
    F
    license
    -
    quality
    A Python-based system that provides AI-powered code reviews through simulated expert personas like Martin Fowler and Robert C. Martin, using the Model Context Protocol (MCP).
    Python
    • Apple
    • Linux
  • Why this server?

    Supports exporting fine-tuned models to Ollama format for local deployment and inference.

    -
    security
    F
    license
    -
    quality
    Provides tools for optimizing, fine-tuning, and deploying large language models with Unsloth, enabling 2x faster training with 80% less memory through model loading, fine-tuning, text generation, and model export capabilities.
    1
    JavaScript
  • Why this server?

    Generates vector embeddings for emails using models like nomic-embed-text for enhanced semantic search capabilities

    -
    security
    F
    license
    -
    quality
    Processes emails from Outlook with date filtering, storing them in SQLite databases while generating vector embeddings for semantic search capabilities in MongoDB.
    5
    Python
    • Apple
    • Linux
  • Why this server?

    Integrates with Ollama as a local LLM provider for context-aware querying. Allows users to send prompts to Ollama models with context from local files.

    -
    security
    F
    license
    -
    quality
    This server provides an API to query Large Language Models using context from local files, supporting various models and file types for context-aware responses.
    1
    TypeScript
  • Why this server?

    Enables seamless communication with local Ollama LLM instances, providing capabilities for task decomposition, result evaluation, and direct model execution with configurable parameters.

    -
    security
    F
    license
    -
    quality
    A server that enables seamless integration between local Ollama LLM instances and MCP-compatible applications, providing advanced task decomposition, evaluation, and workflow management capabilities.
    1
    Python
    • Apple
  • Why this server?

    Uses Ollama for efficient embedding generation, requiring it to be installed and running for vector operations

    -
    security
    F
    license
    -
    quality
    Provides a project memory bank and RAG context provider for enhanced code understanding and management through vector embeddings, integrated with RooCode and Cline.
    9
    Python
    • Apple
  • Why this server?

    Provides integration with Ollama for local AI model usage and processing

    -
    security
    F
    license
    -
    quality
    Facilitates enhanced interaction with large language models (LLMs) by providing intelligent context management, tool integration, and multi-provider AI model coordination for efficient AI-driven workflows.
    Python
  • Why this server?

    Uses Ollama as a Large Language Model provider to determine user intent and route requests

    -
    security
    F
    license
    -
    quality
    Facilitates executing system commands and retrieving web data using the Brave Search API by interpreting user intents via a Large Language Model (LLM).
    1
    Python
  • Why this server?

    Uses Ollama as the default embedding provider for local embeddings generation, supporting semantic documentation search and vector storage.

    -
    security
    F
    license
    -
    quality
    Enables AI assistants to enhance their responses with relevant documentation through a semantic vector search, offering tools for managing and processing documentation efficiently.
    62
    13
    TypeScript
  • Why this server?

    Integrates with Ollama to use the Deepseek model for AI capabilities through the MCP protocol

    -
    security
    F
    license
    -
    quality
    Enables seamless AI integration via Ollama's Deepseek model, providing protocol compliance and automatic configuration for clean AI-driven interactions.
    1
    Python
  • Why this server?

    Allows querying Ollama models directly from Claude with performance tracking, supporting selection of different models and providing context for queries.

    -
    security
    F
    license
    -
    quality
    Facilitates initiating Ollama queries via Claude and manages a simple note storage system with capabilities to add, summarize, and access notes using custom URIs.
    1
    2
    Python
    • Apple
  • Why this server?

    Leverages Ollama's LLM capabilities to interpret natural language questions, generate SQL queries, and provide AI-powered responses based on database results.

    -
    security
    F
    license
    -
    quality
    An interactive chat interface that combines Ollama's LLM capabilities with PostgreSQL database access through the Model Context Protocol (MCP). Ask questions about your data in natural language and get AI-powered responses backed by real SQL queries.
    28
    TypeScript
  • Why this server?

    Allows communication with locally available Ollama models (like llama2, codellama) while maintaining persistent conversation history.

    -
    security
    F
    license
    -
    quality
    A TypeScript-based server that provides a memory system for Large Language Models (LLMs), allowing users to interact with multiple LLM providers while maintaining conversation history and offering tools for managing providers and model configurations.
    20
    JavaScript
    • Apple