Best Ollama MCP Servers
Ollama is an open-source project that allows you to run large language models (LLMs) locally on your own hardware, providing a way to use AI capabilities privately without sending data to external services.
Why this server?
Integrates with Ollama for local embedding generation, enabling semantic search, similarity matching, and automated metadata suggestions using locally-hosted language models.
AsecurityAlicenseCqualityMCP Server for local knowledge management. Semantic + keywords + tagsLast updated a day ago113MITWhy this server?
Supports local embeddings and LLM processing via Ollama for runtime-configurable document indexing and search.
AsecurityAlicense-qualitySelf-hosted knowledge backend for AI agents. Provides 11 MCP tools for hybrid vector + keyword search, container-isolated knowledge bases, and 4 storage connectors (S3, Azure Blob, MinIO, filesystem). Built with .NET, runs via Docker.Last updated 4 hours ago1118MITWhy this server?
Provides an agent runtime for multi-turn conversations with tool-use capabilities using local LLMs through Ollama.
AsecurityAlicense-qualityGeneral purpose toolbox for AI applicationsLast updated 18 days ago31004MITWhy this server?
Enables AI-powered content processing including query rewriting, page summarization, and research synthesis using configurable Ollama models.
AsecurityAlicenseAqualityEnables AI assistants to perform comprehensive web research through tiered search, secure URL fetching with markdown conversion, and automated multi-source synthesis pipelines. Provides read-only tools with configurable caching, SSRF protection, and optional LLM-powered summarization for search results and content analysis.Last updated 3 days ago31MITWhy this server?
Supports local model execution through Ollama via the Marga multi-model router for privacy-focused or local task processing.
AsecurityAlicense-qualityMultilingual intelligence + memory + safety + voice layer for autonomous AI agentsLast updated 21 days ago211MITWhy this server?
Enables routing to locally-hosted models for zero-cost, privacy-preserving, offline inference as the first tier in fallback chains.
AsecurityAlicenseAqualityRoutes your AI tasks to the best available model across 20+ providers — automatically selecting based on task type, budget, and subscription pressure. Supports text, image, video, and audio with built-in cost optimization and fallback chains.Last updated 2 days ago309MITWhy this server?
Allows for the quantization of HuggingFace models into the GGUF format, making them compatible for local execution and deployment within the Ollama runtime.
AsecurityAlicenseAqualityMCP server for LLM quantization. Compress any HuggingFace model to GGUF, GPTQ, or AWQ format. 6 tools: info, check, recommend, quantize, evaluate, push. Self-contained Python server — no external CLI needed.Last updated 10 days ago6MITWhy this server?
Enables local LLM execution via Ollama for fully private genomic analysis without sending genetic data to external cloud APIs.
AsecurityAlicenseAqualityUpload your raw DNA file from 23andMe, AncestryDNA, MyHeritage, FamilyTreeDNA, or any VCF -- and watch a team of AI agents fan out across 16+ public genomics databases, share discoveries with each other in real time, and produce a comprehensive health report. Everything runs on your machine. Nothing is uploaded anywhere.Last updated 3 days ago2412MITWhy this server?
Supports local Ollama instances as an LLM provider for processing prompt optimizations using self-hosted models.
AsecurityAlicense-qualityAn MCP server that transforms vague prompts into platform-optimized prompts for 58 AI platforms across 7 categories. Send a raw prompt. Get back a version specifically optimized for Midjourney, DALL-E, Sora, Runway, ElevenLabs, Claude, ChatGPT, or any of the 58 supported platforms — with the right syntax, parameters, and structure each platform expects.Last updated a month ago751Apache 2.0