Skip to main content
Glama

Ollama MCP Server

by Jadael
CHANGELOG.md2.47 kB
# Changelog ## Version 1.0.0 - Initial Release ### Features #### Core MCP Server - Implemented MCP server using `@modelcontextprotocol/sdk` - Ollama integration via HTTP API (localhost:11434) - Default model: `gemma3:27b` with `gemma3:4b` fallback - 120-second timeout for Ollama requests - Proper error handling for connection issues #### String-Based Tools (7 tools) Tools that accept code as string parameters: 1. `ollama_generate_code` - Generate new code from requirements 2. `ollama_explain_code` - Explain how code works 3. `ollama_review_code` - Review code for issues and improvements 4. `ollama_refactor_code` - Refactor code to improve quality 5. `ollama_fix_code` - Fix bugs or errors in code 6. `ollama_write_tests` - Generate unit tests 7. `ollama_general_task` - Execute any general coding task #### File-Aware Tools (4 tools) - Major Innovation! Tools that read files directly on the MCP server, providing massive token savings: 8. `ollama_review_file` - Review a file by path 9. `ollama_explain_file` - Explain a file by path 10. `ollama_analyze_files` - Analyze multiple files together 11. `ollama_generate_code_with_context` - Generate code using reference files **Token Savings**: File-aware tools reduce conversation token usage by ~98.75% compared to traditional read-then-analyze workflows. ### Documentation - Comprehensive README.md with setup instructions - Detailed test.md with test cases and validation procedures - .gitignore for clean repository - CHANGELOG.md (this file) ### Technical Details - Node.js 18+ required - ES modules (`"type": "module"`) - Dependencies: `@modelcontextprotocol/sdk`, `axios` - Cross-platform support (Windows, macOS, Linux) - Absolute file paths required for file-aware tools ### Performance Characteristics - Small tasks: 30-90 seconds - Medium tasks: 90-180 seconds - Large tasks: 180-300 seconds - Token savings: Up to 98.75% with file-aware tools ### Known Limitations - Timeouts can occur with large models on slower hardware (expected) - No streaming responses (synchronous only) - No caching of file contents (reads on every call) - Requires Ollama to be running locally ### Future Enhancement Ideas - File content caching for repeated operations - Glob pattern support for multi-file operations - Streaming responses for better UX - Auto-context: automatically find related files - File writing capabilities - Configurable timeout per tool - Model selection hints based on task complexity

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/Jadael/OllamaClaude'

If you have feedback or need assistance with the MCP directory API, please join our Discord server