Skip to main content
Glama
orneryd

M.I.M.I.R - Multi-agent Intelligent Memory & Insight Repository

by orneryd
CHANGELOG.md2.06 kB
# Changelog ## [0.1.0] - 2025-11-19 ### Added - **Model Selection Support**: Accept `model` parameter in chat requests (OpenAI-compatible) - Configure via `mimir.model` setting (default: `gpt-4.1`) - Supports any model available in Mimir: GPT-4, Claude, Llama, etc. - Model is preserved across conversation (unless explicitly changed) ### Fixed - **Chat History Handling**: Proper context preservation across conversation - System message (preamble) is only added once per conversation - Follow-up messages reuse the original preamble automatically - Slash commands can override preamble mid-conversation - No more duplicate system messages - **Tool Parameters**: All tool configurations preserved across messages - Vector search depth/limit/similarity persist throughout conversation - Tool enable/disable settings maintained per chat thread ### Improved - **Chat Continuity**: Better conversation flow - Starting with `/research` keeps research mode for all follow-ups - Model selection persists (from settings) - All configurations apply to entire conversation unless changed ## Usage Examples ### Basic Chat ``` @mimir how does graph-rag work? ``` Uses: default preamble (`mimir-v2`) + default model (`gpt-4.1`) ### With Slash Command ``` @mimir /research what are Neo4j best practices? ``` Uses: `research` preamble + default model + higher search depth ### Follow-up Messages ``` @mimir /debug why is my API failing? @mimir can you check the logs? ``` Both use: `debug` preamble (preserves from first message) + same model ### Change Model Mid-Conversation Update `mimir.model` in settings to switch models (takes effect next message) ## Configuration Reference | Setting | Default | Description | |---------|---------|-------------| | `mimir.model` | `gpt-4.1` | LLM model to use | | `mimir.defaultPreamble` | `mimir-v2` | Default chatmode | | `mimir.vectorSearch.depth` | `1` | Graph traversal depth (1-3) | | `mimir.enableTools` | `true` | Enable MCP tools | | `mimir.maxToolCalls` | `3` | Max tool calls per response |

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/orneryd/Mimir'

If you have feedback or need assistance with the MCP directory API, please join our Discord server