Skip to main content
Glama
MidOSresearch

MidOS Research Protocol MCP

agent_handshake

Configure your AI agent's environment by declaring model, context window, languages, frameworks, and project goals to receive optimal settings for MidOS Research Protocol.

Instructions

Personalized agent onboarding. Declare your environment and get optimal config.

Call this FIRST when connecting to MidOS. Pass as much info as you know. Unknown fields can be left empty -- you'll get sensible defaults.

Args: model: Your model ID (e.g. 'claude-opus-4-6', 'gemini-2.5-pro', 'opus') context_window: Your context window in tokens (e.g. 200000). 0 = auto-detect from model. client: Your CLI/IDE (e.g. 'claude-code', 'cursor', 'windsurf', 'cline') languages: Comma-separated languages (e.g. 'python,typescript') frameworks: Comma-separated frameworks (e.g. 'fastapi,react') platform: Your OS (e.g. 'windows', 'linux', 'macos') project_goal: What you're working on (e.g. 'manga engine with SVG rendering')

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
modelNo
context_windowNo
clientNo
languagesNo
frameworksNo
platformNo
project_goalNo

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/MidOSresearch/mid-os-research-protocol-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server