Skip to main content
Glama

agent_handshake

Configure personalized agent onboarding by declaring your environment details to receive optimal settings for connecting to the MidOS Research Protocol server.

Instructions

Personalized agent onboarding. Declare your environment and get optimal config.

Call this FIRST when connecting to MidOS. Pass as much info as you know. Unknown fields can be left empty -- you'll get sensible defaults.

Args: model: Your model ID (e.g. 'claude-opus-4-6', 'gemini-2.5-pro', 'opus') context_window: Your context window in tokens (e.g. 200000). 0 = auto-detect from model. client: Your CLI/IDE (e.g. 'claude-code', 'cursor', 'windsurf', 'cline') languages: Comma-separated languages (e.g. 'python,typescript') frameworks: Comma-separated frameworks (e.g. 'fastapi,react') platform: Your OS (e.g. 'windows', 'linux', 'macos') project_goal: What you're working on (e.g. 'manga engine with SVG rendering')

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
modelNo
context_windowNo
clientNo
languagesNo
frameworksNo
platformNo
project_goalNo

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/MidOSresearch/mid-os-research-protocol'

If you have feedback or need assistance with the MCP directory API, please join our Discord server