We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/wx-b/long-context-mcp'
If you have feedback or need assistance with the MCP directory API, please join our Discord server
{
"v": 1,
"id": "vllm-local-preset",
"request": {
"provider": {
"provider_preset": "vllm_local"
},
"rlm": {
"backend": "openai_compatible",
"environment": "docker",
"model_name": "qwen2.5-coder-7b-instruct",
"other_model_name": "qwen2.5-coder-3b-instruct",
"max_iterations": 12,
"timeout_sec": 90,
"backend_kwargs": {
"temperature": 0.2,
"max_tokens": 1200
},
"other_backend_kwargs": {
"temperature": 0.0,
"max_tokens": 384
}
},
"inputs": {
"prompt": "Replace with your task context",
"root_prompt": "Replace with your root instruction"
}
}
}