Skip to main content
Glama

File Context MCP

# Smithery configuration file: https://smithery.ai/docs/config#smitheryyaml startCommand: type: stdio configSchema: # JSON Schema defining the configuration options for the MCP. type: object required: - togetherApiKey - ollamaBaseUrl - modelName - port properties: togetherApiKey: type: string description: The API key for Together.ai. ollamaBaseUrl: type: string description: The base URL for the Ollama model service. modelName: type: string description: The name of the model to use with the LLM. port: type: number description: The port number on which the server will run. commandFunction: # A function that produces the CLI command to start the MCP on stdio. |- (config) => ({ command: 'node', args: ['dist/server.js'], env: { TOGETHER_API_KEY: config.togetherApiKey, OLLAMA_BASE_URL: config.ollamaBaseUrl, MODEL_NAME: config.modelName, PORT: config.port.toString() } })

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/compiledwithproblems/file-context-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server