Skip to main content
Glama
mcp.json6.09 kB
{ "mcpServers": { "github": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-github" ], "env": { "GITHUB_PERSONAL_ACCESS_TOKEN": "${GITHUB_TOKEN}" } }, "filesystem": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-filesystem", "${workspaceFolder}" ] }, "git": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-git", "--repository", "${workspaceFolder}" ] }, "puppeteer": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-puppeteer" ] }, "brave-search": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-brave-search" ] }, "slack": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-slack" ], "env": { "SLACK_BOT_TOKEN": "${SLACK_BOT_TOKEN}" } }, "postgres": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-postgres", "postgresql://localhost/postgres" ] }, "notion": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-notion" ], "env": { "NOTION_API_KEY": "${NOTION_API_KEY}" } }, "sequential-thinking": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-sequential-thinking" ] }, "memory": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-memory" ] }, "fetch": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-fetch" ] }, "code-intelligence": { "command": "python", "args": [ "C:\\AIStack-MCP\\mcp_intelligence_server.py", "--workspace", "${workspaceFolder}" ], "env": { "OLLAMA_URL": "http://localhost:11434", "QDRANT_URL": "http://localhost:6333" } } } }

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/mjdevaccount/AIStack-MCP'

If you have feedback or need assistance with the MCP directory API, please join our Discord server