Skip to main content
Glama
package.json1.49 kB
{ "name": "ask-mcp", "version": "1.1.0", "description": "A powerful MCP server providing access to multiple AI models (Grok, Gemini, Kimi, Qwen, GLM) through OpenRouter. Perfect for Cursor, Claude Desktop, and other MCP clients.", "main": "dist/index.js", "type": "module", "bin": { "ask-mcp": "dist/index.js" }, "keywords": [ "mcp", "model-context-protocol", "ai", "openrouter", "grok", "gemini", "kimi", "qwen", "glm", "cursor", "claude", "anthropic", "typescript", "llm", "chatbot" ], "author": "SerPepe <serpepexbtt@proton.me>", "license": "MIT", "repository": { "type": "git", "url": "https://github.com/SerPepe/ask-mcp.git" }, "bugs": { "url": "https://github.com/SerPepe/ask-mcp/issues" }, "homepage": "https://github.com/SerPepe/ask-mcp#readme", "files": [ "dist/", "README.md", "LICENSE", "mcp-config-example.json" ], "scripts": { "build": "tsc", "start": "node dist/index.js", "dev": "tsx src/index.ts", "prepublishOnly": "npm run build", "prepare": "npm run build", "setup": "npm run build && node setup-oauth.js" }, "dependencies": { "@modelcontextprotocol/sdk": "^0.4.0", "axios": "^1.6.0", "dotenv": "^17.2.1", "express": "^5.1.0", "open": "^10.2.0" }, "devDependencies": { "@types/express": "^5.0.3", "@types/node": "^20.0.0", "tsx": "^4.0.0", "typescript": "^5.0.0" } }

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/SerPepe/Ask-MCP'

If you have feedback or need assistance with the MCP directory API, please join our Discord server