Skip to main content
Glama

TianGong-AI-MCP

by linancn
DEV_EN.md1.1 kB
# TianGong-AI-MCP [中文](https://github.com/linancn/tiangong-ai-mcp-local/blob/main/DEV_CN.md) | [English](https://github.com/linancn/tiangong-ai-mcp-local/blob/main/DEV_EN.md) TianGong AI Model Context Protocol (MCP) Local Server supports Streamable Http protocol. ## Development ### Environment Setup ```bash # Install Node.js curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.40.3/install.sh | bash nvm install 22 nvm use # Install dependencies npm install # Update dependencies npm update && npm ci ``` ### Code Formatting ```bash # Format code using the linter npm run lint ``` ### Local Testing #### Streamable Http Server ```bash # Launch the Streamable Http Server using MCP Inspector npm start:server ``` #### Launch MCP Inspector ```bash npx @modelcontextprotocol/inspector ``` ### Publish NPM Package ```bash npm login npm run build && npm publish ``` ### Push Docker Image ```bash # Build MCP server image using Dockerfile docker build -t linancn/tiangong-ai-mcp-server-local:0.0.1 . # Push Docker image docker push linancn/tiangong-ai-mcp-server-local:0.0.1 ```

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/linancn/tiangong-ai-mcp-local'

If you have feedback or need assistance with the MCP directory API, please join our Discord server