Skip to main content
Glama

TianGong-AI-MCP

by linancn
README.md1.11 kB
# TianGong-AI-MCP [中文](https://github.com/linancn/tiangong-ai-mcp-local/blob/main/README_CN.md) | [English](https://github.com/linancn/tiangong-ai-mcp-local/blob/main/README.md) TianGong AI Model Context Protocol (MCP) Local Server supports Streamable Http protocol. ## Starting MCP Server ### Streamable Http Server ```bash npm install -g @tiangong-ai/mcp-server-local npx dotenv -e .env -- \ npx -y -p @tiangong-ai/mcp-server-local tiangong-ai-mcp-http npm i -g pm2 pm2 start "npx --no-install tiangong-ai-mcp-http" --name tiangong-mcp-local --time pm2 restart tiangong-mcp-local pm2 stop tiangong-mcp-local pm2 logs tiangong-mcp-local pm2 delete tiangong-mcp-local pm2 status ``` ### Using Docker ```bash # Build MCP server image using Dockerfile (optional) docker build -t linancn/tiangong-ai-mcp-server-local:0.0.1 . # Pull MCP server image docker pull linancn/tiangong-ai-mcp-server-local:0.0.1 # Start MCP server using Docker docker run -d \ --name tiangong-ai-mcp-server-local \ --publish 9279:9279 \ --env-file .env \ linancn/tiangong-ai-mcp-server-local:0.0.1 ```

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/linancn/tiangong-ai-mcp-local'

If you have feedback or need assistance with the MCP directory API, please join our Discord server