Skip to main content
Glama

TianGong-AI-MCP

中文 | English

TianGong AI Model Context Protocol (MCP) Local Server supports Streamable Http protocol.

Starting MCP Server

Streamable Http Server

npm install -g @tiangong-ai/mcp-server-local

npx dotenv -e .env -- \
npx -y -p @tiangong-ai/mcp-server-local tiangong-ai-mcp-http

npm i -g pm2

pm2 start "npx --no-install tiangong-ai-mcp-http" --name tiangong-mcp-local --time

pm2 restart tiangong-mcp-local
pm2 stop tiangong-mcp-local
pm2 logs tiangong-mcp-local

pm2 delete tiangong-mcp-local

pm2 status

Using Docker

# Build MCP server image using Dockerfile (optional)
docker build -t linancn/tiangong-ai-mcp-server-local:0.0.1 .

# Pull MCP server image
docker pull linancn/tiangong-ai-mcp-server-local:0.0.1

# Start MCP server using Docker
docker run -d \
    --name tiangong-ai-mcp-server-local \
    --publish 9279:9279 \
    --env-file .env \
    linancn/tiangong-ai-mcp-server-local:0.0.1
-
security - not tested
A
license - permissive license
-
quality - not tested

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/linancn/tiangong-ai-mcp-local'

If you have feedback or need assistance with the MCP directory API, please join our Discord server