Skip to main content
Glama

OpenAPI to MCP Server

by jeweis

OpenAPI to MCP Server

这是一个OpenAPI转换服务器,提供了以下功能:

  • 将OpenAPI规范转换为MCP服务
  • 支持http header透传
  • 支持本地mcp 通用stdio使用方式、也支持docker部署为mcp server后通过streamable http方式使用

MCP 使用配置

本项目支持通过多种客户端配置 MCP 服务器,以便与各种 IDE 或工具集成。以下是一些常见客户端的配置示例:

Windsurf / Cursor / Claude

对于基于 Windsurf 框架的客户端(如 Cursor 和 Claude),您可以在 ~/.codeium/windsurf/mcp_config.json 文件中配置 MCP 服务器。以下是一个示例配置:

{ "mcpServers": { "openapi2mcpserver": { "disabled": false, "command": "uvx", "args": [ "openapi2mcpserver" ], "env": { "BASE_URL": "your_openapi3.0_host", "OPEN_API_DOC_JSON_URL": "your_openapi3.0_host doc json url" } } } }

请将 BASE_URL, OPEN_API_DOC_JSON_URL 替换为您的实际 OpenAPI 服务器地址和文档 JSON URL。

Cline

对于 Cline 客户端,您可以在其配置文件中添加类似的 MCP 服务器配置。具体的配置方式请参考 Cline 的官方文档。通常,您需要指定服务器的名称、命令、参数和环境变量。

// Cline 配置文件示例 (具体格式请参考 Cline 文档) { "mcpServers": { "openapi2mcpserver": { "command": "uvx", "args": [ "openapi2mcpserver" ], "env": { "BASE_URL": "your_openapi3.0_host", "OPEN_API_DOC_JSON_URL": "your_openapi3.0_host doc json url" } } } }

请将示例中的占位符替换为您的实际 OpenAPI 服务器地址和文档 JSON URL,并根据 Cline 的具体配置格式进行调整。

安装

  1. 克隆仓库
  2. 安装依赖:pip install -r requirements.txt
  3. 配置环境变量(参见下文)

配置

在项目根目录创建.env文件,包含以下环境变量:

BASE_URL=https://api.xxx.com OPEN_API_DOC_JSON_URL=https://api.xxx.com/api/v3/api-docs/default

运行

使用uvx安装并运行(推荐)

uvx --from openapi2mcpserver

docker部署方式

docker run -d -p 9087:9087 -e BASE_URL=http://xxx.xx.xx.xxx:8044/xiaogj-ai-api -e OPEN_API_DOC_JSON_URL=http://xxx.xx.xx.xxx:8044/xiaogj-ai-api/v3/api-docs -e SERVER_NAME=OpenAPI2MCP-Docker-Server -e "ROUTE_MAPS=[{\"methods\":[\"GET\",\"POST\"],\"pattern\":\"^/tool/.*\"}]" jeweis/openapi2mcp:latest

说明

  1. “xxx.xx.xx.xxx”需要换成本机ip
  2. 环境变量按实际需求配置
-
security - not tested
F
license - not found
-
quality - not tested

remote-capable server

The server can be hosted and run remotely because it primarily relies on remote services or has no dependency on the local environment.

A service that converts OpenAPI specifications into MCP tools, enabling AI assistants to interact with your API endpoints through natural language.

  1. MCP 使用配置
    1. Windsurf / Cursor / Claude
    2. Cline
  2. 安装
    1. 配置
      1. 运行
        1. 使用uvx安装并运行(推荐)
      2. docker部署方式
        1. 说明

      Related MCP Servers

      • -
        security
        A
        license
        -
        quality
        A command-line tool that transforms any OpenAPI service into a Model Context Protocol (MCP) server, enabling seamless integration with AI agents and tools that support the MCP specification.
        Last updated -
        3
        5
        TypeScript
        MIT License
      • -
        security
        F
        license
        -
        quality
        A dynamic proxy that converts OpenAPI Specification (OAS) endpoints into Message Communication Protocol (MCP) tools, allowing AI agents to use existing REST APIs as if they were native MCP tools without manual implementation.
        Last updated -
        16
        TypeScript
      • -
        security
        F
        license
        -
        quality
        An auto-generated MCP server that enables interaction with the OpenAI API, allowing users to access OpenAI's models and capabilities through the Multi-Agent Conversation Protocol.
        Last updated -
        Python
      • A
        security
        A
        license
        A
        quality
        An MCP server that automatically converts OpenAPI specifications into MCP tools by scanning a folder for spec files, requiring no configuration files or separate servers.
        Last updated -
        10
        3
        TypeScript
        MIT License

      View all related MCP servers

      MCP directory API

      We provide all the information about MCP servers via our MCP API.

      curl -X GET 'https://glama.ai/api/mcp/v1/servers/jeweis/openapi2mcpserver'

      If you have feedback or need assistance with the MCP directory API, please join our Discord server