AsyncAPI-MCP
AsyncAPI MCP Server
An MCP (Model Context Protocol) server that gives AI assistants access to the AsyncAPI specification. Search, explore, and retrieve any version of the spec directly from your coding tool.
Features
Search the AsyncAPI specification by keyword
Retrieve specific sections by heading or slug
List all stable spec versions available as GitHub tags
Get metadata about the spec (version, source, cache info, size)
Version-aware — query any released spec version, or default to the latest
Caching — ETag/Last-Modified-based HTTP caching with a 10-minute TTL on tag lookups
Prerequisites
Node.js v20 or later
Installation
npm installRunning the Server
Streamable HTTP (local development)
npm run devThe server starts on http://localhost:3000/mcp by default. Set the PORT environment variable to use a different port:
PORT=8080 npm run devStdio Transport (for remote deployment)
npm startThis runs the server over stdio, which is the standard transport for hosted MCP servers.
Building
npm run buildCompiles TypeScript from src/ to dist/.
Available Tools
Tool | Description | Parameters |
| List stable AsyncAPI spec versions available as GitHub tags | None |
| Return source, version, cache, and size metadata for a spec |
|
| Search the spec and return matching snippets |
|
| Return a section by heading text or slug |
|
Available Resources
Resource | URI | Description |
Latest AsyncAPI Spec |
| The latest AsyncAPI markdown specification from the master branch |
AsyncAPI Spec by Version |
| A specific version of the spec fetched from the matching GitHub release tag |
Configuration for AI Coding Tools
The server uses the Streamable HTTP transport. Make sure the server is running before connecting your AI tool.
Claude Desktop
Add to ~/Library/Application Support/Claude/claude_desktop_config.json (macOS) or %APPDATA%\Claude\claude_desktop_config.json (Windows):
{
"mcpServers": {
"asyncapi": {
"url": "http://localhost:3000/mcp"
}
}
}Cursor
Add to .cursor/mcp.json in your project root:
{
"mcpServers": {
"asyncapi": {
"url": "http://localhost:3000/mcp"
}
}
}VS Code Copilot
Add to .vscode/mcp.json in your project root:
{
"servers": {
"asyncapi": {
"url": "http://localhost:3000/mcp",
"type": "http"
}
}
}Windsurf
Add to your Windsurf MCP settings:
{
"mcpServers": {
"asyncapi": {
"url": "http://localhost:3000/mcp"
}
}
}Cline
In Cline's MCP settings, add:
{
"mcpServers": {
"asyncapi": {
"url": "http://localhost:3000/mcp"
}
}
}OpenCode
Add to your OpenCode configuration:
{
"mcp": {
"servers": {
"asyncapi": {
"url": "http://localhost:3000/mcp"
}
}
}
}Zed
Add to your Zed settings.json:
{
"context_servers": {
"asyncapi": {
"url": "http://localhost:3000/mcp"
}
}
}Usage Examples
Once configured, you can ask your AI assistant questions like:
"What does the AsyncAPI spec say about server objects?"
"Search the AsyncAPI spec for 'channels'"
"Get the Info Object section from version 2.6.0"
"List all available AsyncAPI spec versions"
"What are the differences between messages in AsyncAPI 2.x and 3.x?"
"Show me the spec section about schema definitions"
Development
# Install dependencies
npm install
# Build TypeScript to dist/
npm run build
# Run the HTTP server (local development)
npm run dev
# Run the stdio server (for deployment)
npm startLicense
MIT
Latest Blog Posts
MCP directory API
We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/Souvikns/asyncapi-mcp'
If you have feedback or need assistance with the MCP directory API, please join our Discord server