Skip to main content
Glama

Vibe Check MCP

# Agent Quickstart Vibe Check MCP is a lightweight oversight layer for AI agents. It exposes two tools: - **vibe_check** – prompts you with clarifying questions to prevent tunnel vision. - **vibe_learn** – optional logging of mistakes and successes for later review. The server supports Gemini, OpenAI and OpenRouter LLMs. History is maintained across requests when a `sessionId` is provided. ## Setup 1. Install dependencies and build: ```bash npm install npm run build ``` 2. Supply the following environment variables as needed: - `GEMINI_API_KEY` - `OPENAI_API_KEY` - `OPENROUTER_API_KEY` - `DEFAULT_LLM_PROVIDER` (gemini | openai | openrouter) - `DEFAULT_MODEL` (e.g., gemini-2.5-pro) 3. Start the server: ```bash npm start ``` ## Testing Run unit tests with `npm test`. Example request generators are provided: - `alt-test-gemini.js` - `alt-test-openai.js` - `alt-test.js` (OpenRouter) Each script writes a `request.json` file that you can pipe to the server: ```bash node build/index.js < request.json ``` ## Integration Tips Call `vibe_check` regularly with your goal, plan and current progress. Use `vibe_learn` whenever you want to record a resolved issue. Full API details are in `docs/technical-reference.md`.

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/PV-Bhat/vibe-check-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server