With the OpenAI MCP Server, you can query OpenAI models directly from Claude using the MCP protocol.
Ask OpenAI models questions: Use the
ask-openaiendpoint to interact with GPT-4 or GPT-3.5-turbo modelsCustomize query parameters: Control response length with
max_tokens(1-4000) and creativity/randomness withtemperature(0-2)Integration with Claude: Configure the server in Claude Desktop for seamless use of OpenAI models
Local development: Clone, install, and test the server locally for debugging or customization
Allows querying OpenAI models directly from Claude using MCP protocol
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@OpenAI MCP Serversummarize this article in 3 bullet points"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
OpenAI MCP Server
Query OpenAI models directly from Claude using MCP protocol.

Setup
Add to claude_desktop_config.json:
Related MCP server: OpenAI MCP Server
Development
Testing
License
MIT License