With the OpenAI MCP Server, you can query OpenAI models directly from Claude using the MCP protocol.
Ask OpenAI models questions: Use the
ask-openaiendpoint to interact with GPT-4 or GPT-3.5-turbo modelsCustomize query parameters: Control response length with
max_tokens(1-4000) and creativity/randomness withtemperature(0-2)Integration with Claude: Configure the server in Claude Desktop for seamless use of OpenAI models
Local development: Clone, install, and test the server locally for debugging or customization
Allows querying OpenAI models directly from Claude using MCP protocol
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@OpenAI MCP Serversummarize this article in 3 bullet points"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
OpenAI MCP Server
Query OpenAI models directly from Claude using MCP protocol.

Setup
Add to claude_desktop_config.json:
{
"mcpServers": {
"openai-server": {
"command": "python",
"args": ["-m", "src.mcp_server_openai.server"],
"env": {
"PYTHONPATH": "C:/path/to/your/mcp-server-openai",
"OPENAI_API_KEY": "your-key-here"
}
}
}
}Related MCP server: OpenAI MCP Server
Development
git clone https://github.com/pierrebrunelle/mcp-server-openai
cd mcp-server-openai
pip install -e .Testing
# Run tests from project root
pytest -v test_openai.py -s
# Sample test output:
Testing OpenAI API call...
OpenAI Response: Hello! I'm doing well, thank you for asking...
PASSEDLicense
MIT License
Resources
Looking for Admin?
Admins can modify the Dockerfile, update the server description, and track usage metrics. If you are the server author, to access the admin panel.