With the OpenAI MCP Server, you can query OpenAI models directly from Claude using the MCP protocol.
Ask OpenAI models questions: Use the
ask-openaiendpoint to interact with GPT-4 or GPT-3.5-turbo modelsCustomize query parameters: Control response length with
max_tokens(1-4000) and creativity/randomness withtemperature(0-2)Integration with Claude: Configure the server in Claude Desktop for seamless use of OpenAI models
Local development: Clone, install, and test the server locally for debugging or customization
Allows querying OpenAI models directly from Claude using MCP protocol
OpenAI MCP Server
Query OpenAI models directly from Claude using MCP protocol.

Setup
Add to claude_desktop_config.json:
Related MCP server: OpenAI MCP Server
Development
Testing
License
MIT License