Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@Venice MCP Servergenerate an image of a sunset over mountains"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
Venice MCP Server
A Model Context Protocol (MCP) server for Venice AI. Gives AI assistants like Claude access to Venice's text generation, image creation, text-to-speech, embeddings, and account management APIs.
What It Does
This MCP server lets Claude (or any MCP-compatible assistant) use Venice AI's features directly:
Chat with Llama, DeepSeek, Qwen, and other open models
Generate images using Stable Diffusion, FLUX, and more
Create speech from text with natural-sounding voices
Generate embeddings for search and RAG applications
Manage API keys and monitor rate limits
Quick Start
1. Get a Venice API Key
Sign up at venice.ai and create an API key in Settings → API.
2. Install
npm install -g venice-mcp3. Configure Claude Desktop
Add to your config file:
macOS:
~/Library/Application Support/Claude/claude_desktop_config.jsonWindows:
%APPDATA%\Claude\claude_desktop_config.jsonLinux:
~/.config/Claude/claude_desktop_config.json
{
"mcpServers": {
"venice": {
"command": "npx",
"args": ["-y", "venice-mcp"],
"env": {
"VENICE_API_KEY": "your-api-key-here"
}
}
}
}Restart Claude Desktop. You'll see Venice tools available in the 🔧 menu.
Available Tools
Generation
Tool | Description |
| Chat with Venice AI language models |
| Generate images from text descriptions |
| Upscale and enhance images |
| Convert text to spoken audio |
| Generate text embeddings |
Discovery
Tool | Description |
| List available models (text, image, code) |
| List character personas |
Account Management
Requires an admin-level API key:
Tool | Description |
| List all API keys |
| Create a new API key |
| Get details for a key |
| Delete an API key |
| View rate limits and usage |
| View rate limit history |
Examples
Ask Claude to generate an image:
"Use Venice to create an image of a sunset over mountains"
Chat with a specific model:
"Ask Venice's DeepSeek model to explain quantum computing"
Check your usage:
"Show my Venice API rate limits"
Supported Models
Language Models: Llama 3.3 70B, DeepSeek R1, Qwen 3, Mistral, and more
Image Models: Stable Diffusion 3.5, FLUX, Fluently XL
Embeddings: BGE-M3 (1024 dimensions)
TTS: Kokoro with multiple voice options
Run venice_list_models to see all currently available models.
Requirements
Node.js 18+
Venice AI API key
Development
Setup
git clone https://github.com/georgeglarson/venice-mcp.git
cd venice-mcp
npm install
npm run buildTesting
npm test # Run tests
npm run test:watch # Watch mode
npm run test:coverage # Coverage reportContributing
Contributions are welcome! Please see CONTRIBUTING.md for guidelines.
Links
Venice AI - Platform home
Venice API Docs - API reference
MCP Documentation - Protocol spec
License
MIT - see LICENSE
Resources
Looking for Admin?
Admins can modify the Dockerfile, update the server description, and track usage metrics. If you are the server author, to access the admin panel.