DeepSeek MCP Server
MCP server for DeepSeek AI models (Chat + Reasoner). Supports stdio and HTTP transports, Docker deployment, multi-turn sessions, model fallback with circuit breaker, function calling, thinking mode, JSON output, multimodal input, and cost tracking.
Compatible with:
Claude Code CLI
Gemini CLI
Any MCP-compatible client (Cursor, Windsurf, etc.)
Note: This is an unofficial community project and is not affiliated with DeepSeek.
Quick Start
Remote (No Install)
Use the hosted endpoint directly — no npm install, no Node.js required. Bring your own DeepSeek API key:
Claude Code:
claude mcp add --transport http deepseek \
https://deepseek-mcp.tahirl.com/mcp \
--header "Authorization: Bearer YOUR_DEEPSEEK_API_KEY"Cursor / Windsurf / VS Code:
{
"mcpServers": {
"deepseek": {
"url": "https://deepseek-mcp.tahirl.com/mcp",
"headers": {
"Authorization": "Bearer ${DEEPSEEK_API_KEY}"
}
}
}
}Local (stdio)
Claude Code:
claude mcp add -s user deepseek npx @arikusi/deepseek-mcp-server -e DEEPSEEK_API_KEY=your-key-hereGemini CLI:
gemini mcp add deepseek npx @arikusi/deepseek-mcp-server -e DEEPSEEK_API_KEY=your-key-hereScope options (Claude Code):
-s user: Available in all your projects (recommended)-s local: Only in current project (default)-s project: Project-specific.mcp.jsonfile
Get your API key: https://platform.deepseek.com
Features
DeepSeek V3.2: Both models now run DeepSeek-V3.2 (since Sept 2025)
Multi-Turn Sessions: Conversation context preserved across requests via
session_idparameterModel Fallback & Circuit Breaker: Automatic fallback between models with circuit breaker protection against cascading failures
MCP Resources:
deepseek://models,deepseek://config,deepseek://usage— query model info, config, and usage statsThinking Mode: Enable thinking on deepseek-chat with
thinking: {type: "enabled"}JSON Output Mode: Structured JSON responses with
json_mode: trueFunction Calling: OpenAI-compatible tool use with up to 128 tool definitions
Cache-Aware Cost Tracking: Automatic cost calculation with cache hit/miss breakdown
Session Management Tool: List, delete, and clear sessions via
deepseek_sessionstoolConfigurable: Environment-based configuration with validation
12 Prompt Templates: Templates for debugging, code review, function calling, and more
Streaming Support: Real-time response generation
Multimodal Ready: Content part types for text + image input (enable with
ENABLE_MULTIMODAL=true)Remote Endpoint: Hosted at
deepseek-mcp.tahirl.com/mcp— BYOK (Bring Your Own Key), no install neededHTTP Transport: Self-hosted remote access via Streamable HTTP with
TRANSPORT=httpDocker Ready: Multi-stage Dockerfile with health checks for containerized deployment
Tested: 253 tests with 90%+ code coverage
Type-Safe: Full TypeScript implementation
MCP Compatible: Works with any MCP-compatible CLI (Claude Code, Gemini CLI, etc.)
Installation
Prerequisites
Node.js 18+
A DeepSeek API key (get one at https://platform.deepseek.com)
Manual Installation
If you prefer to install manually:
npm install -g @arikusi/deepseek-mcp-serverFrom Source
Clone the repository
git clone https://github.com/arikusi/deepseek-mcp-server.git
cd deepseek-mcp-serverInstall dependencies
npm installBuild the project
npm run buildUsage
Once configured, your MCP client will have access to deepseek_chat and deepseek_sessions tools, plus 3 MCP resources.
Example prompts:
"Use DeepSeek to explain quantum computing"
"Ask DeepSeek Reasoner to solve: If I have 10 apples and buy 5 more..."Your MCP client will automatically call the deepseek_chat tool.
Manual Configuration (Advanced)
If your MCP client doesn't support the add command, manually add to your config file:
{
"mcpServers": {
"deepseek": {
"command": "npx",
"args": ["@arikusi/deepseek-mcp-server"],
"env": {
"DEEPSEEK_API_KEY": "your-api-key-here"
}
}
}
}Config file locations:
Claude Code:
~/.claude.json(add toprojects["your-project-path"].mcpServerssection)Other MCP clients: Check your client's documentation for config file location
Available Tools
deepseek_chat
Chat with DeepSeek AI models with automatic cost tracking and function calling support.
Parameters:
messages(required): Array of conversation messagesrole: "system" | "user" | "assistant" | "tool"content: Message texttool_call_id(optional): Required for tool role messages
model(optional): "deepseek-chat" (default) or "deepseek-reasoner"temperature(optional): 0-2, controls randomness (default: 1.0). Ignored when thinking mode is enabled.max_tokens(optional): Maximum tokens to generate (deepseek-chat: max 8192, deepseek-reasoner: max 65536)stream(optional): Enable streaming mode (default: false)tools(optional): Array of tool definitions for function calling (max 128)tool_choice(optional): "auto" | "none" | "required" |{type: "function", function: {name: "..."}}thinking(optional): Enable thinking mode{type: "enabled"}json_mode(optional): Enable JSON output mode (supported by both models)session_id(optional): Session ID for multi-turn conversations. Previous context is automatically prepended.
Response includes:
Content with formatting
Function call results (if tools were used)
Request information (tokens, model, cost in USD)
Structured data with
cost_usdandtool_callsfields
Example:
{
"messages": [
{
"role": "user",
"content": "Explain the theory of relativity in simple terms"
}
],
"model": "deepseek-chat",
"temperature": 0.7,
"max_tokens": 1000
}DeepSeek Reasoner Example:
{
"messages": [
{
"role": "user",
"content": "If I have 10 apples and eat 3, then buy 5 more, how many do I have?"
}
],
"model": "deepseek-reasoner"
}The reasoner model will show its thinking process in <thinking> tags followed by the final answer.
Function Calling Example:
{
"messages": [
{
"role": "user",
"content": "What's the weather in Istanbul?"
}
],
"tools": [
{
"type": "function",
"function": {
"name": "get_weather",
"description": "Get current weather for a location",
"parameters": {
"type": "object",
"properties": {
"location": {
"type": "string",
"description": "City name"
}
},
"required": ["location"]
}
}
}
],
"tool_choice": "auto"
}When the model decides to call a function, the response includes tool_calls with the function name and arguments. You can then send the result back using a tool role message with the matching tool_call_id.
Thinking Mode Example:
{
"messages": [
{
"role": "user",
"content": "Analyze the time complexity of quicksort"
}
],
"model": "deepseek-chat",
"thinking": { "type": "enabled" }
}When thinking mode is enabled, temperature, top_p, frequency_penalty, and presence_penalty are automatically ignored.
JSON Output Mode Example:
{
"messages": [
{
"role": "user",
"content": "Return a json object with name, age, and city fields for a sample user"
}
],
"model": "deepseek-chat",
"json_mode": true
}JSON mode ensures the model outputs valid JSON. Include the word "json" in your prompt for best results. Supported by both deepseek-chat and deepseek-reasoner.
Multi-Turn Session Example:
{
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
],
"session_id": "my-session-1"
}Use the same session_id across requests to maintain conversation context. The server stores messages in memory and automatically prepends history to each request.
deepseek_sessions
Manage conversation sessions.
Parameters:
action(required): "list" | "clear" | "delete"session_id(optional): Required when action is "delete"
Examples:
{"action": "list"}
{"action": "delete", "session_id": "my-session-1"}
{"action": "clear"}Available Resources
MCP Resources provide read-only data about the server:
Resource URI | Description |
| Available models with capabilities, context limits, and pricing |
| Current server configuration (API key masked) |
| Real-time usage statistics (requests, tokens, costs, sessions) |
Model Fallback & Circuit Breaker
When a model fails with a retryable error (429, 503, timeout), the server automatically falls back to the other model:
deepseek-chatfails → triesdeepseek-reasonerdeepseek-reasonerfails → triesdeepseek-chat
The circuit breaker protects against cascading failures:
After
CIRCUIT_BREAKER_THRESHOLDconsecutive failures (default: 5), the circuit opens (fast-fail mode)After
CIRCUIT_BREAKER_RESET_TIMEOUTms (default: 30000), it enters half-open state and sends a probe requestIf the probe succeeds, the circuit closes and normal operation resumes
Fallback can be disabled with FALLBACK_ENABLED=false.
Available Prompts
Prompt templates (12 total):
Core Reasoning
debug_with_reasoning: Debug code with step-by-step analysis
code_review_deep: Comprehensive code review (security, performance, quality)
research_synthesis: Research topics and create structured reports
strategic_planning: Create strategic plans with reasoning
explain_like_im_five: Explain complex topics in simple terms
Advanced
mathematical_proof: Prove mathematical statements rigorously
argument_validation: Analyze arguments for logical fallacies
creative_ideation: Generate creative ideas with feasibility analysis
cost_comparison: Compare LLM costs for tasks
pair_programming: Interactive coding with explanations
Function Calling
function_call_debug: Debug function calling issues with tool definitions and messages
create_function_schema: Generate JSON Schema for function calling from natural language
Each prompt is optimized for the DeepSeek Reasoner model to provide detailed reasoning.
Models
Both models run DeepSeek-V3.2 with unified pricing.
deepseek-chat
Best for: General conversations, coding, content generation
Speed: Fast
Context: 128K tokens
Max Output: 8K tokens (default 4K)
Mode: Non-thinking (can enable thinking via parameter)
Features: Thinking mode, JSON mode, function calling, FIM completion
Pricing: $0.028/1M cache hit, $0.28/1M cache miss, $0.42/1M output
deepseek-reasoner
Best for: Complex reasoning, math, logic problems, multi-step tasks
Speed: Slower (shows thinking process)
Context: 128K tokens
Max Output: 64K tokens (default 32K)
Mode: Thinking (always active, chain-of-thought reasoning)
Features: JSON mode, function calling
Output: Both reasoning process and final answer
Pricing: $0.028/1M cache hit, $0.28/1M cache miss, $0.42/1M output
Configuration
The server is configured via environment variables. All settings except DEEPSEEK_API_KEY are optional.
Variable | Default | Description |
| (required) | Your DeepSeek API key |
|
| Custom API endpoint |
|
| Default model for requests |
|
| Show cost info in responses |
|
| Request timeout in milliseconds |
|
| Maximum retry count for failed requests |
|
| Skip startup API connection test |
|
| Maximum message content length (characters) |
|
| Session time-to-live in minutes |
|
| Maximum number of concurrent sessions |
|
| Enable automatic model fallback on errors |
|
| Consecutive failures before circuit opens |
|
| Milliseconds before circuit half-opens |
|
| Max messages per session (sliding window) |
|
| Enable multimodal (image) input support |
|
| Transport mode: |
|
| HTTP server port (when TRANSPORT=http) |
Example with custom config:
claude mcp add -s user deepseek npx @arikusi/deepseek-mcp-server \
-e DEEPSEEK_API_KEY=your-key \
-e SHOW_COST_INFO=false \
-e REQUEST_TIMEOUT=30000Development
Project Structure
deepseek-mcp-server/
├── worker/ # Cloudflare Worker (remote BYOK endpoint)
│ ├── src/index.ts # Worker entry point
│ ├── wrangler.toml # Cloudflare config
│ └── package.json
├── src/
│ ├── index.ts # Entry point, bootstrap
│ ├── server.ts # McpServer factory (auto-version)
│ ├── deepseek-client.ts # DeepSeek API wrapper (circuit breaker + fallback)
│ ├── config.ts # Centralized config with Zod validation
│ ├── cost.ts # Cost calculation and formatting
│ ├── schemas.ts # Zod input validation schemas
│ ├── types.ts # TypeScript types + type guards
│ ├── errors.ts # Custom error classes
│ ├── session.ts # In-memory session store (multi-turn)
│ ├── circuit-breaker.ts # Circuit breaker pattern
│ ├── usage-tracker.ts # Usage statistics tracker
│ ├── transport-http.ts # Streamable HTTP transport (Express)
│ ├── tools/
│ │ ├── deepseek-chat.ts # deepseek_chat tool (sessions + fallback)
│ │ ├── deepseek-sessions.ts # deepseek_sessions tool
│ │ └── index.ts # Tool registration aggregator
│ ├── resources/
│ │ ├── models.ts # deepseek://models resource
│ │ ├── config.ts # deepseek://config resource
│ │ ├── usage.ts # deepseek://usage resource
│ │ └── index.ts # Resource registration aggregator
│ └── prompts/
│ ├── core.ts # 5 core reasoning prompts
│ ├── advanced.ts # 5 advanced prompts
│ ├── function-calling.ts # 2 function calling prompts
│ └── index.ts # Prompt registration aggregator
├── dist/ # Compiled JavaScript
├── llms.txt # AI discoverability index
├── llms-full.txt # Full docs for LLM context
├── vitest.config.ts # Test configuration
├── package.json
├── tsconfig.json
└── README.mdBuilding
npm run buildWatch Mode (for development)
npm run watchTesting
# Run all tests
npm test
# Watch mode
npm run test:watch
# With coverage report
npm run test:coverageTesting Locally
# Set API key
export DEEPSEEK_API_KEY="your-key"
# Run the server
npm startThe server will start and wait for MCP client connections via stdio.
Remote Endpoint (Hosted)
A hosted BYOK (Bring Your Own Key) endpoint is available at:
https://deepseek-mcp.tahirl.com/mcpSend your DeepSeek API key as Authorization: Bearer <key>. No server-side API key stored — your key is used directly per request. Powered by Cloudflare Workers (global edge, zero cold start).
Note: The
deepseek-reasonermodel may take over 30 seconds for complex queries. Some MCP clients (e.g. Claude Code) have built-in tool call timeouts that may interrupt long-running requests. For complex tasks,deepseek-chatis recommended.
# Test health
curl https://deepseek-mcp.tahirl.com/health
# Test MCP (requires auth)
curl -X POST https://deepseek-mcp.tahirl.com/mcp \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_KEY" \
-d '{"jsonrpc":"2.0","method":"initialize","params":{"capabilities":{}},"id":1}'HTTP Transport (Self-Hosted)
Run your own HTTP endpoint:
TRANSPORT=http HTTP_PORT=3000 DEEPSEEK_API_KEY=your-key node dist/index.jsTest the health endpoint:
curl http://localhost:3000/healthThe MCP endpoint is available at POST /mcp (Streamable HTTP protocol).
Docker
# Build
docker build -t deepseek-mcp-server .
# Run
docker run -d -p 3000:3000 -e DEEPSEEK_API_KEY=your-key deepseek-mcp-server
# Or use docker-compose
DEEPSEEK_API_KEY=your-key docker compose up -dThe Docker image defaults to HTTP transport on port 3000 with a built-in health check.
Troubleshooting
"DEEPSEEK_API_KEY environment variable is not set"
Option 1: Use the correct installation command
# Make sure to include -e flag with your API key
claude mcp add deepseek npx @arikusi/deepseek-mcp-server -e DEEPSEEK_API_KEY=your-key-hereOption 2: Manually edit the config file
If you already installed without the API key, edit your config file:
For Claude Code: Open
~/.claude.json(Windows:C:\Users\USERNAME\.claude.json)Find the
"mcpServers"section under your project pathAdd the
envfield with your API key:
"deepseek": {
"type": "stdio",
"command": "npx",
"args": ["@arikusi/deepseek-mcp-server"],
"env": {
"DEEPSEEK_API_KEY": "your-api-key-here"
}
}Save and restart Claude Code
"Failed to connect to DeepSeek API"
Check your API key is valid
Verify you have internet connection
Check DeepSeek API status at https://status.deepseek.com
Server not appearing in your MCP client
Verify the path to
dist/index.jsis correctMake sure you ran
npm run buildCheck your MCP client's logs for errors
Restart your MCP client completely
Permission Denied on macOS/Linux
Make the file executable:
chmod +x dist/index.jsPublishing to npm
To share this MCP server with others:
Run
npm loginRun
npm publish --access public
Users can then install with:
npm install -g @arikusi/deepseek-mcp-serverContributing
Contributions are welcome! Please read our Contributing Guidelines before submitting PRs.
Reporting Issues
Found a bug or have a feature request? Please open an issue using our templates.
Development
# Clone the repo
git clone https://github.com/arikusi/deepseek-mcp-server.git
cd deepseek-mcp-server
# Install dependencies
npm install
# Build in watch mode
npm run watch
# Run tests
npm test
# Lint
npm run lintChangelog
See CHANGELOG.md for version history and updates.
License
MIT License - see LICENSE file for details
Support
Resources
DeepSeek Platform - Get your API key
Model Context Protocol - MCP specification
DeepSeek API Documentation - API reference
Acknowledgments
Built with Model Context Protocol SDK
Uses OpenAI SDK for API compatibility
Created for the MCP community
Made by
This is an unofficial community project and is not affiliated with DeepSeek.