/ Directory / Playground / deepseek-mcp-server
● Community DMontgomery40 ⚡ Instant

deepseek-mcp-server

by DMontgomery40 · DMontgomery40/deepseek-mcp-server

Delegate specific tasks to DeepSeek models from inside Claude — cheaper reasoning via R1, or code execution via DeepSeek's code-exec mode.

deepseek-mcp-server exposes DeepSeek's chat/completions/models/balance endpoints as MCP tools. Supports a novel code-execution mode where DeepSeek writes a short program that calls tools directly in a runtime, reducing back-and-forth latency and context pressure.

Why use it

Key features

Live Demo

What it looks like in practice

deepseek.replay ▶ ready
0/0

Install

Pick your client

~/Library/Application Support/Claude/claude_desktop_config.json  · Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "deepseek": {
      "command": "npx",
      "args": [
        "-y",
        "deepseek-mcp-server"
      ],
      "_inferred": true
    }
  }
}

Open Claude Desktop → Settings → Developer → Edit Config. Restart after saving.

~/.cursor/mcp.json · .cursor/mcp.json
{
  "mcpServers": {
    "deepseek": {
      "command": "npx",
      "args": [
        "-y",
        "deepseek-mcp-server"
      ],
      "_inferred": true
    }
  }
}

Cursor uses the same mcpServers schema as Claude Desktop. Project config wins over global.

VS Code → Cline → MCP Servers → Edit
{
  "mcpServers": {
    "deepseek": {
      "command": "npx",
      "args": [
        "-y",
        "deepseek-mcp-server"
      ],
      "_inferred": true
    }
  }
}

Click the MCP Servers icon in the Cline sidebar, then "Edit Configuration".

~/.codeium/windsurf/mcp_config.json
{
  "mcpServers": {
    "deepseek": {
      "command": "npx",
      "args": [
        "-y",
        "deepseek-mcp-server"
      ],
      "_inferred": true
    }
  }
}

Same shape as Claude Desktop. Restart Windsurf to pick up changes.

~/.continue/config.json
{
  "mcpServers": [
    {
      "name": "deepseek",
      "command": "npx",
      "args": [
        "-y",
        "deepseek-mcp-server"
      ]
    }
  ]
}

Continue uses an array of server objects rather than a map.

~/.config/zed/settings.json
{
  "context_servers": {
    "deepseek": {
      "command": {
        "path": "npx",
        "args": [
          "-y",
          "deepseek-mcp-server"
        ]
      }
    }
  }
}

Add to context_servers. Zed hot-reloads on save.

claude mcp add deepseek -- npx -y deepseek-mcp-server

One-liner. Verify with claude mcp list. Remove with claude mcp remove.

Use Cases

Real-world ways to use deepseek-mcp-server

Delegate cheap-but-long reasoning to DeepSeek from inside Claude

👤 Cost-conscious devs building agent flows ⏱ ~10 min intermediate

When to use: You have a step that needs chain-of-thought but not Claude-level quality — save 90% on tokens.

Prerequisites
  • DeepSeek API key — platform.deepseek.com — cheap
Flow
  1. Route specific subtask
    Use deepseek chat (model deepseek-reasoner) to classify these 100 support tickets into 5 buckets. Return only the classifications.✓ Copied
    → Classifications completed cheaply

Outcome: Same quality for classification/extraction at a fraction of the cost.

Pitfalls
  • Quality gap on nuanced writing tasks — Profile first — DeepSeek shines on structured tasks, trails on creative/long-form

Let DeepSeek write a short script to call multiple tools in one go

👤 Agent builders tired of 10-round tool call loops ⏱ ~20 min advanced

When to use: A task needs 8 tool calls in sequence; going round-trip each time is slow and burns tokens.

Flow
  1. Enable code-exec mode
    Using deepseek code-exec mode: fetch 5 URLs with fetch MCP, summarize each, then combine into a table. One shot.✓ Copied
    → Script runs inside exec runtime; returns table

Outcome: Fewer tokens, fewer round-trips for pipeline-style tasks.

Pitfalls
  • Script has bugs in the execution runtime — Fall back to per-tool calls when debugging
  • Security — exec runtime runs LLM-authored code — Keep tool access narrow; don't grant filesystem/shell unless essential
Combine with: fetch · firecrawl

Monitor DeepSeek API spend from your agent

👤 Ops people running DeepSeek-backed agents ⏱ ~5 min beginner

When to use: You want an agent to halt when spend is low or to alert weekly.

Flow
  1. Pull balance
    What's my DeepSeek balance?✓ Copied
    → Current balance
  2. Alert if low
    If balance < $5, post a Slack message #ops-alerts with the number.✓ Copied
    → Conditional alert

Outcome: Never run out of credits mid-pipeline.

Combinations

Pair with other MCPs for X10 leverage

deepseek + fetch

Scrape then classify with DeepSeek cheaply

Fetch 20 news URLs, then deepseek-classify each by topic (5 buckets).✓ Copied

Tools

What this MCP exposes

ToolInputsWhen to callCost
chat messages: Msg[], model?: str, stream?: bool Delegate a reasoning or classification task DeepSeek tokens
completions prompt, model?, max_tokens? Non-chat completion calls DeepSeek tokens
models List available DeepSeek models free
balance Check spend / alert on low credit free

Cost & Limits

What this costs to run

API quota
Per DeepSeek account
Tokens per call
DeepSeek is 10-50x cheaper than frontier models for many tasks
Monetary
DeepSeek V3 ~$0.27/M input, R1 ~$0.55/M input (as of 2026)
Tip
Use the balance tool to hard-stop agents when credits drop below your threshold.

Security

Permissions, secrets, blast radius

Credential storage: DEEPSEEK_API_KEY for local; DEEPSEEK_MCP_AUTH_TOKEN for remote endpoint
Data egress: Your prompts go to api.deepseek.com (China-based)
Never grant: Wide tool surface to the code-exec mode without reviewing which tools are exposed

Troubleshooting

Common errors and fixes

401 from DeepSeek

Check API key. Also verify account hasn't been suspended.

Verify: curl -H 'Authorization: Bearer $KEY' https://api.deepseek.com/v1/models
Rate limit

DeepSeek has tight per-minute limits on new accounts. Add retry + backoff.

Code-exec mode returns 'tool not available'

Code-exec sees only the tool set you exposed. Add the needed tool to its allowlist.

Alternatives

deepseek-mcp-server vs others

AlternativeWhen to use it insteadTradeoff
OpenAI / Anthropic via direct APIYou want frontier quality and are not cost-bound5-50x the cost
Groq MCPYou want ultra-low latency on open modelsDifferent model lineup; no code-exec mode

More

Resources

📖 Read the official README on GitHub

🐙 Browse open issues

🔍 Browse all 400+ MCP servers and Skills