/ Verzeichnis / Playground / volcano-agent-sdk
● Community Kong ⚡ Sofort

volcano-agent-sdk

von Kong · Kong/volcano-agent-sdk

Build TypeScript AI agents that chain LLM reasoning with MCP tools — 100+ models, parallel execution, built-in OTel tracing.

Volcano Agent SDK by Kong is a TypeScript SDK (not an MCP server) for building multi-provider AI agents that consume MCP tools. Supports OpenAI, Anthropic, Mistral, Bedrock, Vertex, Azure. Autoselects tools from configured MCP endpoints, streams tokens, retries on failure, ships OpenTelemetry traces.

Warum nutzen

Hauptfunktionen

Live-Demo

In der Praxis

volcano-agent-sdk.replay ▶ bereit
0/0

Installieren

Wählen Sie Ihren Client

~/Library/Application Support/Claude/claude_desktop_config.json  · Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "volcano-agent-sdk": {
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ],
      "_inferred": true
    }
  }
}

Öffne Claude Desktop → Settings → Developer → Edit Config. Nach dem Speichern neu starten.

~/.cursor/mcp.json · .cursor/mcp.json
{
  "mcpServers": {
    "volcano-agent-sdk": {
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ],
      "_inferred": true
    }
  }
}

Cursor nutzt das gleiche mcpServers-Schema wie Claude Desktop. Projektkonfiguration schlägt die globale.

VS Code → Cline → MCP Servers → Edit
{
  "mcpServers": {
    "volcano-agent-sdk": {
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ],
      "_inferred": true
    }
  }
}

Klicken Sie auf das MCP-Servers-Symbol in der Cline-Seitenleiste, dann "Edit Configuration".

~/.codeium/windsurf/mcp_config.json
{
  "mcpServers": {
    "volcano-agent-sdk": {
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ],
      "_inferred": true
    }
  }
}

Gleiche Struktur wie Claude Desktop. Windsurf neu starten zum Übernehmen.

~/.continue/config.json
{
  "mcpServers": [
    {
      "name": "volcano-agent-sdk",
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ]
    }
  ]
}

Continue nutzt ein Array von Serverobjekten statt einer Map.

~/.config/zed/settings.json
{
  "context_servers": {
    "volcano-agent-sdk": {
      "command": {
        "path": "npx",
        "args": [
          "-y",
          "volcano-agent-sdk"
        ]
      }
    }
  }
}

In context_servers hinzufügen. Zed lädt beim Speichern neu.

claude mcp add volcano-agent-sdk -- npx -y volcano-agent-sdk

Einzeiler. Prüfen mit claude mcp list. Entfernen mit claude mcp remove.

Anwendungsfälle

Praxisnahe Nutzung: volcano-agent-sdk

Build a coding agent that uses GitHub + Sentry MCPs

👤 TS devs building internal automation ⏱ ~60 min advanced

Wann einsetzen: You want a programmable agent, not a chat session.

Voraussetzungen
  • Node 20+ — Standard
  • MCP endpoints for the tools you want (github, sentry, etc.) — Either existing public or your own deploys
Ablauf
  1. Install + scaffold
    npm install @volcano.dev/agent and write a minimal agent that connects to github + sentry MCPs with Anthropic as the model.✓ Kopiert
    → Running TS project
  2. Write the task
    The agent's task: every 15 min, find new Sentry errors, correlate to GitHub commits via MCP, draft revert PRs for the obvious ones.✓ Kopiert
    → Agent executes task autonomously
  3. Instrument
    Enable OTel traces and pipe to Honeycomb/Grafana.✓ Kopiert
    → Spans visible

Ergebnis: A production-ready automation agent with observability.

Fallstricke
  • Agent hallucinates tool calls that don't exist — Restrict the MCP set passed to the agent; fewer, well-documented tools > more
  • Retries amplify transient upstream issues — Tune retry policy and add exponential backoff
Kombinieren mit: github · sentry

Compose a multi-agent crew for research tasks

👤 Developers exploring agent patterns ⏱ ~45 min advanced

Wann einsetzen: A task benefits from specialization (researcher + writer + reviewer).

Ablauf
  1. Define agents
    Create agents: Researcher (web search MCP), Writer (drafts), Reviewer (fact-checks Researcher's sources).✓ Kopiert
    → Three typed agent instances
  2. Delegate
    Run the crew: topic 'state of MCP in 2026'. Have Researcher gather, Writer draft, Reviewer verify claims.✓ Kopiert
    → Coordinated output

Ergebnis: Higher-quality output than a single-pass agent on complex tasks.

Kombinieren mit: omnisearch

Build a streaming chatbot with tool access

👤 Product devs integrating AI into a TS app ⏱ ~40 min advanced

Wann einsetzen: User-facing feature that needs real-time streaming + MCP tool calls.

Ablauf
  1. Wire streaming
    Build a chat endpoint that streams tokens to the client and calls tools mid-stream as the model requests.✓ Kopiert
    → Working streaming endpoint
  2. Explainability
    After each response, expose agent.summary() so the UI can show which tools were used.✓ Kopiert
    → Tool trail visible

Ergebnis: A production chat UI with transparent tool use.

Kombinationen

Mit anderen MCPs für 10-fache Wirkung

volcano-agent-sdk + github + sentry

Autonomous triage agent across incident + code

Build an agent that, given a Sentry alert, fetches the stack, finds the offending commit via GitHub, and opens a PR with a minimal fix.✓ Kopiert
volcano-agent-sdk + vurb-ts

Vurb builds the server side; Volcano builds the agent side

Expose my business data via a Vurb MCP; build a Volcano agent that uses it to answer user questions.✓ Kopiert

Werkzeuge

Was dieses MCP bereitstellt

WerkzeugEingabenWann aufrufenKosten
(SDK) You write TS; SDK picks tools from configured MCPs automatically N/A — Volcano Agent SDK is a library you build with, not an MCP server you call n/a — depends on model + tools

Kosten & Limits

Was der Betrieb kostet

API-Kontingent
LLM provider limits apply; MCP upstream limits apply
Tokens pro Aufruf
Depends on model + conversation length
Kosten in €
SDK free; LLM usage billed by provider
Tipp
Use cheaper models (Haiku/GPT-4o-mini) for routing and reserve expensive models for reasoning; Volcano supports per-step model choice.

Sicherheit

Rechte, Secrets, Reichweite

Credential-Speicherung: LLM provider keys + MCP credentials in env; SDK injects them
Datenabfluss: LLM provider APIs + configured MCP endpoints
Niemals gewähren: LLM keys to untrusted code paths in the same process

Fehlerbehebung

Häufige Fehler und Lösungen

MCP connection fails on startup

Verify the MCP endpoint URL and auth. SDK logs full error when --debug is set.

Prüfen: Curl the MCP endpoint directly
Model refuses to use available tools

Tool descriptions may be poorly phrased; rewrite for clarity or force via agent config.

Prüfen: Inspect tools via agent.listTools()
High token cost on simple tasks

Check that system prompt isn't dragging MCP tool defs into every call; use lazy tool-load mode.

Prüfen: agent.summary() shows token breakdown

Alternativen

volcano-agent-sdk vs. andere

AlternativeWann stattdessenKompromiss
LangChain / LangGraph (TS)You want the largest ecosystem of integrationsHeavier abstraction; slower cold path
Vercel AI SDKYou want tight Next.js integrationLess focus on multi-agent patterns
Anthropic SDK rawYou only need Anthropic and minimal abstractionYou reimplement tool routing, retries, multi-provider

Mehr

Ressourcen

📖 Offizielle README auf GitHub lesen

🐙 Offene Issues ansehen

🔍 Alle 400+ MCP-Server und Skills durchsuchen