/ Annuaire / Playground / volcano-agent-sdk
● Communauté Kong ⚡ Instantané

volcano-agent-sdk

par Kong · Kong/volcano-agent-sdk

Build TypeScript AI agents that chain LLM reasoning with MCP tools — 100+ models, parallel execution, built-in OTel tracing.

Volcano Agent SDK by Kong is a TypeScript SDK (not an MCP server) for building multi-provider AI agents that consume MCP tools. Supports OpenAI, Anthropic, Mistral, Bedrock, Vertex, Azure. Autoselects tools from configured MCP endpoints, streams tokens, retries on failure, ships OpenTelemetry traces.

Pourquoi l'utiliser

Fonctionnalités clés

Démo en direct

Aperçu en pratique

volcano-agent-sdk.replay ▶ prêt
0/0

Installer

Choisissez votre client

~/Library/Application Support/Claude/claude_desktop_config.json  · Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "volcano-agent-sdk": {
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ],
      "_inferred": true
    }
  }
}

Ouvrez Claude Desktop → Settings → Developer → Edit Config. Redémarrez après avoir enregistré.

~/.cursor/mcp.json · .cursor/mcp.json
{
  "mcpServers": {
    "volcano-agent-sdk": {
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ],
      "_inferred": true
    }
  }
}

Cursor utilise le même schéma mcpServers que Claude Desktop. La config projet l'emporte sur la globale.

VS Code → Cline → MCP Servers → Edit
{
  "mcpServers": {
    "volcano-agent-sdk": {
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ],
      "_inferred": true
    }
  }
}

Cliquez sur l'icône MCP Servers dans la barre latérale Cline, puis "Edit Configuration".

~/.codeium/windsurf/mcp_config.json
{
  "mcpServers": {
    "volcano-agent-sdk": {
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ],
      "_inferred": true
    }
  }
}

Même format que Claude Desktop. Redémarrez Windsurf pour appliquer.

~/.continue/config.json
{
  "mcpServers": [
    {
      "name": "volcano-agent-sdk",
      "command": "npx",
      "args": [
        "-y",
        "volcano-agent-sdk"
      ]
    }
  ]
}

Continue utilise un tableau d'objets serveur plutôt qu'une map.

~/.config/zed/settings.json
{
  "context_servers": {
    "volcano-agent-sdk": {
      "command": {
        "path": "npx",
        "args": [
          "-y",
          "volcano-agent-sdk"
        ]
      }
    }
  }
}

Ajoutez dans context_servers. Zed recharge à chaud à la sauvegarde.

claude mcp add volcano-agent-sdk -- npx -y volcano-agent-sdk

Une seule ligne. Vérifiez avec claude mcp list. Supprimez avec claude mcp remove.

Cas d'usage

Usages concrets : volcano-agent-sdk

Build a coding agent that uses GitHub + Sentry MCPs

👤 TS devs building internal automation ⏱ ~60 min advanced

Quand l'utiliser : You want a programmable agent, not a chat session.

Prérequis
  • Node 20+ — Standard
  • MCP endpoints for the tools you want (github, sentry, etc.) — Either existing public or your own deploys
Déroulement
  1. Install + scaffold
    npm install @volcano.dev/agent and write a minimal agent that connects to github + sentry MCPs with Anthropic as the model.✓ Copié
    → Running TS project
  2. Write the task
    The agent's task: every 15 min, find new Sentry errors, correlate to GitHub commits via MCP, draft revert PRs for the obvious ones.✓ Copié
    → Agent executes task autonomously
  3. Instrument
    Enable OTel traces and pipe to Honeycomb/Grafana.✓ Copié
    → Spans visible

Résultat : A production-ready automation agent with observability.

Pièges
  • Agent hallucinates tool calls that don't exist — Restrict the MCP set passed to the agent; fewer, well-documented tools > more
  • Retries amplify transient upstream issues — Tune retry policy and add exponential backoff
Combiner avec : github · sentry

Compose a multi-agent crew for research tasks

👤 Developers exploring agent patterns ⏱ ~45 min advanced

Quand l'utiliser : A task benefits from specialization (researcher + writer + reviewer).

Déroulement
  1. Define agents
    Create agents: Researcher (web search MCP), Writer (drafts), Reviewer (fact-checks Researcher's sources).✓ Copié
    → Three typed agent instances
  2. Delegate
    Run the crew: topic 'state of MCP in 2026'. Have Researcher gather, Writer draft, Reviewer verify claims.✓ Copié
    → Coordinated output

Résultat : Higher-quality output than a single-pass agent on complex tasks.

Combiner avec : omnisearch

Build a streaming chatbot with tool access

👤 Product devs integrating AI into a TS app ⏱ ~40 min advanced

Quand l'utiliser : User-facing feature that needs real-time streaming + MCP tool calls.

Déroulement
  1. Wire streaming
    Build a chat endpoint that streams tokens to the client and calls tools mid-stream as the model requests.✓ Copié
    → Working streaming endpoint
  2. Explainability
    After each response, expose agent.summary() so the UI can show which tools were used.✓ Copié
    → Tool trail visible

Résultat : A production chat UI with transparent tool use.

Combinaisons

Associez-le à d'autres MCPs pour un effet X10

volcano-agent-sdk + github + sentry

Autonomous triage agent across incident + code

Build an agent that, given a Sentry alert, fetches the stack, finds the offending commit via GitHub, and opens a PR with a minimal fix.✓ Copié
volcano-agent-sdk + vurb-ts

Vurb builds the server side; Volcano builds the agent side

Expose my business data via a Vurb MCP; build a Volcano agent that uses it to answer user questions.✓ Copié

Outils

Ce que ce MCP expose

OutilEntréesQuand appelerCoût
(SDK) You write TS; SDK picks tools from configured MCPs automatically N/A — Volcano Agent SDK is a library you build with, not an MCP server you call n/a — depends on model + tools

Coût et limites

Coût d'exécution

Quota d'API
LLM provider limits apply; MCP upstream limits apply
Tokens par appel
Depends on model + conversation length
Monétaire
SDK free; LLM usage billed by provider
Astuce
Use cheaper models (Haiku/GPT-4o-mini) for routing and reserve expensive models for reasoning; Volcano supports per-step model choice.

Sécurité

Permissions, secrets, portée

Stockage des identifiants : LLM provider keys + MCP credentials in env; SDK injects them
Sortie de données : LLM provider APIs + configured MCP endpoints
Ne jamais accorder : LLM keys to untrusted code paths in the same process

Dépannage

Erreurs courantes et correctifs

MCP connection fails on startup

Verify the MCP endpoint URL and auth. SDK logs full error when --debug is set.

Vérifier : Curl the MCP endpoint directly
Model refuses to use available tools

Tool descriptions may be poorly phrased; rewrite for clarity or force via agent config.

Vérifier : Inspect tools via agent.listTools()
High token cost on simple tasks

Check that system prompt isn't dragging MCP tool defs into every call; use lazy tool-load mode.

Vérifier : agent.summary() shows token breakdown

Alternatives

volcano-agent-sdk vs autres

AlternativeQuand l'utiliserCompromis
LangChain / LangGraph (TS)You want the largest ecosystem of integrationsHeavier abstraction; slower cold path
Vercel AI SDKYou want tight Next.js integrationLess focus on multi-agent patterns
Anthropic SDK rawYou only need Anthropic and minimal abstractionYou reimplement tool routing, retries, multi-provider

Plus

Ressources

📖 Lire le README officiel sur GitHub

🐙 Voir les issues ouvertes

🔍 Parcourir les 400+ serveurs MCP et Skills