/ Annuaire / Playground / buyer-eval-skill
● Communauté salespeak-ai ⚡ Instantané

buyer-eval-skill

par salespeak-ai · salespeak-ai/buyer-eval-skill

B2B vendor evaluation skill — domain-expert questions, evidence-based scoring, and structured vendor interviews.

buyer-eval-skill makes Claude your B2B procurement lead: it asks domain-expert questions of the vendor (or their AI agent), records evidence against your evaluation rubric, scores each criterion, and produces a comparison matrix when you're evaluating multiple vendors. Reduces the time from 'let's look at vendors' to 'here's my recommendation with receipts'.

Pourquoi l'utiliser

Fonctionnalités clés

Démo en direct

Aperçu en pratique

buyer-eval-skill.replay ▶ prêt
0/0

Installer

Choisissez votre client

~/Library/Application Support/Claude/claude_desktop_config.json  · Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "buyer-eval-skill": {
      "command": "git",
      "args": [
        "clone",
        "https://github.com/salespeak-ai/buyer-eval-skill",
        "~/.claude/skills/buyer-eval-skill"
      ],
      "_inferred": true
    }
  }
}

Ouvrez Claude Desktop → Settings → Developer → Edit Config. Redémarrez après avoir enregistré.

~/.cursor/mcp.json · .cursor/mcp.json
{
  "mcpServers": {
    "buyer-eval-skill": {
      "command": "git",
      "args": [
        "clone",
        "https://github.com/salespeak-ai/buyer-eval-skill",
        "~/.claude/skills/buyer-eval-skill"
      ],
      "_inferred": true
    }
  }
}

Cursor utilise le même schéma mcpServers que Claude Desktop. La config projet l'emporte sur la globale.

VS Code → Cline → MCP Servers → Edit
{
  "mcpServers": {
    "buyer-eval-skill": {
      "command": "git",
      "args": [
        "clone",
        "https://github.com/salespeak-ai/buyer-eval-skill",
        "~/.claude/skills/buyer-eval-skill"
      ],
      "_inferred": true
    }
  }
}

Cliquez sur l'icône MCP Servers dans la barre latérale Cline, puis "Edit Configuration".

~/.codeium/windsurf/mcp_config.json
{
  "mcpServers": {
    "buyer-eval-skill": {
      "command": "git",
      "args": [
        "clone",
        "https://github.com/salespeak-ai/buyer-eval-skill",
        "~/.claude/skills/buyer-eval-skill"
      ],
      "_inferred": true
    }
  }
}

Même format que Claude Desktop. Redémarrez Windsurf pour appliquer.

~/.continue/config.json
{
  "mcpServers": [
    {
      "name": "buyer-eval-skill",
      "command": "git",
      "args": [
        "clone",
        "https://github.com/salespeak-ai/buyer-eval-skill",
        "~/.claude/skills/buyer-eval-skill"
      ]
    }
  ]
}

Continue utilise un tableau d'objets serveur plutôt qu'une map.

~/.config/zed/settings.json
{
  "context_servers": {
    "buyer-eval-skill": {
      "command": {
        "path": "git",
        "args": [
          "clone",
          "https://github.com/salespeak-ai/buyer-eval-skill",
          "~/.claude/skills/buyer-eval-skill"
        ]
      }
    }
  }
}

Ajoutez dans context_servers. Zed recharge à chaud à la sauvegarde.

claude mcp add buyer-eval-skill -- git clone https://github.com/salespeak-ai/buyer-eval-skill ~/.claude/skills/buyer-eval-skill

Une seule ligne. Vérifiez avec claude mcp list. Supprimez avec claude mcp remove.

Cas d'usage

Usages concrets : buyer-eval-skill

How to evaluate three SaaS vendors against your requirements

👤 Buyers, RevOps, procurement leads ⏱ ~120 min intermediate

Quand l'utiliser : You've shortlisted 3 vendors and need to pick one defensibly.

Prérequis
  • Skill cloned — git clone https://github.com/salespeak-ai/buyer-eval-skill ~/.claude/skills/buyer-eval-skill
Déroulement
  1. Define the rubric
    We're buying a customer support platform. Propose a weighted rubric (feature depth, integrations, price, security, support).✓ Copié
    → Weighted criteria with rationale
  2. Interview each vendor
    Use the skill's question bank to interview vendor A. Their docs / AI agent is at <url>. Record evidence per criterion.✓ Copié
    → Per-criterion evidence snippets with sources
  3. Score and compare
    Score all three and produce a comparison matrix with a recommendation.✓ Copié
    → Matrix + recommendation + named risks

Résultat : A defensible vendor pick with evidence, not vibes.

Pièges
  • Rubric favors the vendor you already like — Lock the rubric before seeing demos
  • Evidence is just vendor marketing claims — Weight third-party sources (docs, G2, case studies) over marketing
Combiner avec : filesystem

Compare RFP responses against your requirements

👤 Procurement teams drowning in RFP responses ⏱ ~90 min intermediate

Quand l'utiliser : You sent an RFP, got 5 responses, and need to rank them.

Déroulement
  1. Load the responses
    Here are 5 RFP response PDFs in rfp/. Extract answers per requirement.✓ Copié
    → Per-requirement matrix
  2. Flag gaps and fluff
    Where did a vendor answer the wrong question or punt? Flag fluff.✓ Copié
    → Honest read of each response
  3. Rank
    Rank the responses and write a short memo for the evaluation committee.✓ Copié
    → Ranked memo

Résultat : A fair, fast RFP review.

Pièges
  • Rewarding the best-written response over the best fit — Score on substance; penalize fluff

Combinaisons

Associez-le à d'autres MCPs pour un effet X10

buyer-eval-skill + filesystem

Read PDFs/docs from a local folder during evaluation

Read rfp/responses/ and extract answers per requirement.✓ Copié
buyer-eval-skill + google-ai-mode-skill

Cross-check vendor claims against public info

For each vendor claim that isn't in their docs, run a quick web check.✓ Copié

Outils

Ce que ce MCP expose

OutilEntréesQuand appelerCoût
rubric-design domain + must-haves Start of any evaluation 0
vendor-interview vendor source (docs/AI agent) Per vendor 0
scoring-and-matrix evidence per vendor + rubric After interviews 0
recommendation-memo matrix Final step 0

Coût et limites

Coût d'exécution

Quota d'API
none beyond LLM
Tokens par appel
10–40k per full eval
Monétaire
free
Astuce
Interview one vendor at a time — keeps context focused and evidence clean

Sécurité

Permissions, secrets, portée

Stockage des identifiants : none
Sortie de données : only if the skill calls out to vendor agents you point it at

Dépannage

Erreurs courantes et correctifs

Evidence is shallow

Point Claude at primary docs; fallback to published case studies

Scores feel gamed

Re-lock the rubric and re-run scoring blind to vendor identity if possible

Alternatives

buyer-eval-skill vs autres

AlternativeQuand l'utiliserCompromis
creative-director-skillEvaluating creative concepts, not B2B softwareDifferent judgment domain

Plus

Ressources

📖 Lire le README officiel sur GitHub

🐙 Voir les issues ouvertes

🔍 Parcourir les 400+ serveurs MCP et Skills