/ 디렉터리 / 플레이그라운드 / SkillCompass
● 커뮤니티 Evol-ai ⚡ 바로 사용

SkillCompass

제작: Evol-ai · Evol-ai/SkillCompass

Evaluate Agent Skill quality — find the weakest link, fix it, and prove the fix worked with before/after metrics.

SkillCompass scores your Agent Skills on clarity, activation rate, downstream correctness, and context cost. It highlights the skill most likely to be hurting your agent's performance, suggests a fix, and re-runs the evaluation so you can show the improvement. Useful when you have a shelf of skills and don't know which are actually earning their context weight.

왜 쓰나요

핵심 기능

라이브 데모

실제 사용 모습

skillcompass-skill.replay ▶ 준비됨
0/0

설치

클라이언트 선택

~/Library/Application Support/Claude/claude_desktop_config.json  · Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "skillcompass-skill": {
      "command": "git",
      "args": [
        "clone",
        "https://github.com/Evol-ai/SkillCompass",
        "~/.claude/skills/SkillCompass"
      ],
      "_inferred": true
    }
  }
}

Claude Desktop → Settings → Developer → Edit Config 열기. 저장 후 앱 재시작.

~/.cursor/mcp.json · .cursor/mcp.json
{
  "mcpServers": {
    "skillcompass-skill": {
      "command": "git",
      "args": [
        "clone",
        "https://github.com/Evol-ai/SkillCompass",
        "~/.claude/skills/SkillCompass"
      ],
      "_inferred": true
    }
  }
}

Cursor는 Claude Desktop과 동일한 mcpServers 스키마 사용. 프로젝트 설정이 전역보다 우선.

VS Code → Cline → MCP Servers → Edit
{
  "mcpServers": {
    "skillcompass-skill": {
      "command": "git",
      "args": [
        "clone",
        "https://github.com/Evol-ai/SkillCompass",
        "~/.claude/skills/SkillCompass"
      ],
      "_inferred": true
    }
  }
}

Cline 사이드바의 MCP Servers 아이콘 클릭 후 "Edit Configuration" 선택.

~/.codeium/windsurf/mcp_config.json
{
  "mcpServers": {
    "skillcompass-skill": {
      "command": "git",
      "args": [
        "clone",
        "https://github.com/Evol-ai/SkillCompass",
        "~/.claude/skills/SkillCompass"
      ],
      "_inferred": true
    }
  }
}

Claude Desktop과 같은 형식. Windsurf 재시작 후 적용.

~/.continue/config.json
{
  "mcpServers": [
    {
      "name": "skillcompass-skill",
      "command": "git",
      "args": [
        "clone",
        "https://github.com/Evol-ai/SkillCompass",
        "~/.claude/skills/SkillCompass"
      ]
    }
  ]
}

Continue는 맵이 아닌 서버 오브젝트 배열 사용.

~/.config/zed/settings.json
{
  "context_servers": {
    "skillcompass-skill": {
      "command": {
        "path": "git",
        "args": [
          "clone",
          "https://github.com/Evol-ai/SkillCompass",
          "~/.claude/skills/SkillCompass"
        ]
      }
    }
  }
}

context_servers에 추가. 저장 시 Zed가 핫 리로드.

claude mcp add skillcompass-skill -- git clone https://github.com/Evol-ai/SkillCompass ~/.claude/skills/SkillCompass

한 줄 명령. claude mcp list로 확인, claude mcp remove로 제거.

사용 사례

실전 활용법: SkillCompass

Find the skill pulling your agent's performance down

👤 Skill authors with 5+ skills installed ⏱ ~45 min advanced

언제 쓸까: You feel the agent has gotten worse, not better, as you added skills.

사전 조건
  • Node 20+ — nvm install 20
  • Skill cloned and installed — git clone https://github.com/Evol-ai/SkillCompass ~/.claude/skills/SkillCompass; npm i
흐름
  1. Run the evaluator
    Score all skills in ~/.claude/skills/ — show me the weakest link.✓ 복사됨
    → Ranked skill list with per-dimension scores
  2. Diagnose the loser
    For the weakest skill, what specifically is wrong?✓ 복사됨
    → Concrete critique (vague description, conflicting with other skill, etc.)
  3. Propose a fix
    Suggest a minimal edit to SKILL.md to fix it.✓ 복사됨
    → Small, reviewable diff
  4. Re-evaluate
    Re-run the eval and show before/after.✓ 복사됨
    → Metrics improved, with evidence

결과: A measurably better skill bundle, with a reproducible eval process.

함정
  • Gaming the eval metric instead of helping real tasks — Include task-level downstream metrics (actual agent outcomes), not just text-level

Review a new skill before you publish it

👤 Skill authors shipping their first bundle ⏱ ~20 min intermediate

언제 쓸까: Before pushing to GitHub and telling the world about your skill.

흐름
  1. Score the draft
    Evaluate my draft skill at ./my-skill/.✓ 복사됨
    → Dimension scores
  2. Fix obvious issues
    Apply the low-hanging suggestions✓ 복사됨
    → Edits in SKILL.md

결과: A publication-ready skill rather than a rough draft.

함정
  • Chasing a perfect score — Ship when scores plateau — diminishing returns

조합

다른 MCP와 조합해 10배 효율

skillcompass-skill + skill-optimizer-skill

Two complementary tools: SkillCompass ranks, skill-optimizer drills into SKILL.md patterns

Use SkillCompass to pick the worst skill; use skill-optimizer to deeply analyze its SKILL.md.✓ 복사됨
skillcompass-skill + filesystem

Operate across the full ~/.claude/skills/ directory

Evaluate every skill in ~/.claude/skills/ and give me a CSV.✓ 복사됨

도구

이 MCP가 노출하는 것

도구입력언제 호출비용
skill-scoring skill path(s) Periodic audits eval compute
weakest-link-id bundle scores After scoring 0
fix-suggestion weak skill + critique Before editing 0
before-after-eval pre/post SKILL.md After applying fixes eval compute

비용 및 제한

운영 비용

API 쿼터
none beyond your LLM provider (evals use LLM calls)
호출당 토큰
evals can be heavy — budget 20–100k tokens for a full bundle scan
금액
free, MIT
Run on one skill at a time during iteration; bundle runs only for audits

보안

권한, 시크릿, 파급범위

자격 증명 저장: none at skill level
데이터 외부 송신: none beyond your LLM provider

문제 해결

자주 발생하는 오류와 해결

Node errors on install

Ensure Node 20+; npm i inside the skill directory.

확인: node -v
Evals are inconsistent run-to-run

Fix the task seed and use a non-stochastic sample; record provider+model.

대안

SkillCompass 다른 것과 비교

대안언제 쓰나단점/장점
skill-optimizer-skillYou want a single skill analyzed deeply rather than a bundle rankedDepth over breadth
manual reviewYou have 1–2 skills totalDoesn't scale

더 보기

리소스

📖 GitHub에서 공식 README 읽기

🐙 열린 이슈 보기

🔍 400+ MCP 서버 및 Skills 전체 보기