/ Verzeichnis / Playground / Bright Data
● Offiziell brightdata 🔑 Eigener Schlüssel nötig

Bright Data

von brightdata · brightdata/brightdata-mcp

Search, scrape, and unblock the web at scale — residential proxies + SERP API + browser automation in one MCP.

Bright Data's official MCP bundles three capabilities: live SERP search results across Google/Bing/DuckDuckGo, scraping of single or bulk URLs through their unblocker/proxy network, and a fleet of pre-built structured scrapers for specific targets (Amazon, LinkedIn public, Instagram public, Zillow, etc.). Credits are metered; use sampling and caching.

Warum nutzen

Hauptfunktionen

Live-Demo

In der Praxis

brightdata.replay ▶ bereit
0/0

Installieren

Wählen Sie Ihren Client

~/Library/Application Support/Claude/claude_desktop_config.json  · Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "brightdata": {
      "command": "npx",
      "args": [
        "-y",
        "@brightdata/mcp"
      ]
    }
  }
}

Öffne Claude Desktop → Settings → Developer → Edit Config. Nach dem Speichern neu starten.

~/.cursor/mcp.json · .cursor/mcp.json
{
  "mcpServers": {
    "brightdata": {
      "command": "npx",
      "args": [
        "-y",
        "@brightdata/mcp"
      ]
    }
  }
}

Cursor nutzt das gleiche mcpServers-Schema wie Claude Desktop. Projektkonfiguration schlägt die globale.

VS Code → Cline → MCP Servers → Edit
{
  "mcpServers": {
    "brightdata": {
      "command": "npx",
      "args": [
        "-y",
        "@brightdata/mcp"
      ]
    }
  }
}

Klicken Sie auf das MCP-Servers-Symbol in der Cline-Seitenleiste, dann "Edit Configuration".

~/.codeium/windsurf/mcp_config.json
{
  "mcpServers": {
    "brightdata": {
      "command": "npx",
      "args": [
        "-y",
        "@brightdata/mcp"
      ]
    }
  }
}

Gleiche Struktur wie Claude Desktop. Windsurf neu starten zum Übernehmen.

~/.continue/config.json
{
  "mcpServers": [
    {
      "name": "brightdata",
      "command": "npx",
      "args": [
        "-y",
        "@brightdata/mcp"
      ]
    }
  ]
}

Continue nutzt ein Array von Serverobjekten statt einer Map.

~/.config/zed/settings.json
{
  "context_servers": {
    "brightdata": {
      "command": {
        "path": "npx",
        "args": [
          "-y",
          "@brightdata/mcp"
        ]
      }
    }
  }
}

In context_servers hinzufügen. Zed lädt beim Speichern neu.

claude mcp add brightdata -- npx -y @brightdata/mcp

Einzeiler. Prüfen mit claude mcp list. Entfernen mit claude mcp remove.

Anwendungsfälle

Praxisnahe Nutzung: Bright Data

Track your keyword rankings on Google across locations

👤 SEO teams ⏱ ~20 min intermediate

Wann einsetzen: You want daily rank tracking for 50 keywords in US/UK/DE without running your own proxies.

Voraussetzungen
  • Bright Data API token — brightdata.com → dashboard → API tokens
  • Budget: ~$0.001–$0.003 per SERP query — Credit balance on Bright Data account
Ablauf
  1. Run the SERP for each kw/country
    For each keyword in [list], run a Google SERP search from country=us. Capture top 10 organic results (url, title, position).✓ Kopiert
    → Per-keyword ranked list
  2. Locate our domain
    For each result set, find where mydomain.com appears (or 'not in top 10'). Output kw → position.✓ Kopiert
    → Rank table
  3. Diff vs yesterday
    Compare to yesterday's JSON [paste]. Highlight moves > 3 positions.✓ Kopiert
    → Daily movers report

Ergebnis: A daily rank-tracking workflow at ~$0.15/day for 50 keywords, no proxy ops.

Fallstricke
  • Each country/device combo counts as a separate query — Only track what you need; 50 kw × 3 countries × 7 days is 1050 queries/week
Kombinieren mit: postgres · notion

Fetch a page that blocks datacenter IPs

👤 Devs whose normal scraping target got behind Cloudflare-Turnstile ⏱ ~15 min intermediate

Wann einsetzen: fetch / Firecrawl get 403 or an interstitial; you need residential IPs.

Ablauf
  1. Try once with unlocker
    Fetch <url> via Web Unlocker. Return the rendered HTML + HTTP status.✓ Kopiert
    → 200 + real HTML
  2. Extract what you need
    From that HTML, extract [list the fields]. Return as JSON.✓ Kopiert
    → Structured data
  3. Respect the site
    If the page says 'robots.txt disallow' or a clear anti-scrape notice, abort and tell me.✓ Kopiert
    → Consent-aware fallback

Ergebnis: The data you need without maintaining a proxy pool.

Fallstricke
  • Unblocker can still fail on hardcore targets (banking, SaaS login pages) — These are intentionally private; pick an official API or a different approach
  • Costs escalate quickly on large crawls — Firecrawl or fetch is cheaper for unprotected sites — only pay Bright Data when you hit a block
Kombinieren mit: firecrawl

Pull a structured Amazon product dataset via prebuilt scraper

👤 E-commerce analysts ⏱ ~20 min intermediate

Wann einsetzen: You want 500 Amazon ASINs with clean title/price/rating/bsr fields, not raw HTML.

Voraussetzungen
  • List of ASINs or category URLs — CSV or text input
Ablauf
  1. Kick off the prebuilt Amazon scraper
    Run the Bright Data Amazon product scraper for ASINs [list]. Return a job id.✓ Kopiert
    → Job id issued
  2. Poll until ready
    Poll the job. When done, fetch the dataset.✓ Kopiert
    → Full dataset delivered
  3. Cache to avoid re-runs
    Save the dataset to /data/amazon-<date>.jsonl. Flag any ASIN that errored.✓ Kopiert
    → Persisted dataset + error list

Ergebnis: A clean, re-runnable Amazon product dataset at ~$X/1000 products (see current pricing).

Fallstricke
  • Public LinkedIn/Instagram scrapers legally vary by region — Stay within public profile data; do not bypass authentication — know your jurisdiction
Kombinieren mit: postgres · filesystem

Daily news search for a brand across SERP

👤 PR / comms ⏱ ~15 min beginner

Wann einsetzen: You want a daily digest of what's being said about your brand in news search.

Ablauf
  1. Run a Google News SERP
    Google News search for '<brand>' last 24h, country=us. Return top 20 results with source, title, url, snippet.✓ Kopiert
    → News SERP
  2. Classify sentiment from snippets
    Score each result as positive/neutral/negative based on title+snippet. Flag anything flagged as negative for review.✓ Kopiert
    → Scored list
  3. Deliver the digest
    Format as a markdown digest: counts by sentiment, negative items with links, top positive items.✓ Kopiert
    → Digest ready

Ergebnis: A focused PR digest without scraping individual news sites.

Fallstricke
  • Sentiment from headlines alone is noisy — Only flag as negative if both title and snippet are clearly negative; human-review the flags
Kombinieren mit: notion

Kombinationen

Mit anderen MCPs für 10-fache Wirkung

brightdata + postgres

Store daily rank-tracking rows for trend analysis

Run SERP for kw list, INSERT into keyword_ranks table with today's date.✓ Kopiert
brightdata + firecrawl

Cheap-first, unblock-fallback crawling

Try Firecrawl first; if 403/blocked, fall back to Bright Data Unlocker for that URL only.✓ Kopiert
brightdata + notion

Weekly PR digest posted to Notion

Run brand SERP for the last 7 days, create a Notion page with the digest.✓ Kopiert

Werkzeuge

Was dieses MCP bereitstellt

WerkzeugEingabenWann aufrufenKosten
search_engine engine: 'google'|'bing'|'duckduckgo', query, country?, lang?, device? SERP / rank-tracking workflows ~$0.001–0.003 per query
scrape_as_markdown url Fetch a single page through unblocker as clean markdown 1 credit per page
scrape_as_html url You need raw HTML to parse yourself 1 credit per page
web_data_<target> urls: str[] or params Prebuilt structured scraper (amazon, linkedin, zillow, etc.) per-scraper pricing
scraping_browser_* url, actions Multi-step / JS-heavy flows browser-session pricing

Kosten & Limits

Was der Betrieb kostet

API-Kontingent
Bounded by account credits; concurrent requests per plan
Tokens pro Aufruf
SERP: 500–2000 tokens. Scraped page: 1000–5000 tokens.
Kosten in €
Pay-as-you-go; typical SERP $0.001–$0.003, unblocker ~$3 per 1000 pages, prebuilt scrapers priced per 1000 records.
Tipp
Cache aggressively — most data doesn't change hourly. Use cheaper fetch/Firecrawl for unprotected targets.

Sicherheit

Rechte, Secrets, Reichweite

Minimale Scopes: API token with zone access for the relevant products
Credential-Speicherung: BRIGHTDATA_API_TOKEN in env
Datenabfluss: All requests through Bright Data's proxy network; they see target URLs and responses
Niemals gewähren: Admin-level account tokens for everyday scraping

Fehlerbehebung

Häufige Fehler und Lösungen

401 Invalid token

BRIGHTDATA_API_TOKEN missing/expired. Regenerate in dashboard.

Prüfen: curl -H 'Authorization: Bearer $BRIGHTDATA_API_TOKEN' https://api.brightdata.com/zone/list
402 Insufficient credits

Top up account balance or reduce query volume; check dashboard for burn rate.

Scraping job SUCCEEDED but dataset empty

Wrong input schema for the prebuilt scraper. Read the scraper's doc page for required fields.

Target site still blocks despite unlocker

Some sites use advanced fingerprinting; switch to Scraping Browser with stealth, or abandon the target.

Alternativen

Bright Data vs. andere

AlternativeWann stattdessenKompromiss
Firecrawl MCPUnprotected sites, generic scrapingFails on hostile targets
Apify MCPYou want a broader Actor marketplace and cheaper pricing for common targetsProxy network quality varies per Actor
SerpAPI MCPYou only need SERP, not full scrapingNo unblocker / prebuilt scrapers

Mehr

Ressourcen

📖 Offizielle README auf GitHub lesen

🐙 Offene Issues ansehen

🔍 Alle 400+ MCP-Server und Skills durchsuchen