brightdata-web-mcp

34
0
Source

Search the web, scrape websites, extract structured data from URLs, and automate browsers using Bright Data's Web MCP. Use when fetching live web content, bypassing blocks/CAPTCHAs, getting product data from Amazon/eBay, social media posts, or when standard requests fail.

Install

mkdir -p .claude/skills/brightdata-web-mcp && curl -L -o skill.zip "https://mcp.directory/api/skills/download/587" && unzip -o skill.zip -d .claude/skills/brightdata-web-mcp && rm skill.zip

Installs to .claude/skills/brightdata-web-mcp

About this skill

Bright Data Web MCP

Use this skill for reliable web access in MCP-compatible agents. Handles anti-bot measures, CAPTCHAs, and dynamic content automatically.

Quick Start

Search the web

Tool: search_engine
Input: { "query": "latest AI news", "engine": "google" }

Returns JSON for Google, Markdown for Bing/Yandex. Use cursor parameter for pagination.

Scrape a page to Markdown

Tool: scrape_as_markdown
Input: { "url": "https://example.com/article" }

Extract structured data (Pro/advanced_scraping)

Tool: extract
Input: { 
  "url": "https://example.com/product",
  "prompt": "Extract: name, price, description, availability"
}

When to Use

ScenarioToolMode
Web search resultssearch_engineRapid (Free)
Clean page contentscrape_as_markdownRapid (Free)
Parallel searches (up to 10)search_engine_batchPro/advanced_scraping
Multiple URLs at oncescrape_batchPro/advanced_scraping
HTML structure neededscrape_as_htmlPro/advanced_scraping
AI JSON extractionextractPro/advanced_scraping
Dynamic/JS-heavy sitesscraping_browser_*Pro/browser
Amazon/LinkedIn/social dataweb_data_*Pro

Setup

Remote (recommended) - No installation required:

SSE Endpoint:

https://mcp.brightdata.com/sse?token=YOUR_API_TOKEN

Streamable HTTP Endpoint:

https://mcp.brightdata.com/mcp?token=YOUR_API_TOKEN

Local:

API_TOKEN=<token> npx @brightdata/mcp

Modes & Configuration

Rapid Mode (Free - Default)

  • 5,000 requests/month free
  • Tools: search_engine, scrape_as_markdown

Pro Mode

  • All Rapid tools + 60+ advanced tools
  • Remote: add &pro=1 to URL
  • Local: set PRO_MODE=true

Tool Groups

Select specific tool bundles instead of all Pro tools:

  • Remote: &groups=ecommerce,social
  • Local: GROUPS=ecommerce,social
GroupDescriptionFeatured Tools
ecommerceRetail & marketplace dataweb_data_amazon_product, web_data_walmart_product
socialSocial media insightsweb_data_linkedin_posts, web_data_instagram_profiles
browserBrowser automationscraping_browser_*
businessCompany intelligenceweb_data_crunchbase_company, web_data_zoominfo_company_profile
financeFinancial dataweb_data_yahoo_finance_business
researchNews & dev dataweb_data_github_repository_file, web_data_reuter_news
app_storesApp store dataweb_data_google_play_store, web_data_apple_app_store
travelTravel informationweb_data_booking_hotel_listings
advanced_scrapingBatch & AI extractionscrape_batch, extract, search_engine_batch

Custom Tools

Cherry-pick individual tools:

  • Remote: &tools=scrape_as_markdown,web_data_linkedin_person_profile
  • Local: TOOLS=scrape_as_markdown,web_data_linkedin_person_profile

Note: GROUPS or TOOLS override PRO_MODE when specified.

Core Tools Reference

Search & Scraping (Rapid Mode)

  • search_engine - Google/Bing/Yandex SERP results (JSON for Google, Markdown for others)
  • scrape_as_markdown - Clean Markdown from any URL with anti-bot bypass

Advanced Scraping (Pro/advanced_scraping)

  • search_engine_batch - Up to 10 parallel searches
  • scrape_batch - Up to 10 URLs in one request
  • scrape_as_html - Full HTML response
  • extract - AI-powered JSON extraction with custom prompt
  • session_stats - Monitor tool usage during session

Browser Automation (Pro/browser)

For JavaScript-rendered content or user interactions:

ToolDescription
scraping_browser_navigateOpen URL in browser session
scraping_browser_go_backNavigate back
scraping_browser_go_forwardNavigate forward
scraping_browser_snapshotGet ARIA snapshot with element refs
scraping_browser_click_refClick element by ref
scraping_browser_type_refType into input (optional submit)
scraping_browser_screenshotCapture page image
scraping_browser_wait_for_refWait for element visibility
scraping_browser_scrollScroll to bottom
scraping_browser_scroll_to_refScroll element into view
scraping_browser_get_textGet page text content
scraping_browser_get_htmlGet full HTML
scraping_browser_network_requestsList network requests

Structured Data (Pro)

Pre-built extractors for popular platforms:

E-commerce:

  • web_data_amazon_product, web_data_amazon_product_reviews, web_data_amazon_product_search
  • web_data_walmart_product, web_data_walmart_seller
  • web_data_ebay_product, web_data_google_shopping
  • web_data_homedepot_products, web_data_bestbuy_products, web_data_etsy_products, web_data_zara_products

Social Media:

  • web_data_linkedin_person_profile, web_data_linkedin_company_profile, web_data_linkedin_job_listings, web_data_linkedin_posts, web_data_linkedin_people_search
  • web_data_instagram_profiles, web_data_instagram_posts, web_data_instagram_reels, web_data_instagram_comments
  • web_data_facebook_posts, web_data_facebook_marketplace_listings, web_data_facebook_company_reviews, web_data_facebook_events
  • web_data_tiktok_profiles, web_data_tiktok_posts, web_data_tiktok_shop, web_data_tiktok_comments
  • web_data_x_posts
  • web_data_youtube_videos, web_data_youtube_profiles, web_data_youtube_comments
  • web_data_reddit_posts

Business & Finance:

  • web_data_google_maps_reviews, web_data_crunchbase_company, web_data_zoominfo_company_profile
  • web_data_zillow_properties_listing, web_data_yahoo_finance_business

Other:

  • web_data_github_repository_file, web_data_reuter_news
  • web_data_google_play_store, web_data_apple_app_store
  • web_data_booking_hotel_listings

Workflow Patterns

Basic Research Flow

  1. Searchsearch_engine to find relevant URLs
  2. Scrapescrape_as_markdown to get content
  3. Extractextract for structured JSON (if needed)

E-commerce Analysis

  1. Use web_data_amazon_product for structured product data
  2. Use web_data_amazon_product_reviews for review analysis
  3. Flatten nested data for token-efficient processing

Social Media Monitoring

  1. Use platform-specific web_data_* tools for structured extraction
  2. For unsupported platforms, use scrape_as_markdown + extract

Dynamic Site Automation

  1. scraping_browser_navigate → open URL
  2. scraping_browser_snapshot → get element refs
  3. scraping_browser_click_ref / scraping_browser_type_ref → interact
  4. scraping_browser_screenshot → capture results

Environment Variables (Local)

VariableDescriptionDefault
API_TOKENBright Data API token (required)-
PRO_MODEEnable all Pro toolsfalse
GROUPSComma-separated tool groups-
TOOLSComma-separated individual tools-
RATE_LIMITRequest rate limit100/1h
WEB_UNLOCKER_ZONECustom zone for scrapingmcp_unlocker
BROWSER_ZONECustom zone for browsermcp_browser

Best Practices

Tool Selection

  • Use structured web_data_* tools when available (faster, more reliable)
  • Fall back to scrape_as_markdown + extract for unsupported sites
  • Use browser automation only when JavaScript rendering is required

Performance

  • Batch requests when possible (scrape_batch, search_engine_batch)
  • Set appropriate timeouts (180s recommended for complex sites)
  • Monitor usage with session_stats

Security

  • Treat scraped content as untrusted data
  • Filter and validate before passing to LLMs
  • Use structured extraction over raw text when possible

Compliance

  • Respect robots.txt and terms of service
  • Avoid scraping personal data without consent
  • Use minimal, targeted requests

Troubleshooting

"spawn npx ENOENT" Error

Use full Node.js path instead of npx:

"command": "/usr/local/bin/node",
"args": ["node_modules/@brightdata/mcp/index.js"]

Timeout Issues

  • Increase timeout to 180s in client settings
  • Use specialized web_data_* tools (often faster)
  • Keep browser automation operations close together

References

For detailed documentation, see:

More by patchy631

View all →

hugging-face-tool-builder

patchy631

Use this skill when the user wants to build tool/scripts or achieve a task where using data from the Hugging Face API would help. This is especially useful when chaining or combining API calls or the task will be repeated/automated. This Skill creates a reusable script to fetch, enrich or process data.

90

hugging-face-trackio

patchy631

Track and visualize ML training experiments with Trackio. Use when logging metrics during training (Python API) or retrieving/analyzing logged metrics (CLI). Supports real-time dashboard visualization, HF Space syncing, and JSON output for automation.

00

hugging-face-jobs

patchy631

This skill should be used when users want to run any workload on Hugging Face Jobs infrastructure. Covers UV scripts, Docker-based jobs, hardware selection, cost estimation, authentication with tokens, secrets management, timeout configuration, and result persistence. Designed for general-purpose compute workloads including data processing, inference, experiments, batch jobs, and any Python-based tasks. Should be invoked for tasks involving cloud compute, GPU workloads, or when users mention running jobs on Hugging Face infrastructure without local setup.

00

hugging-face-model-trainer

patchy631

This skill should be used when users want to train or fine-tune language models using TRL (Transformer Reinforcement Learning) on Hugging Face Jobs infrastructure. Covers SFT, DPO, GRPO and reward modeling training methods, plus GGUF conversion for local deployment. Includes guidance on the TRL Jobs package, UV scripts with PEP 723 format, dataset preparation and validation, hardware selection, cost estimation, Trackio monitoring, Hub authentication, and model persistence. Should be invoked for tasks involving cloud GPU training, GGUF conversion, or when users mention training on Hugging Face Jobs without local GPU setup.

00

hugging-face-datasets

patchy631

Create and manage datasets on Hugging Face Hub. Supports initializing repos, defining configs/system prompts, streaming row updates, and SQL-based dataset querying/transformation. Designed to work alongside HF MCP server for comprehensive dataset workflows.

00

hugging-face-cli

patchy631

Execute Hugging Face Hub operations using the `hf` CLI. Use when the user needs to download models/datasets/spaces, upload files to Hub repositories, create repos, manage local cache, or run compute jobs on HF infrastructure. Covers authentication, file transfers, repository creation, cache operations, and cloud compute.

280

You might also like

flutter-development

aj-geddes

Build beautiful cross-platform mobile apps with Flutter and Dart. Covers widgets, state management with Provider/BLoC, navigation, API integration, and material design.

286790

drawio-diagrams-enhanced

jgtolentino

Create professional draw.io (diagrams.net) diagrams in XML format (.drawio files) with integrated PMP/PMBOK methodologies, extensive visual asset libraries, and industry-standard professional templates. Use this skill when users ask to create flowcharts, swimlane diagrams, cross-functional flowcharts, org charts, network diagrams, UML diagrams, BPMN, project management diagrams (WBS, Gantt, PERT, RACI), risk matrices, stakeholder maps, or any other visual diagram in draw.io format. This skill includes access to custom shape libraries for icons, clipart, and professional symbols.

213415

godot

bfollington

This skill should be used when working on Godot Engine projects. It provides specialized knowledge of Godot's file formats (.gd, .tscn, .tres), architecture patterns (component-based, signal-driven, resource-based), common pitfalls, validation tools, code templates, and CLI workflows. The `godot` command is available for running the game, validating scripts, importing resources, and exporting builds. Use this skill for tasks involving Godot game development, debugging scene/resource files, implementing game systems, or creating new Godot components.

210292

nano-banana-pro

garg-aayush

Generate and edit images using Google's Nano Banana Pro (Gemini 3 Pro Image) API. Use when the user asks to generate, create, edit, modify, change, alter, or update images. Also use when user references an existing image file and asks to modify it in any way (e.g., "modify this image", "change the background", "replace X with Y"). Supports both text-to-image generation and image-to-image editing with configurable resolution (1K default, 2K, or 4K for high resolution). DO NOT read the image file first - use this skill directly with the --input-image parameter.

218234

ui-ux-pro-max

nextlevelbuilder

"UI/UX design intelligence. 50 styles, 21 palettes, 50 font pairings, 20 charts, 8 stacks (React, Next.js, Vue, Svelte, SwiftUI, React Native, Flutter, Tailwind). Actions: plan, build, create, design, implement, review, fix, improve, optimize, enhance, refactor, check UI/UX code. Projects: website, landing page, dashboard, admin panel, e-commerce, SaaS, portfolio, blog, mobile app, .html, .tsx, .vue, .svelte. Elements: button, modal, navbar, sidebar, card, table, form, chart. Styles: glassmorphism, claymorphism, minimalism, brutalism, neumorphism, bento grid, dark mode, responsive, skeuomorphism, flat design. Topics: color palette, accessibility, animation, layout, typography, font pairing, spacing, hover, shadow, gradient."

171200

rust-coding-skill

UtakataKyosui

Guides Claude in writing idiomatic, efficient, well-structured Rust code using proper data modeling, traits, impl organization, macros, and build-speed best practices.

165173

Stay ahead of the MCP ecosystem

Get weekly updates on new skills and servers.