MCPHub LabRegistrybrightdata/cli
brightdata

brightdata/cli

Built by brightdata β€’ 144 stars

What is brightdata/cli?

Official Bright Data CLI - scrape, search, and extract structured web data directly from your terminal.

How to use brightdata/cli?

1. Install a compatible MCP client (like Claude Desktop). 2. Open your configuration settings. 3. Add brightdata/cli using the following command: npx @modelcontextprotocol/brightdata-cli 4. Restart the client and verify the new tools are active.
πŸ›‘οΈ Scoped (Restricted)
npx @modelcontextprotocol/brightdata-cli --scope restricted
πŸ”“ Unrestricted Access
npx @modelcontextprotocol/brightdata-cli

Key Features

Native MCP Protocol Support
Real-time Tool Activation & Execution
Verified High-performance Implementation
Secure Resource & Context Handling

Optimized Use Cases

Extending AI models with custom local capabilities
Automating system workflows via natural language
Connecting external data sources to LLM context windows

brightdata/cli FAQ

Q

Is brightdata/cli safe?

Yes, brightdata/cli follows the standardized Model Context Protocol security patterns and only executes tools with explicit user-granted permissions.

Q

Is brightdata/cli up to date?

brightdata/cli is currently active in the registry with 144 stars on GitHub, indicating its reliability and community support.

Q

Are there any limits for brightdata/cli?

Usage limits depend on the specific implementation of the MCP server and your system resources. Refer to the official documentation below for technical details.

Official Documentation

View on GitHub
<p align="center"> <img src="https://raw.githubusercontent.com/brightdata/cli/main/assets/banner.gif" alt="Bright Data CLI" width="800" /> </p> <h1 align="center">Bright Data CLI</h1> <p align="center"> Scrape, search, and extract structured web data β€” directly from your terminal. </p> <p align="center"> <a href="https://www.npmjs.com/package/%40brightdata%2Fcli"><img src="https://img.shields.io/npm/v/%40brightdata%2Fcli?color=black&label=npm" alt="npm version" /></a> <img src="https://img.shields.io/badge/node-%3E%3D20-black" alt="node requirement" /> <img src="https://img.shields.io/badge/license-MIT-black" alt="license" /> </p>

Overview

@brightdata/cli is the official npm package for the Bright Data CLI. It installs the brightdata command (with bdata as a shorthand alias) for access to the full Bright Data API surface:

CommandWhat it does
brightdata scrapeScrape any URL β€” bypasses CAPTCHAs, JS rendering, anti-bot protections
brightdata searchGoogle / Bing / Yandex search with structured JSON output
brightdata pipelinesExtract structured data from 40+ platforms (Amazon, LinkedIn, TikTok…)
brightdata browserControl a real browser via Bright Data's Scraping Browser β€” navigate, snapshot, click, type, and more
brightdata zonesList and inspect your Bright Data proxy zones
brightdata budgetView account balance and per-zone cost & bandwidth
brightdata skillInstall Bright Data AI agent skills into your coding agent
brightdata add mcpAdd the Bright Data MCP server to Claude Code, Cursor, or Codex
brightdata configManage CLI configuration
brightdata initInteractive setup wizard

Table of Contents


Installation

Requires Node.js β‰₯ 20

macOS / Linux

curl -fsSL https://cli.brightdata.com/install.sh | sh

Windows

npm install -g @brightdata/cli

Or install manually on any platform

npm install -g @brightdata/cli

You can also run without installing:

npx --yes --package @brightdata/cli brightdata <command>

Quick Start

# 1. Run the interactive setup wizard
brightdata init

# 2. Scrape a page as markdown
brightdata scrape https://example.com

# 3. Search Google
brightdata search "web scraping best practices"

# 4. Extract a LinkedIn profile
brightdata pipelines linkedin_person_profile "https://linkedin.com/in/username"

# 5. Check your account balance
brightdata budget

# 6. Install the Bright Data MCP server into your coding agent
brightdata add mcp

Authentication

Get your API key from brightdata.com/cp/setting/users.

# Interactive β€” opens browser, saves key automatically
brightdata login

# Non-interactive β€” pass key directly
brightdata login --api-key <your-api-key>

# Environment variable β€” no login required
export BRIGHTDATA_API_KEY=your-api-key

On first login the CLI checks for required zones (cli_unlocker, cli_browser) and creates them automatically if missing.

# Clear saved credentials
brightdata logout

brightdata add mcp uses the API key stored by brightdata login. It does not currently read BRIGHTDATA_API_KEY or the global --api-key flag, so log in first before using it.


Commands

init

Interactive setup wizard. The recommended way to get started.

brightdata init

Walks through: API key detection β†’ zone selection β†’ default output format β†’ quick-start examples.

FlagDescription
--skip-authSkip the authentication step
-k, --api-key <key>Provide API key directly

scrape

Scrape any URL using Bright Data's Web Unlocker. Handles CAPTCHAs, JavaScript rendering, and anti-bot protections automatically.

brightdata scrape <url> [options]
FlagDescription
-f, --format <fmt>markdown Β· html Β· screenshot Β· json (default: markdown)
--country <code>Geo-target by ISO country code (e.g. us, de, jp)
--zone <name>Web Unlocker zone name
--mobileUse a mobile user agent
--asyncSubmit async job, return a snapshot ID
-o, --output <path>Write output to file
--json / --prettyJSON output (raw / indented)
-k, --api-key <key>Override API key

Examples

# Scrape as markdown (default)
brightdata scrape https://news.ycombinator.com

# Scrape as raw HTML
brightdata scrape https://example.com -f html

# US geo-targeting, save to file
brightdata scrape https://amazon.com -f json --country us -o product.json

# Pipe to a markdown viewer
brightdata scrape https://docs.github.com | glow -

# Async β€” returns a snapshot ID you can poll with `status`
brightdata scrape https://example.com --async

search

Search Google, Bing, or Yandex via Bright Data's SERP API. Google results include structured data (organic results, ads, people-also-ask, related searches).

brightdata search <query> [options]
FlagDescription
--engine <name>google Β· bing Β· yandex (default: google)
--country <code>Localized results (e.g. us, de)
--language <code>Language code (e.g. en, fr)
--page <n>Page number, 0-indexed (default: 0)
--type <type>web Β· news Β· images Β· shopping (default: web)
--device <type>desktop Β· mobile
--zone <name>SERP zone name
-o, --output <path>Write output to file
--json / --prettyJSON output (raw / indented)
-k, --api-key <key>Override API key

Examples

# Formatted table output (default)
brightdata search "typescript best practices"

# German localized results
brightdata search "restaurants berlin" --country de --language de

# News search
brightdata search "AI regulation" --type news

# Page 2 of results
brightdata search "web scraping" --page 1

# Extract just the URLs
brightdata search "open source scraping" --json | jq -r '.organic[].link'

# Search Bing
brightdata search "bright data pricing" --engine bing

pipelines

Extract structured data from 40+ platforms using Bright Data's Web Scraper API. Triggers an async collection job, polls until ready, and returns results.

brightdata pipelines <type> [params...] [options]
FlagDescription
--format <fmt>json Β· csv Β· ndjson Β· jsonl (default: json)
--timeout <seconds>Polling timeout (default: 600)
-o, --output <path>Write output to file
--json / --prettyJSON output (raw / indented)
-k, --api-key <key>Override API key
# List all available dataset types
brightdata pipelines list

Examples

# LinkedIn profile
brightdata pipelines linkedin_person_profile "https://linkedin.com/in/username"

# Amazon product β†’ CSV
brightdata pipelines amazon_product "https://amazon.com/dp/B09V3KXJPB" \
  --format csv -o product.csv

# Instagram profile
brightdata pipelines instagram_profiles "https://instagram.com/username"

# Amazon search by keyword
brightdata pipelines amazon_product_search "laptop" "https://amazon.com"

# Google Maps reviews
brightdata pipelines google_maps_reviews "https://maps.google.com/..." 7

# YouTube comments (top 50)
brightdata pipelines youtube_comments "https://youtube.com/watch?v=..." 50

See Dataset Types Reference for the full list.


browser

Control a real browser session powered by Bright Data's Scraping Browser. A lightweight local daemon holds the browser connection open between commands, giving you persistent state without reconnecting on every call.

brightdata browser open <url>              # Start a session and navigate
brightdata browser snapshot                # Get an accessibility tree of the page
brightdata browser screenshot [path]       # Take a PNG screenshot
brightdata browser click <ref>             # Click an element
brightdata browser type <ref> <text>       # Type into an element
brightdata browser fill <ref> <value>      # Fill a form field
brightdata browser select <ref> <value>    # Select a dropdown option
brightdata browser check <ref>             # Check a checkbox / radio
brightdata browser uncheck <ref>           # Uncheck a checkbox
brightdata browser hover <ref>             # Hover over an element
brightdata browser scroll                  # Scroll the page
brightdata browser get text [selector]     # Get text content
brightdata browser get html [selector]     # Get HTML content
brightdata browser back                    # Navigate back
brightdata browser forward                 # Navigate forward
brightdata browser reload                  # Reload the page
brightdata browser network                 # Show captured network requests
brightdata browser cookies                 # Show cookies
brightdata browser status                  # Show session state
brightdata browser sessions                # List all active sessions
brightdata browser close                   # Close session and stop daemon

Global flags (work with every subcommand)

FlagDescription
--session <name>Session name β€” run multiple isolated sessions in parallel (default: default)
--country <code>Geo-target by ISO country code (e.g. us, de). On open, changing country reconnects the browser
--zone <name>Scraping Browser zone (default: cli_browser)
--timeout <ms>IPC command timeout in milliseconds (default: 30000)
--idle-timeout <ms>Daemon auto-shutdown after idle (default: 600000 = 10 min)
--json / --prettyJSON output
-o, --output <path>Write output to file
-k, --api-key <key>Override API key

browser open <url>

Navigate to a URL. Starts the daemon and browser session automatically if not already running.

brightdata browser open https://example.com
brightdata browser open https://amazon.com --country us --session shop
FlagDescription
--country <code>Geo-targeting. Reconnects the browser if the country changes on an existing session
--zone <name>Browser zone name
--idle-timeout <ms>Daemon idle timeout for this session

browser snapshot

Capture the page as a text accessibility tree. This is the primary way AI agents read page content β€” far more token-efficient than raw HTML.

brightdata browser snapshot
brightdata browser snapshot --compact          # Interactive elements + ancestors only
brightdata browser snapshot --interactive      # Interactive elements as a flat list
brightdata browser snapshot --depth 3          # Limit tree depth
brightdata browser snapshot --selector "main"  # Scope to a CSS subtree
brightdata browser snapshot --wrap             # Wrap output in AI-safe content boundaries

Output format:

Page: Example Domain
URL: https://example.com

- heading "Example Domain" [level=1]
- paragraph "This domain is for use in illustrative examples."
- link "More information..." [ref=e1]

Each interactive element gets a ref (e.g. e1, e2) that you pass to click, type, fill, etc.

FlagDescription
--compactOnly interactive elements and their ancestors (70–90% fewer tokens)
--interactiveOnly interactive elements, as a flat list
--depth <n>Limit tree depth to a non-negative integer
--selector <sel>Scope snapshot to elements matching a CSS selector
--wrapWrap output in --- BRIGHTDATA_BROWSER_CONTENT ... --- boundaries (useful for AI agent prompt injection safety)

browser screenshot [path]

Capture a PNG screenshot of the current viewport.

brightdata browser screenshot
brightdata browser screenshot ./result.png
brightdata browser screenshot --full-page -o page.png
brightdata browser screenshot --base64
FlagDescription
[path]Where to save the PNG (default: temp directory)
--full-pageCapture the full scrollable page, not just the viewport
--base64Output base64-encoded PNG data instead of saving to a file

browser click <ref>

Click an element by its snapshot ref.

brightdata browser click e3
brightdata browser click e3 --session shop

browser type <ref> <text>

Type text into an element. Clears the field first by default.

brightdata browser type e5 "search query"
brightdata browser type e5 " more text" --append   # Append to existing value
brightdata browser type e5 "search query" --submit  # Press Enter after typing
FlagDescription
--appendAppend to existing value using key-by-key simulation
--submitPress Enter after typing

browser fill <ref> <value>

Fill a form field directly (no keyboard simulation). Use type if you need to trigger keydown/keyup events.

brightdata browser fill e2 "user@example.com"

browser select <ref> <value>

Select a dropdown option by its visible label.

brightdata browser select e4 "United States"

browser check <ref> / browser uncheck <ref>

Check or uncheck a checkbox or radio button.

brightdata browser check e7
brightdata browser uncheck e7

browser hover <ref>

Hover the mouse over an element (triggers hover states, tooltips, dropdowns).

brightdata browser hover e2

browser scroll

Scroll the viewport or scroll an element into view.

brightdata browser scroll                        # Scroll down 300px (default)
brightdata browser scroll --direction up
brightdata browser scroll --direction down --distance 600
brightdata browser scroll --ref e10              # Scroll element e10 into view
FlagDescription
--direction <dir>up, down, left, right (default: down)
--distance <px>Pixels to scroll (default: 300)
--ref <ref>Scroll this element into view instead of the viewport

browser get text [selector]

Get the text content of the page or a scoped element.

brightdata browser get text           # Full page text
brightdata browser get text "h1"      # Text of the first h1
brightdata browser get text "#price"  # Text inside #price

browser get html [selector]

Get the HTML of the page or a scoped element.

brightdata browser get html              # Full page outer HTML
brightdata browser get html ".product"   # innerHTML of .product
brightdata browser get html --pretty     # JSON output with selector field

browser network

Show HTTP requests captured since the last navigation.

brightdata browser network
brightdata browser network --json

Example output:

Network Requests (5 total):
[GET] https://example.com/ => [200]
[GET] https://example.com/style.css => [200]
[POST] https://api.example.com/track => [204]

browser cookies

Show cookies for the active session.

brightdata browser cookies
brightdata browser cookies --pretty

browser status

Show the current state of a browser session.

brightdata browser status
brightdata browser status --session shop --pretty

browser sessions

List all active browser daemon sessions.

brightdata browser sessions
brightdata browser sessions --pretty

browser close

Close a session and stop its daemon.

brightdata browser close                   # Close the default session
brightdata browser close --session shop    # Close a named session
brightdata browser close --all             # Close all active sessions

Example: AI agent workflow

# Open a US-targeted session
brightdata browser open https://example.com --country us

# Read the page structure (compact for token efficiency)
brightdata browser snapshot --compact

# Interact using refs from the snapshot
brightdata browser click e3
brightdata browser type e5 "search query" --submit

# Get updated snapshot after interaction
brightdata browser snapshot --compact

# Save a screenshot for visual verification
brightdata browser screenshot ./result.png

# Done
brightdata browser close

Example: multi-session comparison

brightdata browser open https://amazon.com --session us --country us
brightdata browser open https://amazon.com --session de --country de

brightdata browser snapshot --session us --json > us.json
brightdata browser snapshot --session de --json > de.json

brightdata browser close --all

status

Check the status of an async snapshot job (returned by --async or pipelines).

brightdata status <job-id> [options]
FlagDescription
--waitPoll until the job completes
--timeout <seconds>Polling timeout (default: 600)
-o, --output <path>Write output to file
--json / --prettyJSON output (raw / indented)
-k, --api-key <key>Override API key
# Check current status
brightdata status s_abc123xyz

# Block until complete
brightdata status s_abc123xyz --wait --pretty

# Custom timeout (5 minutes)
brightdata status s_abc123xyz --wait --timeout 300

zones

List and inspect your Bright Data proxy zones.

brightdata zones               # List all active zones
brightdata zones info <name>   # Show full details for a zone
# Export all zones as JSON
brightdata zones --json -o zones.json

# Inspect a specific zone
brightdata zones info my_unlocker_zone --pretty

budget

View your account balance and per-zone cost and bandwidth usage. Read-only β€” no writes to the API.

brightdata budget                     # Show account balance (quick view)
brightdata budget balance             # Account balance + pending charges
brightdata budget zones               # Cost & bandwidth table for all zones
brightdata budget zone <name>         # Detailed cost & bandwidth for one zone
FlagDescription
--from <datetime>Start of date range (e.g. 2024-01-01T00:00:00)
--to <datetime>End of date range
--json / --prettyJSON output (raw / indented)
-k, --api-key <key>Override API key
# Current account balance
brightdata budget

# Zone costs for January 2024
brightdata budget zones --from 2024-01-01T00:00:00 --to 2024-02-01T00:00:00

# Detailed view of a specific zone
brightdata budget zone my_unlocker_zone

skill

Install Bright Data AI agent skills into your coding agent (Claude Code, Cursor, Copilot, etc.). Skills provide your agent with context and instructions for using Bright Data APIs effectively.

brightdata skill add              # Interactive picker β€” choose skill + agent
brightdata skill add <name>       # Install a specific skill directly
brightdata skill list             # List all available Bright Data skills

Available skills

SkillDescription
searchSearch Google and get structured JSON results
scrapeScrape any webpage as clean markdown with bot bypass
data-feedsExtract structured data from 40+ websites
bright-data-mcpOrchestrate 60+ Bright Data MCP tools
bright-data-best-practicesReference knowledge base for writing Bright Data code
# Interactive β€” select skills and choose which agents to install to
brightdata skill add

# Install the scrape skill directly
brightdata skill add scrape

# See what's available
brightdata skill list

add mcp

Write a Bright Data MCP server entry into Claude Code, Cursor, or Codex config files using the API key already stored by brightdata login.

brightdata add mcp                               # Interactive agent + scope prompts
brightdata add mcp --agent claude-code --global
brightdata add mcp --agent claude-code,cursor --project
brightdata add mcp --agent codex --global
FlagDescription
--agent <agents>Comma-separated targets: claude-code,cursor,codex
--globalInstall to the agent's global config file
--projectInstall to the current project's config file

Config targets

AgentGlobal pathProject path
Claude Code~/.claude.json.claude/settings.json
Cursor~/.cursor/mcp.json.cursor/mcp.json
Codex$CODEX_HOME/mcp.json or ~/.codex/mcp.jsonNot supported

The command writes the MCP server under mcpServers["bright-data"]:

{
  "mcpServers": {
    "bright-data": {
      "command": "npx",
      "args": ["@brightdata/mcp"],
      "env": {
        "API_TOKEN": "<stored-api-key>"
      }
    }
  }
}

Behavior notes:

  • Existing config is preserved; only mcpServers["bright-data"] is added or replaced.
  • If the target config contains invalid JSON, the CLI warns and offers to overwrite it in interactive mode.
  • In non-interactive mode, pass both --agent and the appropriate scope flag to skip prompts.

config

View and manage CLI configuration.

brightdata config                              # Show all config
brightdata config get <key>                    # Get a single value
brightdata config set <key> <value>            # Set a value
KeyDescription
default_zone_unlockerDefault zone for scrape and search
default_zone_serpDefault zone for search (overrides unlocker zone)
default_formatDefault output format: markdown or json
api_urlOverride the Bright Data API base URL
brightdata config set default_zone_unlocker my_zone
brightdata config set default_format json

login / logout

brightdata login                      # Interactive login
brightdata login --api-key <key>      # Non-interactive
brightdata logout                     # Clear saved credentials

Configuration

Config is stored in an OS-appropriate location:

OSPath
macOS~/Library/Application Support/brightdata-cli/
Linux~/.config/brightdata-cli/
Windows%APPDATA%\brightdata-cli\

Two files are stored:

  • credentials.json β€” API key
  • config.json β€” zones, output format, preferences

Priority order (highest β†’ lowest):

CLI flags  β†’  Environment variables  β†’  config.json  β†’  Defaults

Environment Variables

VariableDescription
BRIGHTDATA_API_KEYAPI key (overrides stored credentials)
BRIGHTDATA_UNLOCKER_ZONEDefault Web Unlocker zone
BRIGHTDATA_SERP_ZONEDefault SERP zone
BRIGHTDATA_POLLING_TIMEOUTDefault polling timeout in seconds
BRIGHTDATA_BROWSER_ZONEDefault Scraping Browser zone (default: cli_browser)
BRIGHTDATA_DAEMON_DIROverride the directory used for browser daemon socket and PID files
BRIGHTDATA_API_KEY=xxx BRIGHTDATA_UNLOCKER_ZONE=my_zone \
  brightdata scrape https://example.com

Output Modes

Every command supports:

ModeFlagBehavior
Human-readable(default)Formatted table or markdown, with colors
JSON--jsonCompact JSON to stdout
Pretty JSON--prettyIndented JSON to stdout
File-o <path>Write to file; format inferred from extension

Auto-detected file formats:

ExtensionFormat
.jsonJSON
.mdMarkdown
.htmlHTML
.csvCSV

Pipe-Friendly Usage

When stdout is not a TTY, colors and spinners are automatically disabled. Errors go to stderr, data to stdout.

# Extract URLs from search results
brightdata search "nodejs tutorials" --json | jq -r '.organic[].link'

# Scrape and view with a markdown reader
brightdata scrape https://docs.github.com | glow -

# Save scraped content to a file
brightdata scrape https://example.com -f markdown > page.md

# Amazon product data as CSV
brightdata pipelines amazon_product "https://amazon.com/dp/xxx" --format csv > product.csv

# Chain search β†’ scrape
brightdata search "top open source projects" --json \
  | jq -r '.organic[0].link' \
  | xargs brightdata scrape

Dataset Types Reference

brightdata pipelines list   # See all types in your terminal

E-Commerce

TypePlatform
amazon_productAmazon product page
amazon_product_reviewsAmazon reviews
amazon_product_searchAmazon search results
walmart_productWalmart product page
walmart_sellerWalmart seller profile
ebay_producteBay listing
bestbuy_productsBest Buy
etsy_productsEtsy
homedepot_productsHome Depot
zara_productsZara
google_shoppingGoogle Shopping

Professional Networks

TypePlatform
linkedin_person_profileLinkedIn person
linkedin_company_profileLinkedIn company
linkedin_job_listingsLinkedIn jobs
linkedin_postsLinkedIn posts
linkedin_people_searchLinkedIn people search
crunchbase_companyCrunchbase
zoominfo_company_profileZoomInfo

Social Media

TypePlatform
instagram_profilesInstagram profiles
instagram_postsInstagram posts
instagram_reelsInstagram reels
instagram_commentsInstagram comments
facebook_postsFacebook posts
facebook_marketplace_listingsFacebook Marketplace
facebook_company_reviewsFacebook reviews
facebook_eventsFacebook events
tiktok_profilesTikTok profiles
tiktok_postsTikTok posts
tiktok_shopTikTok shop
tiktok_commentsTikTok comments
x_postsX (Twitter) posts
youtube_profilesYouTube channels
youtube_videosYouTube videos
youtube_commentsYouTube comments
reddit_postsReddit posts

Other

TypePlatform
google_maps_reviewsGoogle Maps reviews
google_play_storeGoogle Play
apple_app_storeApple App Store
reuter_newsReuters news
github_repository_fileGitHub repository files
yahoo_finance_businessYahoo Finance
zillow_properties_listingZillow
booking_hotel_listingsBooking.com

Troubleshooting

Error: No Web Unlocker zone specified

brightdata config set default_zone_unlocker <your-zone-name>
# or
export BRIGHTDATA_UNLOCKER_ZONE=<your-zone-name>

Error: Invalid or expired API key

brightdata login

Error: Access denied

Check zone permissions in the Bright Data control panel.

Error: Rate limit exceeded

Wait a moment and retry. Use --async for large jobs to avoid timeouts.

Async job is too slow

brightdata pipelines amazon_product <url> --timeout 1200
# or
export BRIGHTDATA_POLLING_TIMEOUT=1200

No active browser session "default"

# Start a session first
brightdata browser open https://example.com

Browser daemon won't start

# Check if a stale socket file exists and clear it
brightdata browser close
# Then retry
brightdata browser open https://example.com

Element ref not found after interaction

Refs are re-assigned on every snapshot call. If you navigate or click (which may cause the page to change), take a fresh snapshot before using refs again:

brightdata browser click e3
brightdata browser snapshot --compact   # refresh refs
brightdata browser type e5 "text"

Garbled output in non-interactive terminal

Colors and spinners are disabled automatically when not in a TTY. If you still see ANSI codes, add | cat at the end of your command.


Links


<p align="center"> <sub>Β© Bright Data Β· ISC License</sub> </p>

Global Ranking

8.5
Trust ScoreMCPHub Index

Based on codebase health & activity.

Manual Config

{ "mcpServers": { "brightdata-cli": { "command": "npx", "args": ["brightdata-cli"] } } }