
Research
Supply Chain Attack on Axios Pulls Malicious Dependency from npm
A supply chain attack on Axios introduced a malicious dependency, plain-crypto-js@4.2.1, published minutes earlier and absent from the project’s GitHub releases.
@brightdata/cli
Advanced tools
Command-line interface for Bright Data. Scrape, search, extract structured data, and automate browsers directly from your terminal.
Scrape, search, and extract structured web data — directly from your terminal.
@brightdata/cli is the official npm package for the Bright Data CLI. It installs the brightdata command (with bdata as a shorthand alias) for access to the full Bright Data API surface:
| Command | What it does |
|---|---|
brightdata scrape | Scrape any URL — bypasses CAPTCHAs, JS rendering, anti-bot protections |
brightdata search | Google / Bing / Yandex search with structured JSON output |
brightdata pipelines | Extract structured data from 40+ platforms (Amazon, LinkedIn, TikTok…) |
brightdata browser | Control a real browser via Bright Data's Scraping Browser — navigate, snapshot, click, type, and more |
brightdata zones | List and inspect your Bright Data proxy zones |
brightdata budget | View account balance and per-zone cost & bandwidth |
brightdata skill | Install Bright Data AI agent skills into your coding agent |
brightdata add mcp | Add the Bright Data MCP server to Claude Code, Cursor, or Codex |
brightdata config | Manage CLI configuration |
brightdata init | Interactive setup wizard |
Requires Node.js ≥ 20
curl -fsSL https://cli.brightdata.com/install.sh | sh
npm install -g @brightdata/cli
npm install -g @brightdata/cli
You can also run without installing:
npx --yes --package @brightdata/cli brightdata <command>
# 1. Run the interactive setup wizard
brightdata init
# 2. Scrape a page as markdown
brightdata scrape https://example.com
# 3. Search Google
brightdata search "web scraping best practices"
# 4. Extract a LinkedIn profile
brightdata pipelines linkedin_person_profile "https://linkedin.com/in/username"
# 5. Check your account balance
brightdata budget
# 6. Install the Bright Data MCP server into your coding agent
brightdata add mcp
Get your API key from brightdata.com/cp/setting/users.
# Interactive — opens browser, saves key automatically
brightdata login
# Non-interactive — pass key directly
brightdata login --api-key <your-api-key>
# Environment variable — no login required
export BRIGHTDATA_API_KEY=your-api-key
On first login the CLI checks for required zones (cli_unlocker, cli_browser) and creates them automatically if missing.
# Clear saved credentials
brightdata logout
brightdata add mcp uses the API key stored by brightdata login. It does not currently read BRIGHTDATA_API_KEY or the global --api-key flag, so log in first before using it.
initInteractive setup wizard. The recommended way to get started.
brightdata init
Walks through: API key detection → zone selection → default output format → quick-start examples.
| Flag | Description |
|---|---|
--skip-auth | Skip the authentication step |
-k, --api-key <key> | Provide API key directly |
scrapeScrape any URL using Bright Data's Web Unlocker. Handles CAPTCHAs, JavaScript rendering, and anti-bot protections automatically.
brightdata scrape <url> [options]
| Flag | Description |
|---|---|
-f, --format <fmt> | markdown · html · screenshot · json (default: markdown) |
--country <code> | Geo-target by ISO country code (e.g. us, de, jp) |
--zone <name> | Web Unlocker zone name |
--mobile | Use a mobile user agent |
--async | Submit async job, return a snapshot ID |
-o, --output <path> | Write output to file |
--json / --pretty | JSON output (raw / indented) |
-k, --api-key <key> | Override API key |
Examples
# Scrape as markdown (default)
brightdata scrape https://news.ycombinator.com
# Scrape as raw HTML
brightdata scrape https://example.com -f html
# US geo-targeting, save to file
brightdata scrape https://amazon.com -f json --country us -o product.json
# Pipe to a markdown viewer
brightdata scrape https://docs.github.com | glow -
# Async — returns a snapshot ID you can poll with `status`
brightdata scrape https://example.com --async
searchSearch Google, Bing, or Yandex via Bright Data's SERP API. Google results include structured data (organic results, ads, people-also-ask, related searches).
brightdata search <query> [options]
| Flag | Description |
|---|---|
--engine <name> | google · bing · yandex (default: google) |
--country <code> | Localized results (e.g. us, de) |
--language <code> | Language code (e.g. en, fr) |
--page <n> | Page number, 0-indexed (default: 0) |
--type <type> | web · news · images · shopping (default: web) |
--device <type> | desktop · mobile |
--zone <name> | SERP zone name |
-o, --output <path> | Write output to file |
--json / --pretty | JSON output (raw / indented) |
-k, --api-key <key> | Override API key |
Examples
# Formatted table output (default)
brightdata search "typescript best practices"
# German localized results
brightdata search "restaurants berlin" --country de --language de
# News search
brightdata search "AI regulation" --type news
# Page 2 of results
brightdata search "web scraping" --page 1
# Extract just the URLs
brightdata search "open source scraping" --json | jq -r '.organic[].link'
# Search Bing
brightdata search "bright data pricing" --engine bing
pipelinesExtract structured data from 40+ platforms using Bright Data's Web Scraper API. Triggers an async collection job, polls until ready, and returns results.
brightdata pipelines <type> [params...] [options]
| Flag | Description |
|---|---|
--format <fmt> | json · csv · ndjson · jsonl (default: json) |
--timeout <seconds> | Polling timeout (default: 600) |
-o, --output <path> | Write output to file |
--json / --pretty | JSON output (raw / indented) |
-k, --api-key <key> | Override API key |
# List all available dataset types
brightdata pipelines list
Examples
# LinkedIn profile
brightdata pipelines linkedin_person_profile "https://linkedin.com/in/username"
# Amazon product → CSV
brightdata pipelines amazon_product "https://amazon.com/dp/B09V3KXJPB" \
--format csv -o product.csv
# Instagram profile
brightdata pipelines instagram_profiles "https://instagram.com/username"
# Amazon search by keyword
brightdata pipelines amazon_product_search "laptop" "https://amazon.com"
# Google Maps reviews
brightdata pipelines google_maps_reviews "https://maps.google.com/..." 7
# YouTube comments (top 50)
brightdata pipelines youtube_comments "https://youtube.com/watch?v=..." 50
See Dataset Types Reference for the full list.
browserControl a real browser session powered by Bright Data's Scraping Browser. A lightweight local daemon holds the browser connection open between commands, giving you persistent state without reconnecting on every call.
brightdata browser open <url> # Start a session and navigate
brightdata browser snapshot # Get an accessibility tree of the page
brightdata browser screenshot [path] # Take a PNG screenshot
brightdata browser click <ref> # Click an element
brightdata browser type <ref> <text> # Type into an element
brightdata browser fill <ref> <value> # Fill a form field
brightdata browser select <ref> <value> # Select a dropdown option
brightdata browser check <ref> # Check a checkbox / radio
brightdata browser uncheck <ref> # Uncheck a checkbox
brightdata browser hover <ref> # Hover over an element
brightdata browser scroll # Scroll the page
brightdata browser get text [selector] # Get text content
brightdata browser get html [selector] # Get HTML content
brightdata browser back # Navigate back
brightdata browser forward # Navigate forward
brightdata browser reload # Reload the page
brightdata browser network # Show captured network requests
brightdata browser cookies # Show cookies
brightdata browser status # Show session state
brightdata browser sessions # List all active sessions
brightdata browser close # Close session and stop daemon
Global flags (work with every subcommand)
| Flag | Description |
|---|---|
--session <name> | Session name — run multiple isolated sessions in parallel (default: default) |
--country <code> | Geo-target by ISO country code (e.g. us, de). On open, changing country reconnects the browser |
--zone <name> | Scraping Browser zone (default: cli_browser) |
--timeout <ms> | IPC command timeout in milliseconds (default: 30000) |
--idle-timeout <ms> | Daemon auto-shutdown after idle (default: 600000 = 10 min) |
--json / --pretty | JSON output |
-o, --output <path> | Write output to file |
-k, --api-key <key> | Override API key |
browser open <url>Navigate to a URL. Starts the daemon and browser session automatically if not already running.
brightdata browser open https://example.com
brightdata browser open https://amazon.com --country us --session shop
| Flag | Description |
|---|---|
--country <code> | Geo-targeting. Reconnects the browser if the country changes on an existing session |
--zone <name> | Browser zone name |
--idle-timeout <ms> | Daemon idle timeout for this session |
browser snapshotCapture the page as a text accessibility tree. This is the primary way AI agents read page content — far more token-efficient than raw HTML.
brightdata browser snapshot
brightdata browser snapshot --compact # Interactive elements + ancestors only
brightdata browser snapshot --interactive # Interactive elements as a flat list
brightdata browser snapshot --depth 3 # Limit tree depth
brightdata browser snapshot --selector "main" # Scope to a CSS subtree
brightdata browser snapshot --wrap # Wrap output in AI-safe content boundaries
Output format:
Page: Example Domain
URL: https://example.com
- heading "Example Domain" [level=1]
- paragraph "This domain is for use in illustrative examples."
- link "More information..." [ref=e1]
Each interactive element gets a ref (e.g. e1, e2) that you pass to click, type, fill, etc.
| Flag | Description |
|---|---|
--compact | Only interactive elements and their ancestors (70–90% fewer tokens) |
--interactive | Only interactive elements, as a flat list |
--depth <n> | Limit tree depth to a non-negative integer |
--selector <sel> | Scope snapshot to elements matching a CSS selector |
--wrap | Wrap output in --- BRIGHTDATA_BROWSER_CONTENT ... --- boundaries (useful for AI agent prompt injection safety) |
browser screenshot [path]Capture a PNG screenshot of the current viewport.
brightdata browser screenshot
brightdata browser screenshot ./result.png
brightdata browser screenshot --full-page -o page.png
brightdata browser screenshot --base64
| Flag | Description |
|---|---|
[path] | Where to save the PNG (default: temp directory) |
--full-page | Capture the full scrollable page, not just the viewport |
--base64 | Output base64-encoded PNG data instead of saving to a file |
browser click <ref>Click an element by its snapshot ref.
brightdata browser click e3
brightdata browser click e3 --session shop
browser type <ref> <text>Type text into an element. Clears the field first by default.
brightdata browser type e5 "search query"
brightdata browser type e5 " more text" --append # Append to existing value
brightdata browser type e5 "search query" --submit # Press Enter after typing
| Flag | Description |
|---|---|
--append | Append to existing value using key-by-key simulation |
--submit | Press Enter after typing |
browser fill <ref> <value>Fill a form field directly (no keyboard simulation). Use type if you need to trigger keydown/keyup events.
brightdata browser fill e2 "user@example.com"
browser select <ref> <value>Select a dropdown option by its visible label.
brightdata browser select e4 "United States"
browser check <ref> / browser uncheck <ref>Check or uncheck a checkbox or radio button.
brightdata browser check e7
brightdata browser uncheck e7
browser hover <ref>Hover the mouse over an element (triggers hover states, tooltips, dropdowns).
brightdata browser hover e2
browser scrollScroll the viewport or scroll an element into view.
brightdata browser scroll # Scroll down 300px (default)
brightdata browser scroll --direction up
brightdata browser scroll --direction down --distance 600
brightdata browser scroll --ref e10 # Scroll element e10 into view
| Flag | Description |
|---|---|
--direction <dir> | up, down, left, right (default: down) |
--distance <px> | Pixels to scroll (default: 300) |
--ref <ref> | Scroll this element into view instead of the viewport |
browser get text [selector]Get the text content of the page or a scoped element.
brightdata browser get text # Full page text
brightdata browser get text "h1" # Text of the first h1
brightdata browser get text "#price" # Text inside #price
browser get html [selector]Get the HTML of the page or a scoped element.
brightdata browser get html # Full page outer HTML
brightdata browser get html ".product" # innerHTML of .product
brightdata browser get html --pretty # JSON output with selector field
browser networkShow HTTP requests captured since the last navigation.
brightdata browser network
brightdata browser network --json
Example output:
Network Requests (5 total):
[GET] https://example.com/ => [200]
[GET] https://example.com/style.css => [200]
[POST] https://api.example.com/track => [204]
browser cookiesShow cookies for the active session.
brightdata browser cookies
brightdata browser cookies --pretty
browser statusShow the current state of a browser session.
brightdata browser status
brightdata browser status --session shop --pretty
browser sessionsList all active browser daemon sessions.
brightdata browser sessions
brightdata browser sessions --pretty
browser closeClose a session and stop its daemon.
brightdata browser close # Close the default session
brightdata browser close --session shop # Close a named session
brightdata browser close --all # Close all active sessions
Example: AI agent workflow
# Open a US-targeted session
brightdata browser open https://example.com --country us
# Read the page structure (compact for token efficiency)
brightdata browser snapshot --compact
# Interact using refs from the snapshot
brightdata browser click e3
brightdata browser type e5 "search query" --submit
# Get updated snapshot after interaction
brightdata browser snapshot --compact
# Save a screenshot for visual verification
brightdata browser screenshot ./result.png
# Done
brightdata browser close
Example: multi-session comparison
brightdata browser open https://amazon.com --session us --country us
brightdata browser open https://amazon.com --session de --country de
brightdata browser snapshot --session us --json > us.json
brightdata browser snapshot --session de --json > de.json
brightdata browser close --all
statusCheck the status of an async snapshot job (returned by --async or pipelines).
brightdata status <job-id> [options]
| Flag | Description |
|---|---|
--wait | Poll until the job completes |
--timeout <seconds> | Polling timeout (default: 600) |
-o, --output <path> | Write output to file |
--json / --pretty | JSON output (raw / indented) |
-k, --api-key <key> | Override API key |
# Check current status
brightdata status s_abc123xyz
# Block until complete
brightdata status s_abc123xyz --wait --pretty
# Custom timeout (5 minutes)
brightdata status s_abc123xyz --wait --timeout 300
zonesList and inspect your Bright Data proxy zones.
brightdata zones # List all active zones
brightdata zones info <name> # Show full details for a zone
# Export all zones as JSON
brightdata zones --json -o zones.json
# Inspect a specific zone
brightdata zones info my_unlocker_zone --pretty
budgetView your account balance and per-zone cost and bandwidth usage. Read-only — no writes to the API.
brightdata budget # Show account balance (quick view)
brightdata budget balance # Account balance + pending charges
brightdata budget zones # Cost & bandwidth table for all zones
brightdata budget zone <name> # Detailed cost & bandwidth for one zone
| Flag | Description |
|---|---|
--from <datetime> | Start of date range (e.g. 2024-01-01T00:00:00) |
--to <datetime> | End of date range |
--json / --pretty | JSON output (raw / indented) |
-k, --api-key <key> | Override API key |
# Current account balance
brightdata budget
# Zone costs for January 2024
brightdata budget zones --from 2024-01-01T00:00:00 --to 2024-02-01T00:00:00
# Detailed view of a specific zone
brightdata budget zone my_unlocker_zone
skillInstall Bright Data AI agent skills into your coding agent (Claude Code, Cursor, Copilot, etc.). Skills provide your agent with context and instructions for using Bright Data APIs effectively.
brightdata skill add # Interactive picker — choose skill + agent
brightdata skill add <name> # Install a specific skill directly
brightdata skill list # List all available Bright Data skills
Available skills
| Skill | Description |
|---|---|
search | Search Google and get structured JSON results |
scrape | Scrape any webpage as clean markdown with bot bypass |
data-feeds | Extract structured data from 40+ websites |
bright-data-mcp | Orchestrate 60+ Bright Data MCP tools |
bright-data-best-practices | Reference knowledge base for writing Bright Data code |
# Interactive — select skills and choose which agents to install to
brightdata skill add
# Install the scrape skill directly
brightdata skill add scrape
# See what's available
brightdata skill list
add mcpWrite a Bright Data MCP server entry into Claude Code, Cursor, or Codex config files using the API key already stored by brightdata login.
brightdata add mcp # Interactive agent + scope prompts
brightdata add mcp --agent claude-code --global
brightdata add mcp --agent claude-code,cursor --project
brightdata add mcp --agent codex --global
| Flag | Description |
|---|---|
--agent <agents> | Comma-separated targets: claude-code,cursor,codex |
--global | Install to the agent's global config file |
--project | Install to the current project's config file |
Config targets
| Agent | Global path | Project path |
|---|---|---|
| Claude Code | ~/.claude.json | .claude/settings.json |
| Cursor | ~/.cursor/mcp.json | .cursor/mcp.json |
| Codex | $CODEX_HOME/mcp.json or ~/.codex/mcp.json | Not supported |
The command writes the MCP server under mcpServers["bright-data"]:
{
"mcpServers": {
"bright-data": {
"command": "npx",
"args": ["@brightdata/mcp"],
"env": {
"API_TOKEN": "<stored-api-key>"
}
}
}
}
Behavior notes:
mcpServers["bright-data"] is added or replaced.--agent and the appropriate scope flag to skip prompts.configView and manage CLI configuration.
brightdata config # Show all config
brightdata config get <key> # Get a single value
brightdata config set <key> <value> # Set a value
| Key | Description |
|---|---|
default_zone_unlocker | Default zone for scrape and search |
default_zone_serp | Default zone for search (overrides unlocker zone) |
default_format | Default output format: markdown or json |
api_url | Override the Bright Data API base URL |
brightdata config set default_zone_unlocker my_zone
brightdata config set default_format json
login / logoutbrightdata login # Interactive login
brightdata login --api-key <key> # Non-interactive
brightdata logout # Clear saved credentials
Config is stored in an OS-appropriate location:
| OS | Path |
|---|---|
| macOS | ~/Library/Application Support/brightdata-cli/ |
| Linux | ~/.config/brightdata-cli/ |
| Windows | %APPDATA%\brightdata-cli\ |
Two files are stored:
credentials.json — API keyconfig.json — zones, output format, preferencesPriority order (highest → lowest):
CLI flags → Environment variables → config.json → Defaults
| Variable | Description |
|---|---|
BRIGHTDATA_API_KEY | API key (overrides stored credentials) |
BRIGHTDATA_UNLOCKER_ZONE | Default Web Unlocker zone |
BRIGHTDATA_SERP_ZONE | Default SERP zone |
BRIGHTDATA_POLLING_TIMEOUT | Default polling timeout in seconds |
BRIGHTDATA_BROWSER_ZONE | Default Scraping Browser zone (default: cli_browser) |
BRIGHTDATA_DAEMON_DIR | Override the directory used for browser daemon socket and PID files |
BRIGHTDATA_API_KEY=xxx BRIGHTDATA_UNLOCKER_ZONE=my_zone \
brightdata scrape https://example.com
Every command supports:
| Mode | Flag | Behavior |
|---|---|---|
| Human-readable | (default) | Formatted table or markdown, with colors |
| JSON | --json | Compact JSON to stdout |
| Pretty JSON | --pretty | Indented JSON to stdout |
| File | -o <path> | Write to file; format inferred from extension |
Auto-detected file formats:
| Extension | Format |
|---|---|
.json | JSON |
.md | Markdown |
.html | HTML |
.csv | CSV |
When stdout is not a TTY, colors and spinners are automatically disabled. Errors go to stderr, data to stdout.
# Extract URLs from search results
brightdata search "nodejs tutorials" --json | jq -r '.organic[].link'
# Scrape and view with a markdown reader
brightdata scrape https://docs.github.com | glow -
# Save scraped content to a file
brightdata scrape https://example.com -f markdown > page.md
# Amazon product data as CSV
brightdata pipelines amazon_product "https://amazon.com/dp/xxx" --format csv > product.csv
# Chain search → scrape
brightdata search "top open source projects" --json \
| jq -r '.organic[0].link' \
| xargs brightdata scrape
brightdata pipelines list # See all types in your terminal
| Type | Platform |
|---|---|
amazon_product | Amazon product page |
amazon_product_reviews | Amazon reviews |
amazon_product_search | Amazon search results |
walmart_product | Walmart product page |
walmart_seller | Walmart seller profile |
ebay_product | eBay listing |
bestbuy_products | Best Buy |
etsy_products | Etsy |
homedepot_products | Home Depot |
zara_products | Zara |
google_shopping | Google Shopping |
| Type | Platform |
|---|---|
linkedin_person_profile | LinkedIn person |
linkedin_company_profile | LinkedIn company |
linkedin_job_listings | LinkedIn jobs |
linkedin_posts | LinkedIn posts |
linkedin_people_search | LinkedIn people search |
crunchbase_company | Crunchbase |
zoominfo_company_profile | ZoomInfo |
| Type | Platform |
|---|---|
instagram_profiles | Instagram profiles |
instagram_posts | Instagram posts |
instagram_reels | Instagram reels |
instagram_comments | Instagram comments |
facebook_posts | Facebook posts |
facebook_marketplace_listings | Facebook Marketplace |
facebook_company_reviews | Facebook reviews |
facebook_events | Facebook events |
tiktok_profiles | TikTok profiles |
tiktok_posts | TikTok posts |
tiktok_shop | TikTok shop |
tiktok_comments | TikTok comments |
x_posts | X (Twitter) posts |
youtube_profiles | YouTube channels |
youtube_videos | YouTube videos |
youtube_comments | YouTube comments |
reddit_posts | Reddit posts |
| Type | Platform |
|---|---|
google_maps_reviews | Google Maps reviews |
google_play_store | Google Play |
apple_app_store | Apple App Store |
reuter_news | Reuters news |
github_repository_file | GitHub repository files |
yahoo_finance_business | Yahoo Finance |
zillow_properties_listing | Zillow |
booking_hotel_listings | Booking.com |
Error: No Web Unlocker zone specified
brightdata config set default_zone_unlocker <your-zone-name>
# or
export BRIGHTDATA_UNLOCKER_ZONE=<your-zone-name>
Error: Invalid or expired API key
brightdata login
Error: Access denied
Check zone permissions in the Bright Data control panel.
Error: Rate limit exceeded
Wait a moment and retry. Use --async for large jobs to avoid timeouts.
Async job is too slow
brightdata pipelines amazon_product <url> --timeout 1200
# or
export BRIGHTDATA_POLLING_TIMEOUT=1200
No active browser session "default"
# Start a session first
brightdata browser open https://example.com
Browser daemon won't start
# Check if a stale socket file exists and clear it
brightdata browser close
# Then retry
brightdata browser open https://example.com
Element ref not found after interaction
Refs are re-assigned on every snapshot call. If you navigate or click (which may cause the page to change), take a fresh snapshot before using refs again:
brightdata browser click e3
brightdata browser snapshot --compact # refresh refs
brightdata browser type e5 "text"
Garbled output in non-interactive terminal
Colors and spinners are disabled automatically when not in a TTY. If you still see ANSI codes, add | cat at the end of your command.
© Bright Data · ISC License
FAQs
Command-line interface for Bright Data. Scrape, search, extract structured data, and automate browsers directly from your terminal.
The npm package @brightdata/cli receives a total of 297 weekly downloads. As such, @brightdata/cli popularity was classified as not popular.
We found that @brightdata/cli demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 7 open source maintainers collaborating on the project.
Did you know?

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Research
A supply chain attack on Axios introduced a malicious dependency, plain-crypto-js@4.2.1, published minutes earlier and absent from the project’s GitHub releases.

Research
Malicious versions of the Telnyx Python SDK on PyPI delivered credential-stealing malware via a multi-stage supply chain attack.

Security News
TeamPCP is partnering with ransomware group Vect to turn open source supply chain attacks on tools like Trivy and LiteLLM into large-scale ransomware operations.