10 Clawbrowser CLI Tips You're Probably Not Using
The Clawbrowser CLI has a short learning curve for basic use. The depth reveals itself over time. Here are ten features worth knowing now.
1. Chain commands with pipes
Extract from multiple pages and pipe results together:
claw extract "product name and price" \
--url https://shop.example.com/category/keyboards \
--array --json | \
jq '[.[] | select(.price < 100)]'Extract → filter with jq → feed to next step. Standard Unix pipelines work.
2. Use --dry-run to preview actions
Before running a destructive or expensive command, preview what Clawbrowser will do:
claw batch extract "product details" \
--urls products.txt \
--dry-runOutput:
DRY RUN — no requests will be made
URLs to process: 247
Estimated time at 30/min: ~8 minutes
Rate limit: 30/minute
Profile: defaultNo requests made. Just the plan.
3. --verbose for debugging
When extraction isn't working as expected, --verbose shows the full execution trace:
claw extract "price" \
--url https://shop.example.com/product/123 \
--verbose[1] Launching browser context (profile: default)
[2] Navigating to https://shop.example.com/product/123
[3] Waiting for DOM stability...
[4] DOM stable after 1240ms (47 mutations settled)
[5] Running extraction: "price"
[6] Candidate elements found: 3
[7] Best match: .product-price (confidence: 0.94)
[8] Extracted value: "$89.99"See exactly what Clawbrowser did and why.
4. Named output formats
Control output format for different pipeline needs:
# JSON (default structured)
claw extract "product details" --url ... --json
# CSV (for spreadsheets)
claw extract "name, price, rating" --url ... --csv
# TSV (for some data pipelines)
claw extract "name, price, rating" --url ... --tsv
# Plain text (for LLM input)
claw extract "main article content" --url ... --text5. Screenshot on failure
When an extraction fails, a screenshot tells you why:
claw extract "price" \
--url https://shop.example.com/product/123 \
--screenshot-on-failure ./debug/If extraction fails, ./debug/failure-1234.png shows you exactly what the browser saw — CAPTCHA, login wall, error page.
6. Profile cloning
Create a new profile pre-populated from an existing one:
# Alice and Bob need similar fingerprints but separate sessions
claw profile clone --from base-profile --to alice
claw profile clone --from base-profile --to bob
# Now they share fingerprint config but have isolated cookies/storage7. Batch processing with concurrency control
claw batch extract "product name and price" \
--urls products.txt \
--json \
--concurrency 5 \ # 5 parallel browser contexts
--rate-limit 30/min \ # total across all contexts
--output results.jsonConcurrency and rate limiting are independent. 5 parallel contexts hitting 30 requests/minute total — Clawbrowser distributes the capacity.
8. Wait for specific content
Instead of timeouts, wait for meaningful signals:
# Wait for specific text to appear
claw navigate https://app.example.com/dashboard \
--wait-for-text "Welcome back"
# Wait for a specific element
claw navigate https://app.example.com/dashboard \
--wait-for "account balance widget"
# Wait for network to go quiet
claw navigate https://app.example.com/dashboard \
--wait-for-idle 500msEach targets a different readiness signal. Use what fits the page.
9. Environment variables for secrets
Don't put credentials in scripts:
# Set once in your environment
export CLAW_PROXY_URL="http://user:pass@proxy.provider.com:8080"
export CLAW_DEFAULT_PROFILE="production"
# CLI picks them up automatically
claw navigate https://target.com
# Uses $CLAW_DEFAULT_PROFILE and $CLAW_PROXY_URLFull list of env vars: claw env --list
10. Config files for shared pipeline settings
For repeatable pipelines, put shared settings in .clawrc:
# .clawrc
default_profile: production
rate_limit: 30/minute
delay_jitter: 25%
screenshot_on_failure: ./debug/
output_format: json
timeout: 30sAll commands in the directory pick up these defaults. No flags needed for common settings.
These ten features don't change what Clawbrowser can do — they change how much effort it takes to do it at scale. Worth five minutes to set up, saves hours over a production pipeline's lifetime.
Full CLI reference: claw --help or the docs.
Ready to start?
Give your agent a browser
Clawbrowser handles fingerprints, proxies, and sessions. Your agent focuses on the task.