Skip to main content

Browser Tool

The Browser tool enables your agents to extract data from websites, perform web scraping, and interact with web pages. It supports multiple scraping engines including Firecrawl for advanced features or Beautiful Soup for simple HTML parsing.
This tool has Default status, meaning it’s production-ready and available on all subscription plans.

Overview

The Browser tool transforms your agents into web scraping specialists capable of:

Web Scraping

Extract data from any publicly accessible website

JavaScript Rendering

Handle dynamic content and single-page applications with Firecrawl

Flexible Extraction

Extract content in Markdown or HTML formats

Multiple Engines

Choose between Firecrawl API and Beautiful Soup engines

Configuration Parameters

scraper
select
required
The scraping engine to use for web data extraction
Options:
  • firecrawl - Professional scraping service with JavaScript rendering and Markdown/HTML extraction
  • beautifulsoup - Python-based HTML parsing for simple text extraction
render
boolean
default:"false"
Enable JavaScript rendering for dynamic content (Firecrawl only)
When enabled, extracts both Markdown and HTML formats. When disabled, extracts Markdown only.
Note: This parameter is ignored when using Beautiful Soup.

Setup Instructions

1

Navigate to Tools

Go to the Tools section in your project dashboard
2

Create Browser Tool

Click Create Tool and select Browser
3

Select Scraper Engine

Choose between Firecrawl API for advanced features or Beautiful Soup for simple scraping
4

Configure JavaScript Rendering

If using Firecrawl, decide whether to enable the render parameter for JavaScript rendering
5

Test Scraping

Use the test button to verify scraping functionality with a sample URL
6

Add to Agent

Assign this tool to your agents in agent settings

Scraper Engines

Firecrawl

Professional scraping service with JavaScript rendering capabilities:
  • JavaScript Rendering: Full browser rendering for dynamic content and SPAs
  • Markdown Extraction: Clean Markdown output from web pages
  • HTML Extraction: Raw HTML output when render is enabled
  • Async Processing: Webhook support for long-running scraping jobs
  • API Key Required: Requires Firecrawl API key configuration

Beautiful Soup

Python-based HTML parsing for lightweight text extraction:
  • Fast Processing: Lightweight HTML parsing
  • Simple Text Extraction: Extracts plain text from HTML
  • No API Key: No external dependencies or authentication required
  • Local Processing: Data processed internally without external calls
  • Low Overhead: Minimal resource usage

Best Practices

Choosing the Right Scraper

  • Use Firecrawl with render: true for JavaScript-heavy sites, single-page applications, or when you need clean Markdown output
  • Use Firecrawl with render: false for faster scraping when only Markdown format is needed
  • Use Beautiful Soup for simple static HTML pages where speed and no external API calls are priorities

Ethical Scraping

Respect Website Terms: Always review and comply with website terms of service and robots.txt files before scraping.
  • Respect Rate Limits: Don’t overwhelm servers with requests
  • Use Public Data: Only scrape publicly accessible information
  • Attribution: Credit sources when republishing scraped content
  • Privacy Compliance: Respect user privacy and data protection laws

Data Protection

  • GDPR Compliance: Handle EU personal data appropriately
  • CCPA Compliance: Respect California privacy rights
  • Data Minimization: Collect only necessary information
  • Secure Storage: Protect scraped data appropriately
  • Access Controls: Limit access to scraped data
  • Fair Use: Understand fair use limitations
  • Attribution: Provide proper source attribution
  • Commercial Use: Review commercial usage rights
  • Content Licensing: Respect content licensing terms
  • DMCA Compliance: Handle takedown requests properly

Next Steps