firecrawl_crawl
Starts a crawl job on a website and extracts content from all pages. **Best for:** Extracting content from multiple related pages, when you need comprehensive coverage. **Not recommended for:** Extracting content from a single page (use scrape); when token limits are a concern (use map + batch_scrape); when you need fast results (crawling can be slow). **Warning:** Crawl responses can be very large and may exceed token limits. Limit the crawl depth and number of pages, or use map + batch_scrape for better control. **Common mistakes:** Setting limit or maxDiscoveryDepth too high (causes token overflow) or too low (causes missing pages); using crawl for a single page (use scrape instead). Using a /* wildcard is not recommended. **Prompt Example:** "Get all blog posts from the first two levels of example.com/blog." **Usage Example:** ```json { "name": "firecrawl_crawl", "arguments": { "url": "https://example.com/blog/*", "maxDiscoveryDepth": 5, "limit": 20, "allowExternalLinks": false, "deduplicateSimilarURLs": true, "sitemap": "include" } } ``` **Returns:** Operation ID for status checking; use firecrawl_check_crawl_status to check progress. **Safe Mode:** Read-only crawling. Webhooks and interactive actions are disabled for security.
Pricing
Per call
$0.05
Model
flat
Pay only for what you use. No subscriptions.
Inputs
excludePaths
arraycrawlEntireDomain
booleanincludePaths
arrayscrapeOptions
objectdeduplicateSimilarURLs
booleanallowExternalLinks
booleanurl *
stringmaxConcurrency
numberdelay
numbermaxDiscoveryDepth
numberlimit
numberallowSubdomains
booleanignoreQueryParameters
booleanprompt
stringsitemap
string
