firecrawl-download

Compare original and translation side by side

🇺🇸

Original

English
🇨🇳

Translation

Chinese

firecrawl download

firecrawl download

Experimental. Convenience command that combines
map
+
scrape
to save an entire site as local files.
Maps the site first to discover pages, then scrapes each one into nested directories under
.firecrawl/
. All scrape options work with download. Always pass
-y
to skip the confirmation prompt.
实验性功能。这是一个便捷命令,结合了
map
scrape
功能,可将整个网站保存为本地文件。
它会先对网站进行映射以发现所有页面,然后将每个页面抓取到
.firecrawl/
目录下的嵌套文件夹中。所有抓取选项均适用于此download命令。请始终添加
-y
参数以跳过确认提示。

When to use

适用场景

  • You want to save an entire site (or section) to local files
  • You need offline access to documentation or content
  • Bulk content extraction with organized file structure
  • 想要将整个网站(或某一板块)保存为本地文件
  • 需要离线访问文档或内容
  • 以结构化的文件结构批量提取内容

Quick start

快速开始

bash
undefined
bash
undefined

Interactive wizard (picks format, screenshots, paths for you)

交互式向导(会为你选择格式、截图设置和路径)

firecrawl download https://docs.example.com
firecrawl download https://docs.example.com

With screenshots

包含截图

firecrawl download https://docs.example.com --screenshot --limit 20 -y
firecrawl download https://docs.example.com --screenshot --limit 20 -y

Multiple formats (each saved as its own file per page)

多种格式(每个页面的每种格式单独保存为一个文件)

firecrawl download https://docs.example.com --format markdown,links --screenshot --limit 20 -y
firecrawl download https://docs.example.com --format markdown,links --screenshot --limit 20 -y

Creates per page: index.md + links.txt + screenshot.png

每个页面会生成:index.md + links.txt + screenshot.png

Filter to specific sections

过滤到特定板块

firecrawl download https://docs.example.com --include-paths "/features,/sdks"
firecrawl download https://docs.example.com --include-paths "/features,/sdks"

Skip translations

跳过翻译版本

firecrawl download https://docs.example.com --exclude-paths "/zh,/ja,/fr,/es,/pt-BR"
firecrawl download https://docs.example.com --exclude-paths "/zh,/ja,/fr,/es,/pt-BR"

Full combo

完整组合命令

firecrawl download https://docs.example.com
--include-paths "/features,/sdks"
--exclude-paths "/zh,/ja"
--only-main-content
--screenshot
-y
undefined
firecrawl download https://docs.example.com
--include-paths "/features,/sdks"
--exclude-paths "/zh,/ja"
--only-main-content
--screenshot
-y
undefined

Download options

下载选项

OptionDescription
--limit <n>
Max pages to download
--search <query>
Filter URLs by search query
--include-paths <paths>
Only download matching paths
--exclude-paths <paths>
Skip matching paths
--allow-subdomains
Include subdomain pages
-y
Skip confirmation prompt (always use in automated flows)
选项描述
--limit <n>
最多下载的页面数量
--search <query>
根据搜索关键词过滤URL
--include-paths <paths>
仅下载匹配指定路径的页面
--exclude-paths <paths>
跳过匹配指定路径的页面
--allow-subdomains
包含子域名页面
-y
跳过确认提示(在自动化流程中请始终使用)

Scrape options (all work with download)

抓取选项(所有选项均适用于download命令)

-f <formats>
,
-H
,
-S
,
--screenshot
,
--full-page-screenshot
,
--only-main-content
,
--include-tags
,
--exclude-tags
,
--wait-for
,
--max-age
,
--country
,
--languages
-f <formats>
,
-H
,
-S
,
--screenshot
,
--full-page-screenshot
,
--only-main-content
,
--include-tags
,
--exclude-tags
,
--wait-for
,
--max-age
,
--country
,
--languages

See also

另请参阅

  • firecrawl-map — just discover URLs without downloading
  • firecrawl-scrape — scrape individual pages
  • firecrawl-crawl — bulk extract as JSON (not local files)
  • firecrawl-map — 仅发现URL而不进行下载
  • firecrawl-scrape — 抓取单个页面
  • firecrawl-crawl — 批量提取为JSON格式(非本地文件)