首页/数据分析/firecrawl-download
F

firecrawl-download

by @firecrawlv1.0.0
0.0(0)

将整个网站下载为本地文件,支持Markdown和截图格式,便于离线查阅和分析,实现内容存档。

File DownloadingWeb Content RetrievalFirecrawl APIData AcquisitionResource ManagementGitHub
安装方式
npx skills add firecrawl/cli --skill firecrawl-download
compare_arrows

Before / After 效果对比

1
使用前

在没有Firecrawl Download之前,下载整个网站的内容通常需要手动逐页保存,或者使用简单的下载工具,但往往会遇到链接断裂、格式不统一、无法保留网站结构等问题。这使得离线查阅或备份变得非常耗时且不完整。

使用后

使用Firecrawl Download后,可以一键将整个网站下载为本地文件,支持Markdown、截图等多种格式,并能保持网站的原始结构。这极大地提高了内容获取的效率和质量,方便了离线使用和文档归档。

description SKILL.md


name: firecrawl-download description: | Download an entire website as local files — markdown, screenshots, or multiple formats per page. Use this skill when the user wants to save a site locally, download documentation for offline use, bulk-save pages as files, or says "download the site", "save as local files", "offline copy", "download all the docs", or "save for reference". Combines site mapping and scraping into organized local directories. allowed-tools:

  • Bash(firecrawl *)
  • Bash(npx firecrawl *)

firecrawl download

Experimental. Convenience command that combines map + scrape to save an entire site as local files.

Maps the site first to discover pages, then scrapes each one into nested directories under .firecrawl/. All scrape options work with download. Always pass -y to skip the confirmation prompt.

When to use

  • You want to save an entire site (or section) to local files
  • You need offline access to documentation or content
  • Bulk content extraction with organized file structure

Quick start

# Interactive wizard (picks format, screenshots, paths for you)
firecrawl download https://docs.example.com

# With screenshots
firecrawl download https://docs.example.com --screenshot --limit 20 -y

# Multiple formats (each saved as its own file per page)
firecrawl download https://docs.example.com --format markdown,links --screenshot --limit 20 -y
# Creates per page: index.md + links.txt + screenshot.png

# Filter to specific sections
firecrawl download https://docs.example.com --include-paths "/features,/sdks"

# Skip translations
firecrawl download https://docs.example.com --exclude-paths "/zh,/ja,/fr,/es,/pt-BR"

# Full combo
firecrawl download https://docs.example.com \
  --include-paths "/features,/sdks" \
  --exclude-paths "/zh,/ja" \
  --only-main-content \
  --screenshot \
  -y

Download options

OptionDescription
--limit <n>Max pages to download
--search <query>Filter URLs by search query
--include-paths <paths>Only download matching paths
--exclude-paths <paths>Skip matching paths
--allow-subdomainsInclude subdomain pages
-ySkip confirmation prompt (always use in automated flows)

Scrape options (all work with download)

-f <formats>, -H, -S, --screenshot, --full-page-screenshot, --only-main-content, --include-tags, --exclude-tags, --wait-for, --max-age, --country, --languages

See also

forum用户评价 (0)

发表评价

效果
易用性
文档
兼容性

暂无评价,来写第一条吧

统计数据

安装量0
评分0.0 / 5.0
版本1.0.0
更新日期2026年3月16日
对比案例1 组

用户评分

0.0(0)
5
0%
4
0%
3
0%
2
0%
1
0%

为此 Skill 评分

0.0

兼容平台

🔧Claude Code
🔧OpenClaw
🔧OpenCode
🔧Codex
🔧Gemini CLI
🔧GitHub Copilot
🔧Amp
🔧Kimi CLI

时间线

创建2026年3月16日
最后更新2026年3月16日