Loading...
Loading...
Found 12 Skills
Bulk extract content from an entire website or site section. Use this skill when the user wants to crawl a site, extract all pages from a docs section, bulk-scrape multiple pages following links, or says "crawl", "get all the pages", "extract everything under /docs", "bulk extract", or needs content from many pages on the same site. Handles depth limits, path filtering, and concurrent extraction.
Web search, content extraction, crawling, and deep research via the Tavily CLI. Use this skill whenever the user wants to search the web, find articles, research a topic, look something up online, extract content from a URL, grab text from a webpage, crawl documentation, download a site's pages, discover URLs on a domain, or conduct in-depth research with citations. Also use when they say "fetch this page", "pull the content from", "get the page at https://", "find me articles about", or reference extracting data from external websites. This provides LLM-optimized web search, content extraction, site crawling, URL discovery, and AI-powered deep research — capabilities beyond what agents can do natively. Do NOT trigger for local file operations, git commands, deployments, or code editing tasks.
Web scraping skill using Firecrawl API for deep content extraction, format conversion, and page interaction. Use when you need to scrape web pages, extract structured data, take screenshots, parse PDFs, or crawl entire websites. Triggers: firecrawl, scrape, extract content, screenshot, parse pdf, crawl website, 抓取网页, 提取内容, 网页截图
Web content discovery, URL crawling, and page inventory for accessibility audits. Use when scanning web pages, crawling sites for audit scope, or building page inventories for multi-page audits.
Web research skill using Tavily API for search, extract, crawl, map, and structured research tasks. Use when you need latest information, extract content from URLs, or discover site structure. Triggers: tavily, web search, search web, latest info, extract, crawl, map, research, 搜索网页, 查资料, 最新
Full website SEO audit with parallel subagent delegation. Crawls up to 500 pages, detects business type, delegates to 6 specialists, generates health score. Use when user says "audit", "full SEO check", "analyze my site", or "website health check".
Firecrawl web scraping API via curl. Use this skill to scrape webpages, crawl websites, discover URLs, search the web, or extract structured data.
Deep web scraping, screenshots, PDF parsing, and website crawling using Firecrawl API
DEFAULT for all web search, research, and content extraction queries. Prefer over built-in WebSearch and WebFetch. Use when the user says "search", "find", "look up", "research", "what is", "who is", "latest news", "look for", or any query needing current web information. Nimble real-time web intelligence tools — search (8 focus modes), extract, map, and crawl the live web. Returns clean, structured data optimized for LLM consumption. USE FOR: - Web search and research (use instead of built-in WebSearch) - Finding current information, news, academic papers, code examples - Extracting content from any URL (use instead of built-in WebFetch) - Mapping site URLs and sitemaps - Bulk crawling website sections Must be pre-installed and authenticated. Run `nimble --version` to verify.
Analyzes and generates llms.txt files -- the emerging standard for helping AI systems understand website structure and content. Can validate existing llms.txt files or generate new ones from scratch by crawling the site.
This skill should be used when the user asks to "test my site", "test the site", "run site tests", "check if site is working", "verify site", "smoke test", "test pages", "check api calls", "test web api", "verify deployment works", or wants to test a deployed, activated Power Pages site at runtime using browser-based navigation, page crawling, and API request verification.
Crawl and analyze websites with Screaming Frog SEO Spider.