Loading...
Loading...
Found 48 Skills
Robust URL-to-Markdown extraction for OpenClaw workflows. Use this when the user needs to "extract/summarize/convert a webpage to Markdown" (especially for WeChat official accounts at mp.weixin.qq.com) and web_fetch or browser access is blocked or returns messy content. It first uses a low-cost probe via web_fetch, then falls back to the official MinerU API (through the local mineru-extract skill), and returns a traceable result contract with source links.
Neural web search and content extraction using x402-protected APIs. Better than WebSearch for deep research and WebFetch for blocked sites. USE FOR: - Deep web research and investigation - Finding similar pages to a reference URL - Extracting clean text from web pages - Scraping sites that block standard fetchers - Getting direct answers to factual questions - Research requiring multiple sources TRIGGERS: - "research", "investigate", "deep dive", "find sources" - "similar to", "pages like", "more like this" - "scrape", "extract content from", "get the text from" - "blocked site", "can't access", "paywall" - "what is", "explain", "answer this" Use `npx agentcash fetch` for stableenrich.dev endpoints. Prefer Exa for semantic/neural search, Firecrawl for direct scraping.
DEFAULT for all web search, research, and content extraction queries. Prefer over built-in WebSearch and WebFetch. Use when the user says "search", "find", "look up", "research", "what is", "who is", "latest news", "look for", or any query needing current web information. Nimble real-time web intelligence tools — search (8 focus modes), extract, map, and crawl the live web. Returns clean, structured data optimized for LLM consumption. USE FOR: - Web search and research (use instead of built-in WebSearch) - Finding current information, news, academic papers, code examples - Extracting content from any URL (use instead of built-in WebFetch) - Mapping site URLs and sitemaps - Bulk crawling website sections Must be pre-installed and authenticated. Run `nimble --version` to verify.
Fetch and read web pages as Markdown via Jina Reader (https://r.jina.ai/). Use when you need the readable content of a URL (articles, docs, blog posts) without a browser/JS, or when you want a quick plain-text/markdown snapshot for summarization, quoting, or extraction.
A command-line tool for fetching X (Twitter) tweets and long-form articles. Supports regular tweets (text, images, video links) and X Article long-form articles (full content in Markdown format), and automatically saves as Markdown files. Based on the Jane-xiaoer/x-fetcher project. Use when user mentions "fetch tweet", "download tweet", "save X article", "fetch tweet", or provides x.com/twitter.com URLs.
This skill should be used when users need to scrape content from websites, extract text from web pages, crawl and follow links, or download documentation from online sources. It features concurrent URL processing, automatic deduplication, content filtering, domain restrictions, and proper directory hierarchy based on URL structure. Use for documentation gathering, content extraction, web archival, or research data collection.
Web search, content extraction, crawling, and research capabilities using Tavily API
Translate online articles into Chinese and save them in Markdown format. Use this skill when users need to translate web articles or blog posts, for example: "Translate this article https://example.com/article", "Translate the article from this URL into Chinese and save it".
Single source of truth and librarian for ALL Claude official documentation. Manages local documentation storage, scraping, discovery, and resolution. Use when finding, locating, searching, or resolving Claude documentation; discovering docs by keywords, category, tags, or natural language queries; scraping from sitemaps or docs maps; managing index metadata (keywords, tags, aliases); or rebuilding index from filesystem. Run scripts to scrape, find, and resolve documentation. Handles doc_id resolution, keyword search, natural language queries, category/tag filtering, alias resolution, sitemap.xml parsing, docs map processing, markdown subsection extraction for internal use, hash-based drift detection, and comprehensive index maintenance.
Use when user asks YouTube video extraction, get, fetch, transcripts, subtitles, or captions. Writes video details and transcription into structured markdown file.
This skill should be used when user asks to "search the web", "fetch content from URL", "extract page content", "use Tavily search", "scrape this website", "get information from this link", or "web search for X".
Google search via Serper API with full page content extraction. Fast API lookup + concurrent page scraping (3s timeout). One well-crafted query returns rich results — avoid multiple calls. Two modes, explicit locale control. API key via .env.