Loading...
Loading...
Found 256 Skills
Web crawling and scraping with analysis. Use for crawling websites, security scanning, and extracting information from web pages.
Web content extraction via Jina AI Reader API. Three modes: read (URL to markdown), search (web search + full content), ground (fact-checking). Extracts clean content without exposing server IP.
Batch-process web pages via headless Playwright browser, extract HTML, convert to markdown using Turndown, and save to timestamped scratchpad file. Use when user asks to "capture these pages as markdown", "save web content", "fetch and convert webpages", or needs clean markdown from HTML. All URLs from one prompt → single file at docs/web-captures/<timestamp>.md.
Web scraping, search, and data extraction using Firecrawl API. Use when users need to fetch web content, discover URLs on sites, search the web, or extract structured data from pages.
Fetch and read web pages as Markdown via Jina Reader (https://r.jina.ai/). Use when you need the readable content of a URL (articles, docs, blog posts) without a browser/JS, or when you want a quick plain-text/markdown snapshot for summarization, quoting, or extraction.
Run the TikSpyder tool to collect TikTok data — search by keyword, username, or hashtag, download videos, extract keyframes, and export structured data. Use this skill whenever the user wants to collect TikTok data, search TikTok profiles or hashtags, search TikTok videos, download TikTok content, run tikspyder, or launch the TikSpyder Streamlit interface. Also trigger when the user mentions data collection from TikTok, even if they don't say "tikspyder" by name.
Unofficial Tavily CLI skill by SYXS for search, extract, crawl, map, and research. USE FOR: - Web research and source discovery - URL content extraction - Website mapping and crawling - Tavily research + follow-up status workflows Must be installed and authenticated. See rules/install.md and rules/security.md.
Scrape any webpage as clean markdown via Bright Data Web Unlocker API. Bypasses bot detection and CAPTCHA. Requires BRIGHTDATA_API_KEY and BRIGHTDATA_UNLOCKER_ZONE environment variables.
通过 Chrome CDP 抓取任意 URL 并转换为 Markdown。支持两种模式——页面加载后自动抓取,或等待用户信号(适用于需要登录的页面)。当用户要求将网页保存为 Markdown 时使用。
Extract and parse article content from web sources. Retrieves text, metadata, and structured information from articles while preserving formatting and context.
Process Obsidian daily notes: classify raw URLs and loose ideas, fetch content (X tweets, GitHub repos, web pages), run deep research on ideas, create structured vault notes, replace raw items with wikilinks. Orchestrates doc-obsidian, res-x, and res-deep skills. Use when: processing daily note links, digesting saved URLs into notes, turning ideas into research, daily note cleanup. Triggers: daily digest, process daily, daily links, triage daily, digest daily note.
Fetch and extract web content as clean Markdown when provided with URLs. Use this skill whenever a user provides a URL (http/https link) that needs to be read, analyzed, summarized, or extracted. Converts web pages to Markdown with 80% fewer tokens than raw HTML. Handles all content types including JS-heavy sites, documentation, articles, and blog posts. Supports three conversion methods (auto, AI, browser rendering). Always use this instead of web_fetch when working with URLs - it's more efficient and provides cleaner output.