Loading...
Loading...
Found 16 Skills
Fetch web page content via Chrome DevTools Protocol (CDP). Full JS rendering, handles redirects (including Google News). Use when you need to read the text content of a web page, scrape articles, or extract information from URLs. Zero dependencies — Python 3 stdlib only. Cross-platform (Mac, Windows, Linux).
Fetches web content with intelligent content extraction, converting HTML to clean markdown. Use for documentation, articles, and reference pages http/https URLs.
Fetch and analyze content from one or more URLs using AI (Gemini 2.5 Flash). Use when you have specific URLs and need to extract or summarize their content. Pairs well with `nansen web search` results.
Fetch and extract readable content from web pages. Use for lightweight page access without browser automation.
Fetch and analyze content from one or more URLs using AI (Gemini 2.5 Flash). Use when you have specific URLs and need to extract or summarize their content. Pairs well with `nansen web search` results.
Web content fetching and URL retrieval via curl and WebFetch — replaces the Fetch MCP server (fetch_html, fetch_json, fetch_markdown, fetch_txt). Use this skill when a specific URL is provided and the user wants its content. Covers HTTP GET/POST, JSON API consumption with jq, HTML retrieval, markdown conversion, plain text extraction, authenticated requests, redirects, cookies, and timeouts. Trigger phrases: "fetch this URL", "get the page content", "download HTML", "call this API", "curl this endpoint", "grab the JSON", "fetch markdown from", "retrieve web content", "hit this endpoint", "scrape this page", "read this URL", "pull data from API", "make an HTTP request", "extract page content", "get article text". NOT for web searches without a URL — use tavily for that.
Fetch and parse web content. Extract readable content, download files, scrape structured data from URLs. No API key required for basic functionality.
Fetch and read web pages, APIs, and online content. Use when users share URLs or ask about web content.
Fetch any URL as clean markdown. ALWAYS use this skill instead of the WebFetch tool when you need to read a URL's content — it has a 5-layer fallback (Jina Reader, defuddle.md, markdown.new, OpenCLI, raw HTML) that produces better results and handles JS-rendered pages (Twitter/X, SPAs), login-required platforms (zhihu, reddit, weibo, xiaohongshu), and complex web pages that WebFetch cannot parse. Invoke whenever the user provides a URL and wants to read, extract, summarize, analyze, or convert its content to markdown. Keywords: 'fetch page', 'read URL', 'grab content from', 'summarize article', 'extract text from webpage', '抓取网页', '读链接', '网页转 markdown'. NOT for: web search without URL, file downloads, screenshots, form filling, or accessibility checks.
Resolve queries or URLs into compact, LLM-ready markdown using a low-cost cascade. Prioritizes llms.txt for structured docs, uses web fetch/search tools for extraction. Use when you need to fetch documentation, resolve web URLs to markdown, search for technical content, or build context from web sources.
Fetches Andrej Karpathy's curated RSS feed, reads ALL original articles, and generates a themed daily briefing document with citation links. This skill should be used when the user asks to generate a Karpathy RSS daily report, fetch Karpathy's curated RSS updates, or create an AI/tech daily briefing based on Karpathy's sources. Trigger phrases include "Karpathy Daily Report", "Karpathy RSS", "Generate Daily Report", "RSS Briefing", "Karpathy Daily".
Track all URLs fetched during SpecStory AI coding sessions. Run when user says "show my link trail", "what URLs did I visit", "list fetched links", or "show web fetches".