Loading...
Loading...
Found 44 Skills
Points to the BlockchainSpider open-source Python/Scrapy toolkit for collecting on-chain data—transfer subgraphs around an address or tx, EVM and Solana block/transaction ingestion, receipts/logs, and optional label plugins. Use when the user wants to build datasets, offline traces, or research pipelines alongside blockchain-analytics-operations and solana-tracing-specialist—not as a substitute for RPC provider ToS, rate limits, or legal review of sensitive crawls.
Comprehensive news aggregator that fetches, filters, and deeply analyzes real-time content from 8 major sources: Hacker News, GitHub Trending, Product Hunt, 36Kr, Tencent News, WallStreetCN, V2EX, and Weibo. Best for 'daily scans', 'tech news briefings', 'finance updates', and 'deep interpretations' of hot topics.
Install, configure, and operate Scrapling for adaptive web scraping with parser-first HTML extraction, fast HTTP fetchers, browser rendering, stealth anti-bot options, CLI extraction, and optional MCP or spider workflows. Use when you need to scrape or crawl websites, choose between static, JavaScript-rendered, or protected targets, parse HTML with CSS or XPath, write Python scrapers, or run Scrapling from the terminal. Triggers on: scrapling, scrape website, crawl site, adaptive scraping, stealthy fetch, cloudflare scraping, mcp scraping server, browser scraping cli, scrapling spider.
WeChat Official Account Article Aggregator. Batch fetch the latest article list of specified WeChat Official Account authors via the mptext.top API, download article content and parse it into Markdown/HTML/plain text formats. It supports fetching by official account name or fakeid, and comes with 8 popular AI tech official accounts pre-configured. Applicable scenarios: fetch the latest articles of specified official accounts, batch crawl content from multiple authors, aggregated reading of official account articles, technical information collection. Keywords: official account article acquisition, WeChat article crawling, official account aggregation, article collection, mptext API.
Extract structured data from websites. Use when: collecting competitor pricing; scraping product listings; extracting contact information; gathering research data; monitoring website changes
Comprehensive web scraping, crawling, and data extraction toolkit powered by Firecrawl API. Provides scripts for single-page scraping (scrape.py), web search (search.py), URL discovery (map.py), multi-page crawling (crawl.py), structured data extraction (extract.py), and autonomous data gathering (agent.py). Use when you need to: (1) extract content from web pages, (2) search and scrape the web, (3) discover URLs on websites, (4) crawl multiple pages, (5) extract structured data with JSON schemas, or (6) autonomously gather data from anywhere on the web. Requires FIRECRAWL_API_KEY environment variable.
Fetch and parse news highlights from CCTV News Broadcast (Xinwen Lianbo) for a given date.
Crawl entire websites using Cloudflare Browser Rendering /crawl API. Initiates async crawl jobs, polls for completion, and saves results as markdown files. Useful for ingesting documentation sites, knowledge bases, or any web content into your project context. Requires CLOUDFLARE_ACCOUNT_ID and CLOUDFLARE_API_TOKEN environment variables.