Loading...
Loading...
Found 175 Skills
Summarize current work, commit, push, and create or update a PR. Automatically write conversation context into the PR description to ensure reviewers can quickly understand the background.
Fetch and extract web content as clean Markdown when provided with URLs. Use this skill whenever a user provides a URL (http/https link) that needs to be read, analyzed, summarized, or extracted. Converts web pages to Markdown with 80% fewer tokens than raw HTML. Handles all content types including JS-heavy sites, documentation, articles, and blog posts. Supports three conversion methods (auto, AI, browser rendering). Always use this instead of web_fetch when working with URLs - it's more efficient and provides cleaner output.
Email Gazette: Transform unread inbox emails into a beautiful newspaper-style HTML front page. Use this skill whenever the user asks for an email briefing, inbox summary, email digest, morning briefing, 'what happened in my inbox', 'catch me up on emails', 'summarize my emails', or anything about turning emails into a readable overview. Also trigger when the user mentions 'gazette' or asks for a newspaper-style view of their emails, daily digest, or email newspaper.
Read and summarize text-based file types only. Prefer read_file for text formats; use execute_shell_command for type detection when needed. PDF/Office/images/archives are handled by other skills.
Generate a handoff document after implementation work is complete — summarizes changes, risks, and review focus areas for the review pipeline. Use when done coding and ready to hand off for review.
Use context-mode tools (ctx_execute, ctx_execute_file) instead of Bash/cat when processing large outputs. Triggers: "analyze logs", "summarize output", "process data", "parse JSON", "filter results", "extract errors", "check build output", "analyze dependencies", "process API response", "large file analysis", "page snapshot", "browser snapshot", "DOM structure", "inspect page", "accessibility tree", "Playwright snapshot", "run tests", "test output", "coverage report", "git log", "recent commits", "diff between branches", "list containers", "pod status", "disk usage", "fetch docs", "API reference", "index documentation", "call API", "check response", "query results", "find TODOs", "count lines", "codebase statistics", "security audit", "outdated packages", "dependency tree", "cloud resources", "CI/CD output". Also triggers on ANY MCP tool output that may exceed 20 lines. Subagent routing is handled automatically via PreToolUse hook.
Automatically collect and summarize daily AI industry news, trends, and hot topics from platforms like GitHub (trending repos), X/Twitter (AI influencers/hashtags), and AI news aggregators. Use this skill when the user asks for "today's AI news", "AI industry updates", "what's trending in AI", or wants a daily digest of AI developments.
Generate high-quality Markdown documents such as weekly reports, work reports, summaries, and introductions. When no draft is provided, search and summarize from the Web; when a draft is provided, organize, polish, and supplement based on the draft. Use this when the user mentions weekly reports, work reports, summaries, introductions, debriefings, or reviews.
Help users study research papers with Paper Breakdown. Use when the user wants to study, understand, ask questions about, summarize, or analyze a paper with Paper Breakdown, including requests like "I want to study paper P with Paper Breakdown", "help me read this paper in PaperBD", "use Paper Breakdown for this arXiv paper", or similar requests about looking up a paper, finding its arXiv ID, checking access in the paperbd CLI, and then answering questions about the paper.
Fetch and analyze content from one or more URLs using AI (Gemini 2.5 Flash). Use when you have specific URLs and need to extract or summarize their content. Pairs well with `nansen web search` results.
Analyze codebases from the bottom up and generate a hierarchical README.md document tree. Start analysis from leaf directories, generate README.md files for each directory containing one-sentence descriptions of files, classes, and functions, and summarize layer by layer upwards to form a complete codebase documentation system. Supports state persistence and resumable analysis, suitable for scenarios such as understanding new projects, generating technical documentation, and analyzing code structures. Use this skill when you need to understand codebase structures, analyze function implementations, or generate code documentation.
Turn many commits into a curated grouped squash summary compatible with the opinionated wording style of git-visual-commits. Use this skill whenever the user asks to squash a branch into a concise summary, write a squash-and-merge summary, summarize a commit range or PR as grouped lines, clean up noisy commit history, or asks for a curated summary without committing. Treat phrases like "squash summary", "squash commit message", "summarize this branch", "turn these commits into one summary", "rewrite these 10+ commits", or "draft the squash summary" as automatic triggers. This skill is non-mutating: it inspects git history and diffs, then returns grouped summary lines only. It preserves technical identifiers where possible, groups by intent rather than chronology, merges overlapping commits, drops low-signal noise, uses strong concrete verbs, favors readable GitHub and terminal output, keeps every output line at or below 72 characters, and does not invent unsupported changes or drift into changelog wording.