Loading...
Loading...
Found 707 Skills
Expert guidance for designing Azure solutions using Azure Architecture. Covers reference architectures, solution ideas, design patterns, technology choices, architecture styles, best practices, anti-patterns, example workloads, and migration guides. Use when selecting architecture patterns, choosing Azure services, or implementing production-ready solutions.
Systematic LLM prompt engineering: analyzes existing prompts for failure modes, generates structured variants (direct, few-shot, chain-of-thought), designs evaluation rubrics with weighted criteria, and produces test case suites for comparing prompt performance. Triggers on: "prompt engineering", "prompt lab", "generate prompt variants", "A/B test prompts", "evaluate prompt", "optimize prompt", "write a better prompt", "prompt design", "prompt iteration", "few-shot examples", "chain-of-thought prompt", "prompt failure modes", "improve this prompt". Use this skill when designing, improving, or evaluating LLM prompts specifically. NOT for evaluating Claude Code skills or SKILL.md files — use skill-evaluator instead.
Audits and enhances API documentation for FastAPI and REST endpoints. Identifies missing descriptions, incomplete response codes, missing examples, and generates enhanced docstrings, Pydantic model examples, and OpenAPI spec improvements. Triggers on: "generate API docs", "document this API", "OpenAPI for", "add examples to", "improve docstrings", "API documentation audit", "FastAPI docs", "document endpoints", "API reference", "swagger docs", "REST API docs", "endpoint documentation", "response documentation". Use this skill when API endpoints need documentation or documentation audit.
Deep code simplification, refactoring, and quality refinement. Analyzes structural complexity, anti-patterns, and readability debt, then applies targeted refactoring preserving exact behavior. Language-agnostic: Python, Go, TypeScript/JavaScript, Rust. Use this skill when the goal is simplification and clarity rather than bug-finding. Triggers on: "simplify this code", "clean up my code", "refactor for clarity", "reduce complexity", "make this more readable", "code quality pass", "tech debt cleanup", "run the code refiner", "simplify recent changes", "this code is messy", "too much nesting", "this function is too long", "clean this up before I PR it", "tidy up my code", cyclomatic complexity, cognitive complexity, code smells.
Distill Opus-level reasoning into optimized instructions for Haiku 4.5 (and Sonnet). Generates explicit, procedural prompts with n-shot examples that maximize smaller model performance on a given task. Use when user says "down-skill", "distill for Haiku", "optimize for Haiku", "make this work on Haiku", "generate Haiku instructions", or needs to delegate a task to a smaller model with high reliability.
DEFAULT search tool for ALL search/lookup needs. Multi-source search and deduplication layer with intent-aware scoring. Integrates Brave Search (web_search), Exa, Tavily, and Grok to provide high-coverage, high-quality results. Automatically classifies query intent and adjusts search strategy, scoring weights, and result synthesis. Use for ANY query that requires web search — factual lookups, research, news, comparisons, resource finding, "what is X", status checks, etc. Do NOT use raw web_search directly; always route through this skill.
Use when the user wants to review a pull request, understand what a PR changes, assess risk of merging, or check for missing test coverage. Examples: "Review this PR", "What does PR #42 change?", "Is this PR safe to merge?"
GitHub Research Assistant. Use this skill when the user wants to analyze a GitHub repository. Analysis dimensions -- 1) Basic information; 2) Purpose, what it can be used for; 3) Tech stack, including frameworks, languages, algorithms, etc.; 4) Usage and examples; 5) Technical architecture and module analysis
Establish a naming convention system for design elements, components, and tokens with clear rules and examples.
Optimizes AI skills for activation, clarity, and cross-model reliability. Use when creating or editing skill packs, diagnosing weak skill uptake, reducing regressions, tuning instruction salience, improving examples, shrinking context cost, or setting benchmark/release gates for skills. Trigger terms: skill optimization, activation gap, benchmark skill, with/without skill delta, regression, context budget, prompt salience.
Structure a performance review with self-assessment, manager template, and calibration prep. Use when review season kicks off and you need a self-assessment template, writing a manager review for a direct report, prepping rating distributions and promotion cases for calibration, or turning vague feedback into specific behavioral examples.
Create videos from a text prompt using HeyGen's Video Agent. Use when: (1) Creating a video from a description or idea, (2) Generating explainer, demo, or marketing videos from a prompt, (3) Making a video without specifying exact avatars, voices, or scenes, (4) Quick video prototyping or drafts, (5) One-shot prompt-to-video generation, (6) User says "make me a video" or "create a video about X".