Loading...
Loading...
Found 18 Skills
Iteratively reviews and fixes Claude Code skill quality issues until they meet standards. Runs automated fix-review cycles using the skill-reviewer agent. Use to fix skill quality issues, improve skill descriptions, run automated skill review loops, or iteratively refine a skill. Triggers on 'fix my skill', 'improve skill quality', 'skill improvement loop'. NOT for one-time reviews—use /skill-reviewer directly.
Comprehensive multi-dimensional skill reviews across structure, content, quality, usability, and integration. Task-based operations with automated validation, manual assessment, scoring rubrics, and improvement recommendations. Use when reviewing skills, ensuring quality, validating production readiness, identifying improvements, or conducting quality assurance.
Reviews and grades an agent skill directory (SKILL.md plus supporting resources) for specification compliance, clarity, token efficiency, safety, robustness, and portability. Use when a user wants a rubric-based critique with a weighted score/grade and concrete, minimal patch suggestions.
Analyze and optimize document quality issues of other Skills, including checks on frontmatter format, progressive disclosure structure, file reference integrity, AI-friendliness, etc. When to use: When users say "optimize this skill", "check skill quality", "review skill", "there's something wrong with the skill", "help me look at this skill", "diagnose the skill"
Review and analyze a skill against best practices for length, intent scope, and trigger patterns
Create, review, and iterate on Claude Code skills using Anthropic's official best practices and latest API documentation. Use when creating skills, reviewing existing skills, writing skill descriptions, designing skill architecture, or when user says "create a skill", "review my skill", "skill best practices", "skill description help". Do NOT use for creating plugins, commands, agents, or hooks - use plugin-dev skills for those.