Loading...
Loading...
Found 6 Skills
Reviews and improves Claude Code skills against official best practices. Supports three modes - self-review (validate your own skills), external review (evaluate others' skills), and auto-PR (fork, improve, submit). Use when checking skill quality, reviewing skill repositories, or contributing improvements to open-source skills.
Use this tool to review, audit, or validate the quality and cross-platform/cross-agent compatibility of Claude Code skills. It is triggered by phrases such as "审查 skill", "review skill", "检查 skill 质量", "skill 兼容性检查", "review 兼容性"
Universal skill reviewer: SKILL mode (D1-D9 + M1-M5) or COMMAND mode (.claude/commands review)
We review Claude Code Skills based on official best practices and provide specific improvement suggestions. It is triggered by requests such as 'Review this skill', 'Check skill quality', 'Validate SKILL.md', 'I want to improve this skill'.
Review skills in any project using a dual-axis method: (1) deterministic code-based checks (structure, scripts, tests, execution safety) and (2) LLM deep review findings. Use when you need reproducible quality scoring for `skills/*/SKILL.md`, want to gate merges with a score threshold (for example 90+), or need concrete improvement items for low-scoring skills. Works across projects via --project-root.
Iteratively reviews and fixes Claude Code skill quality issues until they meet standards. Runs automated fix-review cycles using the skill-reviewer agent. Use to fix skill quality issues, improve skill descriptions, run automated skill review loops, or iteratively refine a skill. Triggers on 'fix my skill', 'improve skill quality', 'skill improvement loop'. NOT for one-time reviews—use /skill-reviewer directly.