Total 44,046 skills, Code Quality has 2066 skills
Showing 12 of 2066 skills
Review a spec or concrete code changes and report evidence-backed bugs, regressions, and risks.
Use when debugging bugs, test failures, build failures, performance regressions, or unexpected behavior and you need root-cause investigation before proposing fixes. Trigger on requests to debug, investigate why something broke, or find the source of a technical issue.
Comprehensive multi-perspective review using specialized judges with debate and consensus building
Adversarial code review with anti-trust verification. Verifies every implementer claim against the actual diff. Use when reviewing PRs, branches, or recent commits.
Systematic debugging with hypothesis-driven investigation. Use when diagnosing bugs, errors, or unexpected behavior. Phases: Reproduce, Hypothesize, Investigate, Fix, Verify, Regression.
Research the codebase to answer questions. Search for relevant files, gather context, and provide clear answers.
Refactor code to simplify it without changing behavior.
Implement one task or scoped change: make the change, add valuable tests, and verify it works.
Reviews code for quality — architecture conformance, anti-patterns, performance issues, maintainability. Read-only analysis that detects circular dependencies, N+1 queries, dead code, naming violations, and layering breaches. Use when the user asks for a code review, wants feedback on code quality, PR review, tech debt analysis, or architecture conformance checks.
Use when the user asks to review code, review changes, review a commit, review a PR, audit code quality, check for security issues, or generate a code review report. Trigger on phrases like "review my changes", "코드 리뷰", "check my code", "review the last commit", "what do you think of this diff", "compare branches", "code audit" — even if they don't say "code review" explicitly. For persistent file output use `code-review-md` (markdown) or `code-review-html` (markdown + HTML).
Validate whether an implementation matches its stated goal. Use this skill when a skill or agent wants a second opinion on its own output, when the user says "check this implementation", "validate what you did", "is this correct?", "review the output", or "did you do this right?". Also spawned automatically as a subagent by other skills (memory-bridge, daily-update) to self-check their outputs before presenting to the user. Returns a structured pass/warn/fail verdict with specific actionable issues.
Language-specific verification for Python, TypeScript/JavaScript, and Go. Checks type safety, language idioms, and best practices. Use when asked to "verify language", "check types", or for language-specific checks.