Loading...
Loading...
Found 65 Skills
Run quality gates on implemented feature: tests, lint, type checks, and custom validation. Use after /feature-implement completes. Use for "validate feature", "run quality gates", "check feature", or "/feature-validate". Do NOT use for ad-hoc linting or debugging.
Collaborative coding with enforced micro-steps: announce, show diff, wait for confirmation, apply, verify. User controls pace with commands. Works with any domain agent as the executor. Use when: "pair program", "pair with me", "let's code together", "step by step coding", "walk me through implementing", "code with me"
Run the full spec-driven workflow automatically. Proposes, implements, verifies, reviews, and archives a change with one mandatory proposal checkpoint plus any extra confirmations required by blocking conditions.
Read-only drift detector. Diffs SPEC.md against current code and reports violations grouped by severity. Writes nothing — suggests remedies via the spec or build skills but never invokes them. Triggers when the user asks to check drift, audit the spec, verify invariants, or ask whether code still matches the spec. Phrasings: "check drift", "audit the spec", "does the code still match §V", "check invariants", "spec vs code".
프로젝트의 모든 verify 스킬을 순차 실행하여 통합 검증 보고서를 생성합니다. 기능 구현 후, PR 전, 코드 리뷰 시 사용.
Database schema and migration workflow. Use when modifying database schemas in tools/atlas/ or SQLC queries in tools/sqlc/.
Phase-level implementation workflow for builder agents. Handles reading phase files, finding references, invoking domain skills, implementing all steps, and running verification (tests + typecheck). Preloaded into builder agents via skills: field — not user-invocable.
Test-driven development workflow with philosophy guide - plan → write tests → implement → validate
Verify completed implementation against the plan. Checks that all tasks were fully implemented, nothing was forgotten, code compiles, tests pass, and quality standards are met. Use after "/aif-implement" completes, or when user says "verify", "check work", "did we miss anything".
Use when you have a written implementation plan to execute with review checkpoints
Code quality gatekeeper and auditor. Enforces strict quality gates, resolves the AI verification gap, and evaluates codebases across 12 critical dimensions with evidence-based scoring. Use when auditing code quality, reviewing AI-generated code, scoring codebases against industry standards, or enforcing pre-commit quality gates. Use for quality audit, code review, codebase evaluation, security assessment, technical debt analysis.
TDD Skill for Team Workflows (Phase 3). Before writing any implementation code, you must first write a failing test and verify the failure, then write the minimal implementation until it passes, and finally refactor and re-test.