Loading...
Loading...
Found 62 Skills
Generate or remediate documentation with human-quality writing and style adherence. Use when creating new documentation, rewriting AI-generated content, or applying style profiles. Do not use for slop detection only (use slop-detector) or learning styles (use style-learner).
Implements stories with code and tests.
Use after planning completes and before any implementation begins. Verifies that the current phase contract, story map, and bead graph are structurally sound, execution-ready, and aligned with locked decisions. Use for prompts like "validate the plan", "is this ready to build?", "check the bead graph", "verify execution readiness", or whenever there is doubt about whether the current phase is safe to execute.
Check whether a design is complete enough to move into implementation planning. Use when a design appears mostly done and needs a final readiness review for missing branches, weak assumptions, unresolved risks, failure handling, validation gaps, or non-functional omissions. Trigger before invoking writing-plans or when the user asks whether the current design is ready to implement. Do not use as a general design-document audit for external docs or as a replacement for initial design work.
Test coverage, code quality, defect metrics, and QA KPIs
FORGE Autopilot — Intelligent autonomous mode. FORGE analyzes the project state, automatically decides the next action, and orchestrates all agents until completion. Configurable checkpoints for human review. Usage: /forge-auto or /forge-auto "specific objective"
BAZDMEG Method workflow checkpoint system for AI-assisted development. Enforce quality gates at three phases: pre-code, post-code, and pre-PR. Use when: (1) starting a new feature or bug fix, (2) finishing AI-generated code before review, (3) preparing a pull request, (4) running a planning interview, (5) auditing automation readiness, (6) preventing AI slop, (7) session bootstrap, (8) source rank, (9) domain gates, (10) bugbook. Triggers: 'bazdmeg', 'pre-code checklist', 'post-code checklist', 'pre-PR checklist', 'planning interview', 'quality gates', 'session bootstrap', 'source rank', 'domain gates', 'bugbook'.
Worker that runs existing tests to catch regressions. Auto-detects framework, reports pass/fail. No status changes or task creation.
Final code review and quality gate — run tests, check coverage, audit security, verify acceptance criteria from spec, and generate ship-ready report. Use when user says "review code", "quality check", "is it ready to ship", "final review", or after /deploy completes. Do NOT use for planning (use /plan) or building (use /build).
Evidence-based test debugging enforcing systematic root cause analysis. Use when tests are failing, pytest errors occur, test suite not passing, debugging test failures, or fixing broken tests. Prevents assumption-based fixes by enforcing proper diagnostic sequence. Works with Python (.py), JavaScript/TypeScript (.js/.ts), Go, Rust test files. Supports pytest, jest, vitest, mocha, go test, cargo test, and other frameworks.
Detects orphaned code (files/functions that exist but are never imported or called in production), preventing "created but not integrated" failures. Use before marking features complete, before moving ADRs to completed, during code reviews, or as part of quality gates. Triggers on "detect orphaned code", "find dead code", "check for unused modules", "verify integration", or proactively before completion. Works with Python modules, functions, classes, and LangGraph nodes. Catches the ADR-013 failure pattern where code exists and tests pass but is never integrated.
Full code review, fix, quality, PR workflow. Chains review-branch, address-review, check-quality, and pr. Use when: code complete and ready for PR, want comprehensive review before shipping.