Loading...
Loading...
Found 10 Skills
Invokes Codex CLI as a second opinion. Use for reviewing plans, code, architectural decisions, or getting an independent perspective from OpenAI's reasoning models.
Repo Updater - Multi-repo synchronization with AI-assisted review orchestration. Parallel sync, agent-sweep for dirty repos, ntm integration, git plumbing. 17K LOC Bash CLI.
Use the @steipete/oracle CLI to bundle a prompt plus the right files and get a second-model review (API or browser) for debugging, refactors, design checks, or cross-validation.
Review and resolve PR issues with Qodo - get AI-powered code review issues and fix them interactively (GitHub, GitLab, Bitbucket, Azure DevOps)
Generate production-ready REVIEW.md and AGENTS.md files for Devin Review's AI code review system. Use this skill whenever the user mentions Devin Review, REVIEW.md, Devin code review setup, PR review instructions for Devin, AI code review configuration for Devin, or wants to create instruction files that Devin's Bug Catcher uses. Also trigger when someone says "set up Devin review", "configure Devin for our repo", "create review rules for Devin", or asks about REVIEW.md / AGENTS.md — even if they don't say "Devin" explicitly but describe wanting AI-powered PR review instructions that work with Devin's auto-review or Bug Catcher.
Orchestrate comprehensive code review across ~12 AI reviewers. 5 persona reviewers (Grug, Carmack, Ousterhout, Beck, Fowler) via Moonbridge, 4 domain specialists (security-sentinel MANDATORY, performance, data integrity, architecture) via Task, plus hindsight-reviewer and synthesis. Use when: code review, PR review, pre-merge quality check.
Run Codex, Claude Code, and Gemini CLI reviews against the current branch concurrently, deduplicate the findings, and fix only the review comments that are still valid for the current codebase.
Use this skill when a user asks to review a pull request for bugs, wants AI code review focused on correctness issues, or runs /bug-review. Trigger on PR review, bug finding, code review, "review this PR", "check for bugs", "find issues in this PR". This is a multi-pass review workflow with 5 parallel passes, majority voting, independent Opus validation, and resolution rate tracking. Also trigger on /bug-review:resolve to classify whether findings were fixed at merge time, and /bug-review:report for resolution rate stats. Even if the user just says "review this" while on a PR branch, trigger this skill.
Generate production-ready Greptile AI code review configuration for any repository. Use this skill whenever the user mentions Greptile, AI code review setup, PR review configuration, automated code review rules, or wants to set up .greptile/ config files. Also trigger when someone says "set up code review", "configure PR reviews", "add review rules", or asks about Greptile configuration — even if they don't say "Greptile" explicitly but describe wanting AI-powered PR review automation. This skill analyzes the actual repository structure and produces tailored config, not generic boilerplate.
Umbrella skill for agent work discipline across development, analysis, and documentation: inspect the repo before restructuring, keep durable truth in repo artifacts instead of chat memory, co-evolve specs/design/steering/user docs with code, apply sound coding patterns, verify work honestly, avoid shortcuts, work efficiently with subagents without hallucinating, and keep moving through the next concrete work item when the human is away. References cover coding patterns, AI-authored code review, and artifact co-evolution. Trigger when the user asks for workflow discipline, coding patterns, doc/artifact maintenance, code review of AI-authored code, project hygiene, execution guardrails, repo normalization, or when a task risks drifting across architecture, storage, specs, continuity, or tooling boundaries.