Loading...
Loading...
Found 2,122 Skills
Toolkit for creating and validating skills and subagents. Use when: creating a new skill (fast or full mode), validating an existing skill, deciding Skills vs Subagents, migrating docs to skills, estimating token cost, or running a security scan. Triggers: "create skill", "build skill", "validate skill", "new subagent", "skills vs subagents", "estimate tokens", "security scan".
Side role: find and correct bad signals, earn leaderboard points per Publisher-approved correction (max 3/day)
Build and use the verification infrastructure coding agents need to prove their work. Use when: a repo has no bootable dev environment, no real-surface tests, or no interaction layer an agent can use; auditing or grading a repo's agent-readiness; verifying changes work end to end on real surfaces; or when harness gaps block reliable agent output.
Verifica a Stack do N8N. Além disso analisa parâmetros, rotas Traefik, volumes, recursos e conformidade do stack N8N de Acordo com as Recomendações da Promovaweb.
Verifica a Stack do Kestra. Além disso analisa parâmetros, rotas Traefik, volumes, recursos e conformidade do stack Kestra de Acordo com as Recomendações da Promovaweb.
Evaluate the output of a journey-builder run, identify instruction gaps, and edit the project root AGENTS.md (or add pitfalls to the gist) to fix those gaps. Does NOT modify the journey-builder skill itself.
Invoke this skill when a user shares test code and questions whether it actually works as intended — not to run or fix the test, but to evaluate whether the test has real value. Triggers on: "is this test any good?", "would this catch a real bug?", "this test always passes — is that normal?", "review these tests before I commit", or "does this test verify anything meaningful?". Also triggers when someone suspects a test is useless, wants a pre-commit quality gate, or is unsure if an auto-generated test is worth keeping. The core question this skill answers: "Would this test fail if the feature broke?" If not, the test gets rejected. Do NOT use for generating new tests, fixing failing tests, or exploring application features.
Start here — turn source material into a validated agent skill. Orchestrates cogworks-encode (synthesis) and cogworks-learn (skill generation).
Walk through a Jira task plan and interactively confirm assumptions, resolve open questions, and validate decisions — using progressive disclosure. Only asks questions relevant to the CURRENT phase or task being executed. Use when the user says "review the plan", "ask me questions", "clarify assumptions", "let's go through the questions", "grill me on the plan", "validate plan for PROJECT-1234", or anything about reviewing, questioning, or validating a task plan. Also triggered by the orchestrating-jira-workflow skill as Phase 3 of the pipeline, and re-invoked during Phase 5 before each task execution. Requires a task plan at docs/<TICKET_KEY>-tasks.md.
Validate multi-skill workflows defined in CLAUDE.md by checking skill existence, inter-skill data contracts (JSON schema compatibility), file naming conventions, and handoff integrity. Use when adding new workflows, modifying skill outputs, or verifying pipeline health before release.
Critically review strategy drafts from edge-strategy-designer for edge plausibility, overfitting risk, sample size adequacy, and execution realism. Use when strategy_drafts/*.yaml exists and needs quality gate before pipeline export. Outputs PASS/REVISE/REJECT verdicts with confidence scores.
Use after the final approved execution scope is complete, or when the user asks whether a feature is done, ready to ship, safe to merge, or needs a quality check. Runs the post-execution quality gate: specialist review, artifact verification, and human UAT against locked decisions and the final exit state. Use for prompts like "review this feature", "is this done?", "can we ship this?", "double-check the implementation", or "run UAT".