Loading...
Loading...
Found 19 Skills
This skill should be used when generating comprehensive test cases from PRD documents or user requirements. Triggers when users request test case generation, QA planning, test scenario creation, or need structured test documentation. Produces detailed test cases covering functional, edge case, error handling, and state transition scenarios.
Eval enablement accelerator — help customers think through "what does good look like" for their AI agent, then generate a structured eval plan and test cases they can use immediately. No running agent required. Works from a description, an idea, or even a vague goal. Use when anyone mentions agent evaluation, eval planning, "what should we test", "how do we know if the agent is good", test case generation, or interpreting eval results.
Generates eval test cases from an eval suite plan (output of /eval-suite-planner) or a plain-English agent description. Supports both single-response and conversation (multi-turn) evaluation modes. Outputs a Copilot Studio test set table, a CSV file for import (single-response only), and a docx report for human review.
Test Case Generator - Based on the theories of Equivalence Partitioning and Boundary Value Analysis, generates high-quality test cases in batches by Test Points (POINT), output in Markdown format. Used when users execute the /testcase-gen command or need to generate test cases.
Website exploration for testing using Playwright MCP
NeuroForge QA is a QA/UX review system grounded in the 30 Laws of UX and QA engineering standards. Works with ANY framework, language, or software — React, Vue, iOS, Android, APIs, wireframes, or plain descriptions. On activation it scans the project and creates (or reads existing) files in a /neuroforge/ folder: project analysis, UX audit, risk register, accessibility audit, and test cases in /neuroforge/test-cases/. Treats these files as single source of truth, updating incrementally. Trigger on: "review my UI", "audit this design", "write test cases", "check my UX", "QA this flow", "critique my wireframe", "write tests for", "find bugs in", any screenshot shared for feedback, or any request for QA or UX analysis of a product, screen, flow, or codebase. When in doubt, trigger.
Debug regex patterns with visual breakdowns, plain English explanations, test case generation, and flavor conversion. Use when user needs help with regular expressions or pattern matching.
Property-based testing with fast-check (TypeScript/JavaScript) and Hypothesis (Python). Generate test cases automatically, find edge cases, and test mathematical properties. Use when user mentions property-based testing, fast-check, Hypothesis, generating test data, QuickCheck-style testing, or finding edge cases automatically.
DEPRECATED: Use testing-best-practices instead. This skill has been retired.
Combinatorial testing with a greedy pairwise matrix generator. Covers all factor pairs in near-minimal test cases.
Use when you need to generate or update `{FEATURE_DIR}/verification/usecase.md` (test cases) during the verification phase of the Spec Pack, and require AC traceability and a structure that supports automated script generation.
Universal generation template for OpenHarmony XTS test cases. Supports test case generation for various subsystems, API definition parsing, test coverage analysis, and code specification checking. Trigger keywords: XTS, test generation, case generation, test case.