Loading...
Loading...
Found 63 Skills
Use when you need to generate or update `{FEATURE_DIR}/verification/suites.md` (test suite) during the verification phase of Spec Pack, organize test cases into executable sets and define blocking rules and execution order.
Use when you need to generate or update `{FEATURE_DIR}/verification/usecase.md` (test cases) during the verification phase of the Spec Pack, and require AC traceability and a structure that supports automated script generation.
Guides QA engineers through daily testing activities—morning review, test case creation, automation, exploratory testing, bug reporting, and end-of-day wrap-up. Use when planning or executing day-to-day testing or when the user asks about daily testing workflow.
Design accessibility testing plans and test cases covering WCAG 2.1 standards, screen readers, and keyboard navigation. Default output is Markdown, with support for requesting Excel/CSV/JSON formats. Use for accessibility testing.
Design API test plans and cases for REST/GraphQL/gRPC interfaces. Default output Markdown; can request Excel, CSV, or JSON. Use for API testing.
Sync tests with TestRail. Use when user mentions "testrail", "test management", "test cases", "test run", "sync test cases", "push results to testrail", or "import from testrail".
Smoke test for alicloud-database-analyticdb-mysql. Validate minimal authentication, API reachability, and one read-only query path.
Systematic LLM prompt engineering: analyzes existing prompts for failure modes, generates structured variants (direct, few-shot, chain-of-thought), designs evaluation rubrics with weighted criteria, and produces test case suites for comparing prompt performance. Triggers on: "prompt engineering", "prompt lab", "generate prompt variants", "A/B test prompts", "evaluate prompt", "optimize prompt", "write a better prompt", "prompt design", "prompt iteration", "few-shot examples", "chain-of-thought prompt", "prompt failure modes", "improve this prompt". Use this skill when designing, improving, or evaluating LLM prompts specifically. NOT for evaluating Claude Code skills or SKILL.md files — use skill-evaluator instead.
Universal generation template for OpenHarmony XTS test cases. Supports test case generation for various subsystems, API definition parsing, test coverage analysis, and code specification checking. Trigger keywords: XTS, test generation, case generation, test case.
Maintain JSONL-only profiler performance test cases under csrc/ops/<op>/test in ascend-kernel. Collect data using torch_npu.profiler (with fixed warmup=5 and active=5), aggregate the Total Time(us) from ASCEND_PROFILER_OUTPUT/op_statistic.csv, and output a unified Markdown comparison report (custom operator vs baseline) that includes a DType column. Do not generate perf_cases.json or *_profiler_results.json. Refer to examples/layer_norm_profiler_reference/ for the reference implementation.
Use when adding or updating Go CLI E2E coverage for one `tests/cli_e2e/{domain}` domain of the compiled `lark-cli`, especially when the work requires live `--help` or `schema` exploration, scenario-based `clie2e.RunCmd` workflows, and per-domain `coverage.md` maintenance.
Convert feature requirements into structured end-to-end test cases as JSON.