Loading...
Loading...
Found 67 Skills
Normalize terminology across a draft (canonical terms + synonym policy) without changing citations or meaning. **Trigger**: terminology, glossary, consistent terms, 术语统一, 统一叫法, 术语表. **Use when**: the draft has concept drift (same thing called 2–3 names) or global-review flags terminology inconsistency. **Skip if**: you are still changing the outline/taxonomy heavily (do that first). **Network**: none. **Guardrail**: do not add/remove citation keys; do not introduce new claims; avoid moving citations across subsections.
Use this skill when optimizing e-commerce sites for search engines - product page SEO, faceted navigation crawl control, category taxonomy, product schema markup, pagination handling, inventory-aware SEO (out-of-stock pages), and e-commerce site architecture. Triggers on any task involving online store search optimization, product listing pages, shopping search results, or e-commerce technical SEO challenges.
Neta API community skill — browse interactive feeds, view collection details, like and interact with content, and browse content by tags and characters in a community context. Use this skill when the user wants to “see what people are making”, “scroll the feed”, or “interact with works”. Do not use it for taxonomy/keyword‑level research (handled by neta-suggest) or for generating images/videos/songs (handled by neta-creative).
Systematic web application QA testing with structured issue taxonomy, health scoring, and regression tracking. Use this skill when the user asks for QA testing, systematic testing, smoke testing, regression testing, web app testing, browser testing, or says "QA this", "test the app", "smoke test", "run QA", "systematic test", "check the site", "regression test", "full QA", "/qa-systematic". Supports full, quick, and regression modes.
Overcome LLM knowledge cutoffs with real-time developer content. daily.dev aggregates articles from thousands of sources, validated by community engagement, with structured taxonomy for precise discovery.
Catalog GitHub starred repositories into a structured Obsidian vault with AI-synthesized summaries, normalized topic taxonomy, graph-optimized wikilinks, and Obsidian Bases (.base) index files for filtered views. Fetches repo metadata and READMEs via gh CLI, classifies repos into categories and normalized topics, generates individual repo notes with frontmatter, and creates hub notes for categories/topics/authors that serve as graph-view connection points. Use this skill when users want to: (1) Catalog or index their GitHub stars into Obsidian (2) Create a searchable knowledge base from starred repos (3) Organize and discover patterns in their GitHub stars (4) Export GitHub stars as structured markdown notes (5) Build a graph of starred repos by topic, language, or author For saving/distilling a specific URL to a note, use kcap instead. For browsing AI tweets, use ai-twitter-radar instead.
Set up, audit, and debug analytics tracking implementation — GA4, Google Tag Manager, event taxonomy, conversion tracking, and data quality. Use when building a tracking plan from scratch, auditing existing analytics for gaps or errors, debugging missing events, or setting up GTM. Trigger keywords: GA4 setup, Google Tag Manager, GTM, event tracking, analytics implementation, conversion tracking, tracking plan, event taxonomy, custom dimensions, UTM tracking, analytics audit, missing events, tracking broken. NOT for analyzing marketing campaign data — use campaign-analytics for that. NOT for BI dashboards — use product-analytics for in-product event analysis.
Use when a security incident has been detected or declared and needs classification, triage, escalation path determination, and forensic evidence collection. Covers SEV1-SEV4 classification, false positive filtering, incident taxonomy, and NIST SP 800-61 lifecycle.
When the user wants to set up product analytics -- including event taxonomy, tracking plans, funnel analysis, or tool selection (Mixpanel, Amplitude, PostHog). Also use when the user says "event tracking," "analytics setup," "tracking plan," "analytics implementation," or "user identification." For PLG metrics, see plg-metrics. For experimentation, see growth-experimentation.
Guide for EVM/solidity smart contract security work: vulnerability taxonomy, review workflow, and where to place resources in README.md.
Read production traces, identify what's failing, and build failure taxonomies using open coding and axial coding methodology. Use when debugging agent or pipeline quality, investigating "why are my outputs bad?", or before building any evaluator — error analysis must come first. Do NOT use when you already have identified failure modes and need evaluators (use build-evaluator) or datasets (use generate-synthetic-dataset).
Standardized artifact creation via tk tickets. Use whenever a skill needs to persist output — research findings, plans, postmortems, reviews, design specs, decisions. Replaces all bespoke output directories (.oracle/, .plans/, etc.) with a single canonical system.