Loading...
Loading...
Found 13 Skills
Record and analyze post-trade outcomes for signals generated by edge pipeline and other skills. Track false positives, missed opportunities, and regime mismatches. Feed results back to edge-signal-aggregator weights and skill improvement backlog.
Applies targeted improvements to an existing pm-skills skill based on feedback, validation reports, or convention changes. Reads current files, previews proposed changes, writes on confirmation, and suggests a version bump. Use when improving a skill after validation or feedback.
Analyze the current session and propose improvements to skills. **Proactively invoke this skill** when you notice user corrections after skill usage, or at the end of skill-heavy sessions. Also use when user says "reflect", "improve skill", or "learn from this".
피치 스킬 사용 중 발견된 문제점/노하우를 구조화하여 docs/스킬피드백/에 문서화하는 범용 스킬. "스킬 개선", "피드백 정리", "문제점 기록", "스킬 리뷰", "개선사항", "스킬 피드백" 키워드로 트리거. 모든 피치 스킬에 범용 적용 가능. 다른 AI 에이전트가 문서를 읽고 스킬을 개선할 수 있도록 구성.
[Hyper] Optimize an existing Codex skill through baseline-first experiments, binary evals, optional guards, and one-mutation-at-a-time iteration. Use for skill autoresearch, measured trigger/workflow improvement, self-optimizing a skill, benchmarking skill changes, or resuming skill experiment artifacts.
Distill session insights into reusable knowledge: Claude rules, skill improvements, and justfile recipes. Use at the end of a session to capture learnings, update existing artifacts, and avoid reinventing solutions. Prioritizes updating over adding.
Extract learnings about skill creation/improvement from a session and propagate them to the central skill learnings file, then sync to appropriate skills. Use when a session revealed patterns, anti-patterns, or insights about structuring skills. Invoke via /update-skill-learnings or after skill creation/improvement sessions.
Evaluate and improve skills through measured testing. Run trigger evaluations to test whether skill descriptions cause correct activation, optimize descriptions via automated train/test loops, benchmark skill output quality with A/B comparisons, and validate skill structure. Use when user says "improve skill", "test skill triggers", "optimize description", "benchmark skill", "eval skill", or "skill quality". Do NOT use for creating new skills (use skill-creator-engineer).
Analyze Claude Code session transcripts to improve existing skills or create new ones. Use when you want to review a past session to identify what worked, what didn't, and how to enhance skill documentation. Extracts session data and provides structured analysis prompts. Triggers on "improve skill", "analyze session", "review session", "skill improvement", "create skill from session", "skill not working", "skill missed", "skill didn't trigger", "enhance skill", "refine skill", "skill feedback", "session transcript", "what went wrong", "skill optimization", "better triggers".
Analyzes and refines agent skills by identifying quality issues, prioritizing fixes (MUST/SHOULD/NICE), gathering user feedback, and implementing improvements. Checks for common problems like time estimates, oversized SKILL.md files, poor structure, redundant content, missing examples, and unclear workflows. Use when reviewing, improving, refactoring, or auditing existing skills. Triggers include "review skill", "improve skill", "refactor skill", "skill quality", "audit skill", "fix skill", "optimize skill", "analyze skill".
Skills retrospective and improvement. Use when: - User asks to "review", "retrospect", "summarize" or "复盘" skills - User wants to analyze skills issues from the conversation - User requests skills optimization or improvement - End of conversation or after significant skill usage
Analyzes your recent Claude Code chat history to identify coding patterns, development gaps, and areas for improvement, generating a personalized growth report with actionable recommendations.