Loading...
Loading...
Found 67 Skills
Stay current with how OpenCode, OpenAI Codex, and Claude Code implement extensibility features (skills, slash commands, subagents, custom prompts). Use when comparing implementations across AI coding assistants, researching how a specific tool implements a feature, or syncing knowledge about agent extensibility patterns. Triggers include questions like "how does X implement skills?", "compare slash commands across tools", "what's the latest on Claude Code sub-agents?", or requests to understand agent extensibility approaches.
Unified setup hub: project init, tool setup, 2-agent config, harness-mem, codex CLI, and rule localization. Use when user mentions setup, initialization, new projects, workflow files, CI setup, LSP setup, MCP setup, codex setup, opencode setup, 2-Agent setup, PM coordination, Cursor setup, harness-mem, claude-mem integration, cross-session memory, localize rules, adapt rules. Do NOT load for: implementation work, reviews, build verification, or deployments.
Use the beans CLI to track issues/tasks alongside your code. Flat-file issue tracker that stores beans as markdown files in .beans/ directory. Integrates with Codex, OpenCode, and Claude Code via beans-prime.
Adapter boundary rules for plugin integrations. Trigger: Changes in plugin scripts/hooks for Claude, OpenCode, Gemini, or Codex.
Rust SDK reference for the OpenCode HTTP API and SSE streaming. Use when implementing, debugging, or reviewing code that integrates with opencode-sdk, including client setup, session/message APIs, event streaming, managed server/runtime workflows, feature-flag behavior, and error handling patterns.
Set up and optimize repositories for AI coding agents. Creates minimal AGENTS.md, CLAUDE.md symlink, docs/REQUIREMENTS.md, docs/BUSINESS-RULES.md, feedback loops, and deterministic enforcement (Claude Code hooks, OpenCode plugins). Use when user wants to make a repo AI-friendly, set up AGENTS.md/CLAUDE.md, document requirements/business rules for AI, add pre-commit hooks for AI workflows, or optimize codebase structure for coding agents.
Universal Cross-session Memory Protocol (Universal Memory Protocol). Enable all AI programming tools to share the same memory system. Applicable to Claude Code / Cursor / Aider / Cline / Codex / Trae / OpenCode. Capabilities: Intelligent Classification / FSRS Decay / Monthly Compression / Multi-layer Retrieval. Triggers: User says "remember"; asks "previous"; sensitive information detected; session ends.
Spawn a single autonomous AI agent with a specific task, personality, and CLI backend (Claude, Gemini, OpenCode, Copilot). Agent accepts task from docs/todo/pending/, selects personality based on task type, and works autonomously with CLI tools. Integrates with docs-first workflow via task signals and progress tracking.
Explore and analyze TUI applications to document their features for cloning. Use when asked to reverse-engineer, analyze, document, or understand a terminal UI like Claude Code, OpenCode, Codex, lazygit, or any ratatui/ncurses-based application. Launches the target TUI in tmux, systematically explores all views and keybindings, captures ASCII diagrams of each screen, and writes findings incrementally to a markdown file (survives context compaction).
Cross-tool AI consultation. Use when user asks to 'consult gemini', 'ask codex', 'get second opinion', 'cross-check with claude', 'consult another AI', 'ask opencode', 'copilot opinion', or wants a second opinion from a different AI tool.
Delegate tasks to the cost-effective opencode/glm-5 model. Use when you need inexpensive task execution, simple research, or delegating work that doesn't require the most powerful models.
Critically review terminal user interfaces for UX quality, responsiveness, visual design, and interactivity. Use when asked to "review my TUI", "test my TUI UX", "audit my terminal UI", "check TUI responsiveness", "review TUI keybindings", "check interactivity", or any request to evaluate the user experience quality of a ratatui/crossterm/ncurses-based terminal application. Launches the TUI in tmux, systematically tests 10 dimensions (responsiveness, input conflicts, visual clarity, navigation, feedback loops, error states, layout, keyboard design, permission flows, visual design & color), and produces a graded report with screenshots and specific findings. Benchmarks against Claude Code, OpenCode, and Codex — the three best-in-class AI terminal UIs.