Loading...
Loading...
Found 7,493 Skills
Claims-based authorization for agents and operations. Grant, revoke, and verify permissions for secure multi-agent coordination. Use when: permission management, access control, secure operations, authorization checks. Skip when: open access, no security requirements, single-agent local work.
CopilotKit integration patterns for providers, runtime wiring, `useCoAgent`, `useCopilotAction`, `useLangGraphInterrupt`, shared state, and HITL with LangGraph. Use when building agent-native product UX.
An AI Agent Skill that enforces a 'Risk Triage -> Align -> Act' protocol. Triggers when requests are vague, conflict-ridden, or high-impact. Do not activate for low-risk or precise, scoped requests.
Produce an LLM Build Pack (prompt+tool contract, data/eval plan, architecture+safety, launch checklist). Use for building with LLMs, GPT/Claude apps, prompt engineering, RAG, and tool-using agents.
Converts a Refound/Lenny Skill into a high-density, agent-executable Skill Pack (Agent Skills standard). Output must be in English.
Fetch and apply Cursor-style workspace rules supporting all rule formats (.cursor/rules/*.md, *.mdc, AGENTS.md, and legacy .cursorrules).
Add an external skill from a GitHub repository to the current workspace. Use when users want to import, install, or add a skill from a GitHub URL (e.g., `/add-skill https://github.com/OpenHands/skills/tree/main/skills/codereview` or "add the codereview skill from https://github.com/OpenHands/skills/"). Handles fetching the skill files and placing them in .agents/skills/.
AI Agent long-term memory system with cross-session, cross-project persistence. Triggers: - /remember - Store memories - /recall - Search memories - /forget - Delete/archive memories - /memory-status - Check status - When needing to persist important conversation insights - When sharing user preferences across projects
Transform extracted engineer expertise into an actionable skill with progressive disclosure, allowing agents to find and apply relevant patterns for specific tasks.
Test, validate, and improve agent instructions (CLAUDE.md, system prompts) using sub-agents as experiment subjects. Measures instruction compliance, context decay, and constraint strength. Use for "test prompt", "validate instructions", "prompt effectiveness", "instruction decay", or when designing robust agent behaviors.
Activate when user provides a prompt, SKILL.md, or agent instruction and requests optimization. Transforms weak instructions into reliable, enforceable agent protocols.
GitHub Discussion CLI for AI agents. Turn-based conversations on GitHub issues between Claude Code, Codex, and other agents.