Total 30,580 skills, AI & Machine Learning has 4942 skills
Showing 12 of 4942 skills
This skill should be used when the user asks about new features, recent changes, or updates in Claude Code — for example "what's new in Claude Code?", "Claude Code changelog", "what did I miss in Claude?", "any recent updates?", "tell me about new Claude features", or "what's changed since version 1.0.30?". It fetches the official changelog, filters for notable features (excluding bug fixes), researches each feature for deeper context on Anthropic's website, and presents mini-article summaries. Supports both automatic tracking (since last check) and explicit version queries.
Expert Suno AI prompt engineering for cinematic, transformative music creation. Use this skill when creating Suno prompts for Vibe OS sessions, meditation tracks, or any AI-generated music that needs professional quality and emotional resonance.
Auditing memory traces and debugging.
Create and configure custom OpenCode agents (primary and subagents) with specialized prompts, tools, permissions, and models. Use when the user wants to create, modify, or configure OpenCode agents, or mentions agent modes, tool permissions, or task delegation.
Create subagent definitions for Claude Code and OpenCode that delegate to skills. Use when creating new subagents or refactoring existing ones to follow the delegation pattern.
AI image generation CLI using Gemini. Use when generating images, checking syntax for resolution, aspect ratio, and reference image options.
A skill that creates 'conversation flow' nodes for vox.ai voice agents. It converts call-center/OB/CS scripts (free text) into a flow-node format with `name + content + transition conditions`. Use this skill whenever users ask to transform call scripts, consultation scenarios, OB scripts, CS manuals, call flows, consultation flows, or conversation-flow nodes into step-based prompts/nodes/flows. It is also triggered by requests such as 'split this script into steps,' 'turn this into flow nodes,' or 'organize this with transition conditions.'
This skill should be used when the agent needs to give a spoken voice update to the user, or when reminded by a Stop hook to provide audio feedback. Use this skill to speak a short summary of what was accomplished.
Configure a Stop hook that surfaces unfinished todos before a session ends and suggests creating GitHub issues for deferred work. Use when you want unfinished Claude Code session tasks automatically flagged for GitHub issue creation at session end.
Launch and manage Codex Cloud tasks from the CLI, including detached background watchers that track completion. Use when users ask to run coding work in cloud/background agents, queue multiple cloud tasks, poll task status, fetch cloud diffs, apply cloud outputs locally, or pair cloud kickoff with `$cas` orchestration.
Perform image understanding and analysis using GLM Vision MCP. Trigger conditions: (1) Users request image analysis, image understanding, or description of image content (2) Need to identify objects, text, or scenes in images (3) Use GLM's visual understanding capabilities
Replace with a description of what this skill does and when Claude should use it. Include keywords that help agents discover and activate this skill.