Total 31,643 skills, AI & Machine Learning has 5103 skills
Showing 12 of 5103 skills
A valid skill with version field for testing
A valid skill that uses $ARGUMENTS placeholder
Build, run, and visualize multi-step AI generation workflows. The AI architect translates natural language descriptions into connected node graphs — chain image generation, video creation, enhancement, and editing into automated pipelines.
Market salary research, accomplishment quantification, negotiation scripts, total compensation analysis, timing strategy.
Use this skill to manage already-installed skills across Claude Code, Codex, Gemini, OpenCode, OpenClaw, Cursor, Copilot, and other configured agent tools by comparing skill status and linking from configured source directories such as ~/.cc-switch/skills/ and ~/.agents/skills/. Trigger it in two major cases: first, when the user wants to sync, remove, repair, or align skills or agent skills across multiple agents; second, when the user does not yet know the current skill state and wants to inspect skill differences, missing skills, per-agent skill coverage, per-skill coverage, or decide what skill changes to make next. Use this skill when the topic is cross-agent skill or agent-skill management, not for general agent comparison, general model capability questions, or creating, editing, or installing skills from GitHub.
Delegate coding tasks to Google Jules AI agent for asynchronous execution. Use when user says: 'have Jules fix', 'delegate to Jules', 'send to Jules', 'ask Jules to', 'check Jules sessions', 'pull Jules results', 'jules add tests', 'jules add docs', 'jules review pr'. Handles: bug fixes, documentation, features, tests, refactoring, code reviews. Works with GitHub repos, creates PRs.
Analyzes and generates llms.txt files -- the emerging standard for helping AI systems understand website structure and content. Can validate existing llms.txt files or generate new ones from scratch by crawling the site.
Convert raster images (photos, illustrations, AI-generated art) into high-quality SVG recreations. Breaks the image into isolated features, builds each as a standalone SVG layer, then composites them. Use when the user wants to recreate an image as SVG, create vector versions of artwork, or extract specific elements from images as scalable graphics.
Paper Workflow: Read papers and create reading cards in one go. Accepts one or more arXiv links, paper URLs, PDFs, or paper titles. For each paper, it runs ljg-paper (generates org-format analysis) followed by ljg-card -l (generates long-form reading card PNG). Trigger this workflow when the user says '论文流', 'paper flow', '读论文并做卡片', '论文卡片', or provides multiple papers and requires both analysis and reading cards.
This skill implements a specific task from a project's ROADMAP.md file. It should be used when the user wants to work on a roadmap action item by its ID (e.g., '1.1', '2.3'). Triggered by requests like '/do-task 1.1', '/do-task 2.3', or 'do task 3.1'. Works alongside the project-init skill (which creates the roadmap) and the checkpoint skill (which commits afterward).
This skill applies when OpenStoryline has been installed, and the user needs to start local MCP/Web services, create or continue a session, send editing instructions, perform multi-round re-editing, verify rendered video outputs, or make Chinese requests such as "启动 OpenStoryline", "把 OpenStoryline 跑起来", "用 OpenStoryline 剪视频".
Build Gradio web UIs and demos in Python. Use when creating or editing Gradio apps, components, event listeners, layouts, or chatbots.