Total 31,167 skills, AI & Machine Learning has 5044 skills
Showing 12 of 5044 skills
This guide applies when designing, writing, or structuring AI courses, tutorials, lectures, and hands-on projects. It is also to be used when users request to create syllabi, write lecture notes, or design coding exercises related to AI/ML/LLM topics.
Claudeception is a continuous learning system that extracts reusable knowledge from work sessions. Triggers: (1) /claudeception command to review session learnings, (2) "save this as a skill" or "extract a skill from this", (3) "what did we learn?", (4) After any task involving non-obvious debugging, workarounds, or trial-and-error discovery. Creates new Claude Code skills when valuable, reusable knowledge is identified.
AI image generation for paid ad creatives. Reads campaign-brief.md and brand-profile.json to produce platform-sized ad images using Gemini (default) or a configured provider. Requires GOOGLE_API_KEY or ADS_IMAGE_PROVIDER + matching key. Triggers on: "generate ads", "create images", "make ad creatives", "generate visuals", "create ad images", "generate campaign images", "make the images", "generate from brief".
MindOS Knowledge Base Operation Guide (Chinese) for Agent tasks on local markdown/csv knowledge bases. It should be automatically triggered whenever tasks involve note files, SOP/workflow documents, profile/context documents, CSV tables, knowledge base organization, cross-Agent handover or decision synchronization, and are executed via the MindOS MCP tool. Typical requests include "update notes", "search knowledge base", "organize files", "execute SOP", "review according to team standards", "hand over tasks to another Agent", "synchronize decisions", "append to CSV", "retrospect this conversation", "extract key experiences", "adaptively update retrospective results to corresponding documents", "route this information to corresponding files", "synchronously update all related documents", etc.; it should be triggered even if the user does not explicitly mention MindOS.
Provides Codex CLI delegation workflows for complex code generation and development tasks using OpenAI's GPT-5.3-codex models, including English prompt formulation, execution flags, sandbox modes, and safe result handling. Use when the user explicitly asks to use Codex for complex programming tasks such as code generation, refactoring, or architectural analysis. Triggers on "use codex", "delegate to codex", "run codex cli", "ask codex", "codex exec", "codex review".
Use this skill to interact with Moorcheh, the Universal Memory Layer for Agentic AI. Provides semantic search with ITS (Information-Theoretic Scoring), namespace management, text and vector data operations, and AI-powered answer generation (RAG). Use when building applications that need semantic search, knowledge bases, document Q&A, AI memory systems, or retrieval-augmented generation.
Ultimate AI agent memory system for Cursor, Claude, ChatGPT & Copilot. WAL protocol + vector search + git-notes + cloud backup. Never lose context again. Vibe-coding ready.
Use when the user says "/plan-execute", "plan execute", "implement plan", or "execute plan" and provides a finalized plan file path to carry out. Claude orchestrates, Codex writes code, Claude reviews, and Codex fixes issues until the quality bar is met.
Delegate coding tasks to Codex CLI for execution, or discuss implementation approaches with it. CodeX is a cost-effective, strong coder — great for batch refactoring, code generation, multi-file changes, test writing, and multi-turn implementation tasks. Use when the plan is clear and needs hands-on coding. Claude handles architecture, strategy, copywriting, and ambiguous problems better.
Use when users want to maintain persistent memory across sessions, track user preferences, store important decisions, manage tasks and reminders, or provide personalized service with cross-session context.
Generate images, videos, audio, and 3D models via RunningHub API (170+ endpoints) and run any RunningHub AI Application (custom ComfyUI workflow) by webappId. Covers text-to-image, image-to-video, text-to-speech, music generation, 3D modeling, image upscaling, AI apps, and more.
General web search and content extraction skill. It supports multi-source parallel search (WebSearch, MCP search tools, ctx7, agent-browser), web page main content extraction (defuddle/WebFetch) and structured article analysis. This skill is used when users need to search for information, research topics, find materials, obtain web content, read articles, or analyze web pages. Trigger scenarios include: search, research, investigation, fetch, check for me, help me find, read this link, analyze this article. Even if the user doesn't explicitly say "search", this skill should be triggered as long as it involves information acquisition and web content processing.