Total 44,264 skills, AI & Machine Learning has 7037 skills
Showing 12 of 7037 skills
Guides technology selection and implementation of AI and ML features in .NET 8+ applications using ML.NET, Microsoft.Extensions.AI (MEAI), Microsoft Agent Framework (MAF), GitHub Copilot SDK, ONNX Runtime, and OllamaSharp. Covers the full spectrum from classic ML through modern LLM orchestration to local inference. Use when adding classification, regression, clustering, anomaly detection, recommendation, LLM integration (text generation, summarization, reasoning), RAG pipelines with vector search, agentic workflows with tool calling, Copilot extensions, or custom model inference via ONNX Runtime to a .NET project. DO NOT USE FOR projects targeting .NET Framework (requires .NET 8+), the task is pure data engineering or ETL with no ML/AI component, or the project needs a custom deep learning training loop (use Python with PyTorch/TensorFlow, then export to ONNX for .NET inference).
Claudeception is a continuous learning system that extracts reusable knowledge from work sessions. Triggers: (1) /claudeception command to review session learnings, (2) "save this as a skill" or "extract a skill from this", (3) "what did we learn?", (4) After any task involving non-obvious debugging, workarounds, or trial-and-error discovery. Creates new Claude Code skills when valuable, reusable knowledge is identified.
Product photography enhancement for ad creatives using AI image generation. Takes a product image and generates 5 professional photography styles for ad use: Studio, Floating, Ingredient, In Use, and Lifestyle. Requires GOOGLE_API_KEY or configured ADS_IMAGE_PROVIDER. Triggers on: "product photo", "product photography", "photoshoot", "enhance product image", "product shoot", "product photos for ads", "generate product photos", "studio shot", "lifestyle photo".
Detect crisis signals in user content using NLP, mental health sentiment analysis, and safe intervention protocols. Implements suicide ideation detection, automated escalation, and crisis resource integration. Use for mental health apps, recovery platforms, support communities. Activate on "crisis detection", "suicide prevention", "mental health NLP", "intervention protocol". NOT for general sentiment analysis, medical diagnosis, or replacing professional help.
Generate AI videos from text prompts using the HeyGen API. Use when: (1) Generating videos from text descriptions, (2) Creating AI-generated video clips for content production, (3) Image-to-video generation with a reference image, (4) Choosing between video generation providers (VEO, Kling, Sora, Runway, Seedance), (5) Working with HeyGen's /v1/workflows/executions endpoint for video generation.
Swap faces in a video using AI via the HeyGen API. Use when: (1) Replacing a face in a video with another face, (2) Face swapping from a source image onto a target video, (3) Creating personalized videos by swapping in a person's face, (4) Working with HeyGen's /v1/workflows/executions endpoint for face swap processing.
Agent-IM Conversation Skill - Create sessions, send messages such as image/video generation requests via OpenAPI, and query session progress. This skill is activated when users need to generate images/videos or query current session messages.
ORCA v1.1 — Hardened dual-mode emerging movers scanner. Every lesson from 5+ days of live trading across 22 agents baked into the code. v1.1 adds the DSL state template directly in scanner output — eliminating the dsl-profile.json override bugs that broke Fox, Grizzly, Jackal, and every Wolf-based agent. XYZ equities banned at scan level. Leverage 7-10x enforced. Stagnation TP mandatory. 10% daily loss limit. 2-hour per-asset cooldown. Conviction-scaled Phase 1 timing per-signal. The agent cannot override any of these — they are in the scanner, not instructions.
The project's all-seeing guide. Sentinel MUST activate before Claude takes any action that modifies, creates, or deletes anything in the project. It understands the codebase, architecture, brand, design system, business model, deployment pipeline, testing strategy, and every convention. Trigger on: action requests (build, fix, add, change, update, refactor, implement, create, remove, delete, migrate, deploy, integrate, improve, configure, install, bump, upgrade, debug, troubleshoot, move, rename); casual requests (can you, I need to, let's, go ahead and, help me, we need to); status reports (X is broken/failing, there's a bug); project questions (how does X work here, where would I add, walk me through); planning (scope this, break this down, write a spec). Do NOT trigger on general knowledge, blog posts, interview prep, or tech comparisons for other projects. Key test: does this need THIS project's context? If yes, trigger. Sentinel guides Claude, it does not execute. No task is too small.
Set up or log in to Karma. Use when user says "set up agent", "configure API key", "connect to Karma", "login to Karma", "log in", or before first use of any Karma skill.
Exposes Claude's reasoning chain as an auditable, decomposable artifact. Quick mode (default) gives assumption inventory + weakest-link in 2 stages. Full mode (--full) adds decision branching, confidence decomposition, and falsification conditions. Triggers on "왜 그렇게 생각해", "reasoning", "근거", "show your work", "어떻게 그 결론이", "trace", "판단 근거", "why do you think that".
Skill for using Paperclip — open-source orchestration platform for running autonomous AI-agent companies with org charts, budgets, governance, and heartbeats.