Loading...
Loading...
Found 11 Skills
See exactly what your AI did on a specific request. Use when you need to debug a wrong answer, trace a specific AI request, profile slow AI pipelines, find which step failed, inspect LM calls, view token usage per request, build audit trails, or understand why a customer got a bad response. Covers DSPy inspection, per-step tracing, OpenTelemetry instrumentation, and trace viewer setup.
Eino orchestration with Graph, Chain, and Workflow. Use when a user needs to build multi-step pipelines, compose components into executable graphs, handle streaming between nodes, use branching or parallel execution, manage state with checkpoints, or understand the Runnable abstraction. Covers Graph (directed graph with cycles), Chain (linear sequential), and Workflow (DAG with field mapping).
Generate production-ready fal.ai workflow JSON files. Use when user requests "create workflow", "chain models", "multi-step generation", "image to video pipeline", or complex AI generation pipelines.
Chain agents together in sequential or branching workflows with data passing
Build DAG-based AI pipelines connecting Gradio Spaces, HuggingFace models, and Python functions into visual workflows. Use when asked to create a workflow, build a pipeline, connect AI models, chain Gradio Spaces, create a daggr app, build multi-step AI applications, or orchestrate ML models. Triggers on: "build a workflow", "create a pipeline", "connect models", "daggr", "chain Spaces", "AI pipeline".
Vision, audio, and multimodal LLM integration patterns. Use when processing images, transcribing audio, generating speech, or building multimodal AI pipelines.
The orchestration layer for AI-native creative production. This skill coordinates multiple AI tools—video, image, audio, digital humans, effects—into cohesive campaigns, productions, and creative systems. As AI tools proliferate, the challenge shifts from "can we create this?" to "how do we orchestrate these capabilities into something coherent?" The AI Creative Director thinks in systems, not tools. In pipelines, not one-offs. In brand consistency across AI-generated assets. This is where creative vision meets technical orchestration. The AI Creative Director doesn't just use AI tools—they compose them into creative instruments that produce at scales and speeds previously impossible. Use when "AI creative director, orchestrate AI, AI campaign, multi-tool, AI workflow, AI pipeline, coordinate AI, AI production, AI creative system, full AI production, AI at scale, orchestration, creative-direction, ai-production, workflow, pipeline, multi-tool, scale, quality-control" mentioned.
Compress LLM responses to pure signal — Rocky's early notation style. Drop articles, filler, hedging. Best for pipelines and coding.
Build new AI method from scratch using the MTHDS standard (.mthds bundle files). Use when user says "create a pipeline", "build a workflow", "new .mthds file", "make a method", "design a pipe", or wants to create any new method from scratch. Guides the user through a 10-phase construction process.
Build a LiveAvatar integration end-to-end — assesses the user's existing stack, recommends the optimal path, and guides implementation. Use when: (1) Building a new LiveAvatar integration, (2) Adding a real-time avatar to an app or site, (3) Connecting LiveAvatar to an existing AI pipeline, (4) User mentions LiveAvatar, real-time avatar, interactive avatar, conversational avatar, or lip-sync avatar, (5) Deciding between Embed, FULL Mode, and LITE Mode, (6) Migrating from HeyGen Interactive Avatar to LiveAvatar.
Optimize a prompt through a critique-compress pipeline with semantic equivalence verification at each stage. Applies think-critically to improve the prompt, then compress-prompt to reduce it, validating that behavior is preserved after each transformation.