Loading...
Loading...
Found 445 Skills
Scan the Obsidian wiki and automatically discover missing cross-references between pages. Use this skill when the user says "link my pages", "find missing links", "cross-reference", "connect my wiki", "add wikilinks", "what pages should be linked", or after any large ingestion to ensure new pages are woven into the existing knowledge graph. Also trigger when the user mentions "orphan pages" in the context of wanting to connect them, or says things like "my wiki feels disconnected" or "pages aren't linked well". This is a write-heavy skill — it actually modifies pages to add links, unlike wiki-lint which just reports issues.
Answer questions by searching the compiled Obsidian wiki. Use this skill when the user asks a question about their knowledge base, wants to find information across their wiki, asks "what do I know about X", "find everything related to Y", or wants synthesized answers with citations from their wiki pages. Also use when the user wants to explore connections between topics in their wiki. Works from any project. Includes an index-only fast mode triggered by "quick answer", "just scan", "don't read the pages", "fast lookup" — returns answers from page summaries and frontmatter without reading page bodies.
Use this skill when users need to create, generate, modify, or validate Salesforce Lightning pages (FlexiPages). Trigger when users mention RecordPage, AppPage, HomePage, Lightning pages, page layouts, adding components to pages, or page customization. Also use when users say things like 'create a Lightning page', 'add a component to a page', 'customize the record page', 'generate a FlexiPage', or when they're working with FlexiPage XML files and need help with components, regions, or deployment errors. Always use this skill for any FlexiPage-related work, even if they just mention 'page' in the context of Salesforce.
Ultra-lightweight channel for feature workflows: No need to write design docs, checklists, or conduct phased reviews. Let AI write code directly as it normally would, but before it starts, tell it where the CodeStable knowledge base in the project is and how to search it. This way, the code it writes will have fewer pitfalls and be more consistent with project conventions. Trigger scenarios: Users say "fast mode", "fastforward", "skip all those steps", "just start coding", "help me make xxx" and the requirement is too small to go through the design process.
Phase 3 of the feature workflow – Complete the acceptance closed-loop. Four tasks: 1. Check layer by layer against {slug}-design.md to verify if the implementation deviates from the plan; fix any deviations on the spot instead of just "noting them" in the report. 2. Incorporate this feature into the project's overall architecture documentation. 3. If this feature changes the user story or boundaries of the corresponding requirement, update the requirement doc accordingly. 4. If this feature originated from a roadmap item, change the status of the corresponding entry in roadmap items.yaml to done and sync it with the main document. Finally, produce a {slug}-acceptance.md as the closed-loop proof for the entire workflow. Prerequisite: cs-feat-impl is completed. Trigger scenarios: User says "The feature is done, let's accept it", "Do the final check", "Prepare for merge", "Generate the acceptance report".
Generates blog post thumbnail images for Orbitant following the brand's visual identity, using Google's Imagen API (Nano Banana 2). Activates when creating blog images, generating thumbnails, designing featured images for articles, or when someone needs a visual for an Orbitant insight/blog post. Use this skill even if the user just says "I need an image for this article", "create a thumbnail", "generate a hero image", or "make a featured image". Also triggers when the user mentions "Nano Banana 2", "image generation", or asks for a prompt for an AI image tool.
Ingest GitHub Copilot CLI session history into an Obsidian wiki as distilled knowledge pages. Use this skill when the user wants to capture their Copilot CLI sessions into a personal wiki — extracting architecture decisions, debug notes, and patterns into searchable Obsidian pages. Triggers on phrases like "ingest my copilot sessions into obsidian", "add my copilot history to my wiki", "pull my copilot session history into the vault", "capture what I've learned from copilot into obsidian", "just the new sessions since last time", or "mine patterns across my copilot sessions". Also triggers when the user mentions session-store.db, ~/.copilot/session-state, or VS Code copilot-chat transcripts in the context of building a wiki or knowledge base. Does NOT trigger for general copilot usage questions, searching sessions, or backing up history.
The meta-skill that powers all other AI tools. Prompt engineering for creative applications is the art and science of communicating with AI models to produce exactly what you envision—in images, video, audio, and text. This isn't just "write better prompts." It's understanding how different models interpret language, how to structure requests for different modalities, how to iterate systematically, and how to build prompt libraries that encode your creative vision. The best prompt engineers have developed intuition for what words trigger what responses in each model. This skill is foundational—it amplifies the effectiveness of every other AI creative skill. Master this, and you master the interface to all AI creation. Use when "prompt, prompting, prompt engineering, better prompts, prompt optimization, how to prompt, prompt strategy, prompt library, prompt template, make AI understand, prompt-engineering, prompting, meta-skill, ai-creative, foundational, optimization, iteration" mentioned.
Automatically intercepts and optimizes prompts using the prompt-learning MCP server. Learns from performance over time via embedding-indexed history. Uses APE, OPRO, DSPy patterns. Activate on "optimize prompt", "improve this prompt", "prompt engineering", or ANY complex task request. Requires prompt-learning MCP server. NOT for simple questions (just answer them), NOT for direct commands (just execute them), NOT for conversational responses (no optimization needed).
World-class color theory expertise combining the scientific precision of Josef Albers' "Interaction of Color," the systematic thinking of color systems from Pantone and RAL, and the perceptual psychology insights from researchers like Bevil Conway. Color is not just aesthetics - it's communication, emotion, and usability compressed into wavelengths. Great color work is invisible when done right. Users don't notice "nice colors" - they notice when they can't read text, when buttons don't look clickable, when errors don't feel urgent, or when the interface feels "off" without knowing why. Color theory is the science of making the right thing feel obvious. Use when "color theory, color palette, color scheme, color harmony, complementary colors, analogous colors, contrast ratio, dark mode colors, light mode, color tokens, semantic colors, color accessibility, color blindness, color psychology, color system, brand colors, data visualization colors, color, design, accessibility, contrast, dark-mode, theming, tokens, wcag, palette, harmony" mentioned.
Unvarnished technical criticism combining Linus Torvalds' precision, Gordon Ramsay's standards, and James Bach's BS-detection. Use when code/tests need harsh reality checks, certification schemes smell fishy, or technical decisions lack rigor. No sugar-coating, just surgical truth about what's broken and why.
A Just-In-Time (JIT) compiler for Python that translates a subset of Python and NumPy code into fast machine code. Developed by Anaconda, Inc. Highly effective for accelerating loops, custom mathematical functions, and complex numerical algorithms. Use for @njit, @vectorize, prange, cuda.jit, numba.typed, JIT compilation, parallel loops, GPU acceleration with CUDA, Monte Carlo simulations, numerical algorithms, and high-performance Python computing.