Loading...
Loading...
Found 277 Skills
Format prompts for different LLM providers with chat templates and HNSW-powered context retrieval
AI/ML APIs, LLM integration, and intelligent application patterns
PyTiDB (pytidb) setup and usage for TiDB from Python. Covers connecting, table modeling (TableModel), CRUD, raw SQL, transactions, vector/full-text/hybrid search, auto-embedding, custom embedding functions, and reference templates/snippets (vector/hybrid/image) plus agent-oriented examples (RAG/memory/text2sql).
Stop your AI from making things up. Use when your AI hallucinates, fabricates facts, isn't grounded in real data, doesn't cite sources, makes unsupported claims, or you need to verify AI responses against source material. Covers citation enforcement, faithfulness verification, grounding via retrieval, and confidence thresholds.
This skill should be used when the user asks to "integrate DSPy with Haystack", "optimize Haystack prompts using DSPy", "use DSPy to improve Haystack pipeline", mentions "Haystack pipeline optimization", "combining DSPy and Haystack", "extract DSPy prompt for Haystack", or wants to use DSPy's optimization capabilities to automatically improve prompts in existing Haystack pipelines.
This skill should be used when the user asks to "create issues from a scan", "prioritize what to fix", "rank the issues", "build a roadmap from scan results", "run Morphiq Rank", or mentions creating a prioritized roadmap from scan results. Consumes a Morphiq Scan Report, applies issue creation criteria with impact/effort weighting, and organizes issues into 4 progressive discovery tiers.
This skill should be used when the user asks to "diagnose context problems", "fix lost-in-middle issues", "debug agent failures", "understand context poisoning", or mentions context degradation, attention patterns, context clash, context confusion, or agent performance degradation. A core context engineering skill — also activates when the user mentions "context engineering" or "context-engineering" in the context of diagnosing and mitigating context failures.
Evidence-based Drug-Drug Interaction (DDI) assessment skill modeled after the Micromedex Drug-Reax methodology. Trigger this skill whenever the user types /drug-drug, mentions "drug interaction", "DDI", "drug-drug", "can I take X with Y", "interaction between", "交互作用", "併用", or asks whether two medications can be used together. This skill performs systematic literature retrieval via PubMed, CrossRef, and WebSearch, then produces a structured assessment report with Severity, Documentation, Onset, Mechanism, Clinical Effects, and Management — mirroring the Micromedex Drug-Reax classification framework. Even casual questions like "is it safe to combine A and B" should trigger this skill.
Explains the ADK Dev Console — what each tab shows, how to read Agent Steps, traces, and other UI features visible at localhost:3001 during adk dev
Design or audit AI-first help centers/knowledge bases/FAQs, including taxonomy, article templates, analytics, and AI support (RAG, chatbot, escalation), using 2025-2026 best practices
Query decomposition for multi-concept retrieval. Use when handling complex queries spanning multiple topics, implementing multi-hop retrieval, or improving coverage for compound questions.
LLM and ML model deployment for inference. Use when serving models in production, building AI APIs, or optimizing inference. Covers vLLM (LLM serving), TensorRT-LLM (GPU optimization), Ollama (local), BentoML (ML deployment), Triton (multi-model), LangChain (orchestration), LlamaIndex (RAG), and streaming patterns.