Loading...
Loading...
Found 774 Skills
Guides the agent through building LLM-powered applications with LangChain and stateful agent workflows with LangGraph. Triggered when the user asks to "create an AI agent", "build a LangChain chain", "create a LangGraph workflow", "implement tool calling", "build RAG pipeline", "create a multi-agent system", "define agent state", "add human-in-the-loop", "implement streaming", or mentions LangChain, LangGraph, chains, agents, tools, retrieval augmented generation, state graphs, or LLM orchestration.
AI/ML APIs, LLM integration, and intelligent application patterns
Produce an LLM Build Pack (prompt+tool contract, data/eval plan, architecture+safety, launch checklist). Use for building with LLMs, GPT/Claude apps, prompt engineering, RAG, and tool-using agents.
Use when building secure AI pipelines or hardening LLM integrations. Defense-in-depth implements 8 validation layers from edge to storage with no single point of failure.
Agentic workflow patterns for autonomous LLM reasoning. Use when building ReAct agents, implementing reasoning loops, or creating LLMs that plan and execute multi-step tasks.
Expert prompt optimization for LLMs and AI systems. Use PROACTIVELY when building AI features, improving agent performance, or crafting system prompts. Masters prompt patterns and techniques.
Scaffold a new AI feature powered by DSPy. Use when adding AI to your app, starting a new AI project, building an AI-powered feature, setting up a DSPy program from scratch, or bootstrapping an LLM-powered backend.
Core technical documentation writing principles for voice, tone, structure, and LLM-friendly patterns. Use when writing or reviewing any documentation.
Use this skill when working with Apple's Foundation Models framework for on-device AI and LLM capabilities in iOS/macOS apps
Model Context Protocol expert for building MCP servers, tools, resources, and client integrationsUse when "mcp server, model context protocol, claude code extension, building ai tools, tool definition, mcp transport, stdio transport, sse transport, resource provider, prompt template, mcp, model-context-protocol, claude-code, ai-tools, llm-integration, anthropic, server, protocol" mentioned.
Hugging Face Transformers best practices including model loading, tokenization, fine-tuning workflows, and inference optimization. Use when working with transformer models, fine-tuning LLMs, implementing NLP tasks, or optimizing transformer inference.
This skill should be used when the user asks to "debug DSPy programs", "trace LLM calls", "monitor production DSPy", "use MLflow with DSPy", mentions "inspect_history", "custom callbacks", "observability", "production monitoring", "cost tracking", or needs to debug, trace, and monitor DSPy applications in development and production.