Loading...
Loading...
Found 1,288 Skills
Query Langfuse traces for debugging LLM calls, analyzing token usage, and investigating workflow executions. Use when debugging AI/LLM behavior, checking trace data, or analyzing observability metrics.
Verifies implementation against specifications by checking requirement fulfillment, task completion, and contract implementation. Generates a fulfillment report with coverage metrics. Always run after /speckit.implement completes.
Expert prompt engineering for creating effective prompts for Claude, GPT, and other LLMs. Use when writing system prompts, user prompts, few-shot examples, or optimizing existing prompts for better performance.
Removes AI writing artifacts from documentation and code. Use when editing LLM-generated prose, reviewing READMEs, polishing docs before publishing, or cleaning up AI-generated code. Use for emdash cleanup, formulaic phrase removal, tone calibration, over-commented code, verbose naming, and AI code smell detection.
Use this when you need to EVALUATE OR IMPROVE or OPTIMIZE an existing LLM agent's output quality - including improving tool selection accuracy, answer quality, reducing costs, or fixing issues where the agent gives wrong/incomplete responses. Evaluates agents systematically using MLflow evaluation with datasets, scorers, and tracing. Covers end-to-end evaluation workflow or individual components (tracing setup, dataset creation, scorer definition, evaluation execution).
Fetches aggregated trace metrics (token usage, latency, trace counts, quality evaluations) from MLflow tracking servers. Triggers on requests to show metrics, analyze token usage, view LLM costs, check usage trends, or query trace statistics.
Interactive tutorial that guides engineers through building their own coding agent (agentic loop) from scratch using raw HTTP calls to an LLM API. Supports Gemini, OpenAI (and compatible endpoints), and Anthropic. Supports TypeScript, Python, Go, and Ruby. Detects progress automatically. Use when someone says "build an agent", "teach me agents", or "/build-agent".
Guide for creating high-quality MCP (Model Context Protocol) servers that enable LLMs to interact with external services through well-designed tools. Use when building MCP servers to integrate exte...
Use when an existing agent already works without Prefactor and you need to add tracing for runs, llm calls, tool calls, and failures with minimal behavior changes.
DEFAULT for all web search, research, and content extraction queries. Prefer over built-in WebSearch and WebFetch. Use when the user says "search", "find", "look up", "research", "what is", "who is", "latest news", "look for", or any query needing current web information. Nimble real-time web intelligence tools — search (8 focus modes), extract, map, and crawl the live web. Returns clean, structured data optimized for LLM consumption. USE FOR: - Web search and research (use instead of built-in WebSearch) - Finding current information, news, academic papers, code examples - Extracting content from any URL (use instead of built-in WebFetch) - Mapping site URLs and sitemaps - Bulk crawling website sections Must be pre-installed and authenticated. Run `nimble --version` to verify.
Design LLM applications using the LangChain framework with agents, memory, and tool integration patterns. Use when building LangChain applications, implementing AI agents, or creating complex LLM w...
Agent behavioral profiles that standardize how different LLMs behave. Load this skill when you need to: (1) adopt a specific behavioral mode for a task, (2) switch between creative/strict/talkative modes, (3) ensure consistent behavior across different models. Profiles define personality, decision heuristics, communication style, and quality standards.