Loading...
Loading...
Found 3 Skills
Systematic improvement of existing agents through performance analysis, prompt engineering, and continuous iteration.
Use this when you need to EVALUATE OR IMPROVE or OPTIMIZE an existing LLM agent's output quality - including improving tool selection accuracy, answer quality, reducing costs, or fixing issues where the agent gives wrong/incomplete responses. Evaluates agents systematically using MLflow evaluation with datasets, scorers, and tracing. Covers end-to-end evaluation workflow or individual components (tracing setup, dataset creation, scorer definition, evaluation execution).
Agent skill for v3-performance-engineer - invoke with $agent-v3-performance-engineer