Loading...
Loading...
Found 80 Skills
Create, improve, or optimize prompts using best practices
A method for iteratively improving text instructions for agents (skills / slash commands / task prompts / CLAUDE.md sections / code generation prompts) by having unbiased executors run them, then evaluating from both perspectives (executor self-report + instruction-side metrics). Repeat until improvement plateaus. Use immediately after creating or significantly revising a prompt or skill, or when you suspect the reason an agent isn't behaving as expected is due to ambiguity in the instructions.
This skill should be used when working with DSPy.rb, a Ruby framework for building type-safe, composable LLM applications. Use this when implementing predictable AI features, creating LLM signatures and modules, configuring language model providers (OpenAI, Anthropic, Gemini, Ollama), building agent systems with tools, optimizing prompts, or testing LLM-powered functionality in Ruby applications.
This skill should be used when the user asks to "create a replit prompt", "write a prompt for replit", "optimize for replit agent", "prepare instructions for replit", or mentions building something with Replit Agent. Transforms user requirements into optimized, structured prompts that Replit Agent understands and executes accurately with minimal iterations.
Test, validate, and improve agent instructions (CLAUDE.md, system prompts) using sub-agents as experiment subjects. Measures instruction compliance, context decay, and constraint strength. Use for "test prompt", "validate instructions", "prompt effectiveness", "instruction decay", or when designing robust agent behaviors.
Generate optimized prompts for AI image and video generation. Triggers on "generate a prompt for", "write me a prompt", "create an image prompt", "create a video prompt", "optimize this prompt".
Find prompt and model quality issues using real conversation data, with specific optimization recommendations. Can implement prompt fixes and model switches directly in your codebase.
Create, optimize, and iteratively refine agent prompts and system prompts. Use when asked to "improve a prompt", "optimize a system prompt", "rewrite an agent prompt", "tune prompt wording", "make this prompt more reliable", or "adapt a prompt for OpenAI, Claude, or Gemini". Handles model-specific prompt guidance, prompt markers/tags, eval design, and meta optimization loops for new and existing prompts.
Ultra-compressed response mode. Cuts token usage by dropping articles, filler, pleasantries, and hedging. Uses symbols for relationships. Technical terms and code blocks remain exact and uncompressed. Use when user says "save tokens", "RTU mode", "compress", or "be brief".
Transform user requests into detailed, precise prompts for AI models. Use when users say "promptify", "promptify this", or explicitly request prompt engineering or improvement of their request for better AI responses.
Conversational guidance for building software with AI agents, covering workflows, tool selection, prompt strategies, parallel agent management, and best practices based on real-world high-volume agentic development experience. Use this skill when users ask about setting up agentic workflows, choosing models, optimizing prompts, managing parallel agents, or improving agent output quality.
This skill should be used when the user asks to "integrate DSPy with Haystack", "optimize Haystack prompts using DSPy", "use DSPy to improve Haystack pipeline", mentions "Haystack pipeline optimization", "combining DSPy and Haystack", "extract DSPy prompt for Haystack", or wants to use DSPy's optimization capabilities to automatically improve prompts in existing Haystack pipelines.