Loading...
Loading...
Found 1,295 Skills
Find prompt and model quality issues using real conversation data, with specific optimization recommendations. Can implement prompt fixes and model switches directly in your codebase.
Use when starting any conversation -- establish how to find and use skills, requiring that the Skill tool be invoked before any response (including clarifying questions)
Starshipit integration. Manage Orders, Products, Customers, Users, Integrations. Use when the user wants to interact with Starshipit data.
Claude Code skill that makes AI agents respond in caveman-speak, cutting ~65-75% of output tokens while preserving full technical accuracy
Use when starting a new project with Maestro or when no .maestro.md context file exists yet. Run once per project.
Ultra-compressed communication mode. Talk like a caveman to reduce token usage by about 75%. Full technical accuracy is maintained. Intensity levels: 3 tiers - Polite, Normal (default), Extreme. Activate by saying "Caveman Mode", "Shorten", "Be Concise", "Save Tokens", or using /genshijin.
Spawn specialized sub-agents with context handoff for complex multi-phase tasks. Enables expertise delegation within a session with automatic context merging and depth limiting to prevent infinite loops.
LangGraph-based agent framework for consistent tool calling with automatic tool loops. Use when you need reliable multi-step task execution with OpenAI-compatible providers (Z.AI/GLM-5, OpenRouter, Groq, DeepSeek, Ollama).
Evaluate the quality of CAW (Cobo Agentic Wallet) Agent in local Claude Code, and generate scoring data and analysis reports. Use when: Users want to run CAW evaluation, conduct evaluation, test Skill, assess Agent quality, generate evaluation reports, or say "run evaluation", "evaluate CAW", "eval", "score". For weak model / openclaw evaluation, please use caw-eval-openclaw (only installed on openclaw servers).
Cost-conscious Claude Code mode. Reduces output tokens 40-70% and overall costs 30-60% by enforcing concise responses, smart model routing, and efficient workflow patterns. Keeps full technical accuracy. Activate with /cost-mode or "enable cost mode". Auto-triggers on mentions of budget, cost, tokens, or spending.
Operate LM Studio's `lms` CLI and local/remote LM Studio servers for model discovery, server status checks, model loading, endpoint smoke tests, and downstream OpenAI-compatible wiring. Use when the user mentions LM Studio, `lms`, a local model server, `/v1/models`, a remote LM Studio host, or wants to connect another tool to LM Studio; even if they only ask to test a local OpenAI-compatible endpoint or choose the correct loaded-model identifier. Triggers on: lmstudio, lm studio, lms, local model server, LM Studio API, LM Studio endpoint, /v1/models, connect Strix to LM Studio, load model in LM Studio.
Systematic debugging for ADK agents — trace reading, log analysis, common failure diagnosis, and the debug loop.