Loading...
Loading...
Found 1,140 Skills
Professionally evaluate story outlines, judge and score from the dimensions of market potential, innovation attributes, and content highlights. Suitable for story outline quality assessment, IP adaptation potential judgment, and project approval decision-making
Defines evaluation criteria and scoring methodologies for deliverable assessment
Discover scientific equations from data using LLM-guided evolutionary search (LLM-SR). Multi-island algorithm with softmax-based cluster sampling, island reset, and LLM-proposed equation mutations. Use for symbolic regression and equation discovery.
Create AI evaluation plans with benchmarks, rubrics, and error analysis workflows.
Iterative refinement workflow for polishing code, documentation, or designs through systematic evaluation and improvement cycles. Use when refining drafts into production-grade quality.
Implement comprehensive evaluation strategies for LLM applications using automated metrics, human feedback, and benchmarking. Use when testing LLM performance, measuring AI application quality, or establishing evaluation frameworks.
Respond in Chinese to user requests; when the user's message is fully in English (ignoring punctuation, digits, emojis, and whitespace), append a brief Chinese evaluation plus a 1-10 score.
Evaluate design effectiveness from a UX perspective. Assesses visual hierarchy, information architecture, emotional resonance, and overall design quality with actionable feedback.
A-share Value Investment Analysis Tool that provides stock screening, in-depth individual stock analysis, industry comparison and valuation calculation functions. Based on value investment theory, it uses akshare to obtain public financial data, suitable for ordinary investors with low-frequency trading.
Patterns and techniques for evaluating and improving AI agent outputs. Use this skill when: - Implementing self-critique and reflection loops - Building evaluator-optimizer pipelines for quality-critical generation - Creating test-driven code refinement workflows - Designing rubric-based or LLM-as-judge evaluation systems - Adding iterative improvement to agent outputs (code, reports, analysis) - Measuring and improving agent response quality
PluginEval quality methodology — dimensions, rubrics, statistical methods, and scoring formulas. Use this skill when understanding how plugin quality is measured, when interpreting a low score on a specific dimension, when deciding how to improve a skill's triggering accuracy or orchestration fitness, when calibrating scoring thresholds for your marketplace, or when explaining quality badges to external partners like Neon.
Analyze roster fit and personality dynamics. Leadership assessment, role clarity, locker room culture, trade/signing impact.