Loading...
Loading...
Found 1,140 Skills
Evaluate a vendor — cost analysis, risk assessment, and recommendation. Use when reviewing a new vendor proposal, deciding whether to renew or replace a contract, comparing two vendors side-by-side, or building a TCO breakdown and negotiation points before procurement sign-off.
Run a single experiment iteration. Edit the target file, evaluate, keep or discard.
Set up a new autoresearch experiment interactively. Collects domain, target file, eval command, metric, direction, and evaluator.
A decision-support framework that evaluates systems, architectures, and strategies through the entropy (decay) vs negentropy (growth) lens, while surfacing tacit knowledge gaps. Use this skill whenever the user is making architecture decisions, evaluating system designs, reviewing technical approaches, choosing between options, auditing existing systems, or planning strategies. Also trigger when the user explicitly asks to "apply the negentropy lens", mentions "entropy", "negentropy", "tacit knowledge", "knowledge engine", or "flip the switch". Nudge activation when you detect the user is at a decision point — even if they haven't asked for this lens — by briefly noting the entropic/negentropic dimension before proceeding.
Design click/first-click tests to evaluate navigation and information findability.
Evaluates interfaces, components, screens, and flows against universal UX/UI principles (heuristics, UX laws, Gestalt, cognitive psychology, accessibility) and delivers concrete, prioritized improvements. Use whenever the user shares UI code, screenshots, components, or mockups and wants feedback — even if they don't use the words "critique" or "review". Also trigger when the user asks "what's wrong with this UI", "how can I improve this", "review my component", "does this look right", "give me feedback on this design", or shares any interface and asks for thoughts. Trigger for partial slices too (a single button, form, or card) — not only full screens.
Evaluate and rank agent results by metric or LLM judge for an AgentHub session.
Expert skill for generating GitHub Copilot skills from ING-internal documentation repositories. Use this skill when asked to create a skill from any ING documentation-as-code repo, generate a knowledge base skill for an ING framework, convert ING tool documentation into a Copilot skill, or turn any docs/ folder into an expert skill file. Also trigger when the user mentions "skill from docs", "generate skill", "create skill from repo", or references ING-internal frameworks like Baker, Merak, Kingsroad, or similar. Includes evaluation framework, grading agents, and benchmark tools for testing generated skills.
Help identify and evaluate communities to build a minimalist business around. Use when someone is looking for a business idea, trying to find their community, or wondering where to start as an entrepreneur.
Instrument, trace, evaluate, and monitor LLM applications and AI agents with LangSmith. Use when setting up observability for LLM pipelines, running offline or online evaluations, managing prompts in the Prompt Hub, creating datasets for regression testing, or deploying agent servers. Triggers on: langsmith, langchain tracing, llm tracing, llm observability, llm evaluation, trace llm calls, @traceable, wrap_openai, langsmith evaluate, langsmith dataset, langsmith feedback, langsmith prompt hub, langsmith project, llm monitoring, llm debugging, llm quality, openevals, langsmith cli, langsmith experiment, annotate llm, llm judge.
Design and evaluate compression strategies for long-running sessions
Score content against GEO optimization criteria. Triggers on "score this", "rate content", "GEO score", "how does this rank", "evaluate content", "content score".