Loading...
Loading...
Use when conducting user research (interviews, usability tests, surveys, A/B tests) or designing research studies. Covers discovery, validation, evaluative methods, research ops, governance, and measurement for software experiences.
npx skill4agent add vasilyu1983/ai-agents-public software-ux-researchsoftware-ui-ux-design| Type | Goal | Primary Outputs |
|---|---|---|
| Discovery | Understand needs and context | JTBD, opportunity areas, constraints |
| Validation | Reduce solution risk | Go/no-go, prioritization signals |
| Evaluative | Improve usability/accessibility | Severity-rated issues + fixes |
What do you need?
├─ WHY / needs / context → interviews, contextual inquiry, diary
├─ HOW / usability → moderated usability test, cognitive walkthrough, heuristic eval
├─ WHAT / scale → analytics/logs + targeted qual follow-ups
└─ WHICH / causal → experiments (if feasible) or preference tests| Question | Best methods | Avoid when | Output |
|---|---|---|---|
| What problems matter most? | Interviews, contextual inquiry, diary | Only surveys/analytics | Problem framing + evidence |
| Can users complete key tasks? | Moderated usability tests, task analysis | Stakeholder review | Task success + issue list |
| Is navigation findable? | Tree test, first-click, card sort | Extremely small audience [Inference] | IA changes + labels |
| What is happening at scale? | Funnels, cohorts, logs, support taxonomy | Instrumentation missing | Baselines + segments + drop-offs |
| Which variant performs better? | A/B, switchback, holdout | Insufficient power or high risk | Decision with confidence + guardrails |
| Stage | Decisions | Primary Methods | Secondary Methods | Output |
|---|---|---|---|---|
| Discovery | What to build and for whom | Interviews, field/diary, journey mapping | Competitive analysis, feedback mining | Opportunity brief + JTBD |
| Concept/MVP | Does the concept work? | Concept test, prototype usability | First-click/tree test | MVP scope + onboarding plan |
| Launch | Is it usable + accessible? | Usability testing, accessibility review | Heuristic eval, session replay | Launch blockers + fixes |
| Growth | What drives adoption/value? | Segmented analytics + qual follow-ups | Churn interviews, surveys | Retention drivers + friction |
| Maturity | What to optimize/deprecate? | Experiments, longitudinal tracking | Unmoderated tests | Incremental roadmap |
| Metric category | What it answers | Pair with |
|---|---|---|
| Adoption | Are people using it? | Outcome/value metric |
| Value | Does it help users succeed? | Adoption + qualitative reasons |
| Reliability | Does it fail in ways users notice? | Error rate + recovery success |
| Accessibility | Can diverse users complete flows? | Assistive-tech coverage + defect trends |
| Indicator | Example | Research Implication |
|---|---|---|
| Multi-step workflows | Draft → approve → publish | Task analysis + state mapping |
| Multi-role permissions | Admin vs editor vs viewer | Test each role + transitions |
| Data dependencies | Requires integrations/sync | Error-path + recovery testing |
| High stakes | Finance, healthcare | Safety checks + confirmations |
| Expert users | Dev tools, analytics | Recruit real experts (not proxies) |
| Approach | Guardrails | Risk Level |
|---|---|---|
| Templated usability tests | Script + task templates provided | Low |
| Customer interviews by PMs | Training + review required | Medium |
| Survey design by anyone | Central review + standard questions | Medium |
| Unsupervised research | Not recommended | High |
| Research Activity | Proxy Metric | Calculation |
|---|---|---|
| Usability testing finding | Prevented dev rework | Hours saved × $150/hr |
| Discovery interview | Prevented build-wrong-thing | Sprint cost × risk reduction % |
| A/B test conclusive result | Improved conversion | (ΔConversion × Traffic × LTV) - Test cost |
| Heuristic evaluation | Early defect detection | Defects found × Cost-to-fix-later |
| Confidence | Evidence requirement | Use for |
|---|---|---|
| High | Multiple methods or sources agree | High-impact decisions |
| Medium | Strong signal from one method + supporting indicators | Prioritization |
| Low | Single source / small sample | Exploratory hypotheses |
| Metric type | Example | Common pitfall |
|---|---|---|
| Adoption | Feature usage rate | “Used” ≠ “helpful” |
| Value/outcome | Task success, goal completion | Harder to instrument |
| Situation | Why it fails | Better method |
|---|---|---|
| Low power/traffic | Inconclusive results | Usability tests + trends |
| Many variables change | Attribution impossible | Prototype tests → staged rollout |
| Need “why” | Experiments don’t explain | Interviews + observation |
| Ethical constraints | Harmful denial | Phased rollout + holdouts |
| Long-term effects | Short tests miss delayed impact | Longitudinal + retention analysis |
Use only when researching automation/AI-powered features. Skip for traditional software UX.2026 benchmark: Trend reports consistently highlight AI-assisted analysis. Use AI for speed while keeping humans responsible for strategy and interpretation. Example reference: https://www.lyssna.com/blog/ux-research-trends/
| Dimension | Question | Methods |
|---|---|---|
| Mental model | What do users think the system can/can’t do? | Interviews, concept tests |
| Trust calibration | When do users over/under-rely? | Scenario tests, log review |
| Explanation usefulness | Does “why” help decisions? | A/B explanation variants, interviews |
| Failure recovery | Do users recover and finish tasks? | Failure-path usability tests |
| Failure type | Typical impact | What to measure |
|---|---|---|
| Wrong output | Rework, lost trust | Verification + override rate |
| Missing output | Manual fallback | Fallback completion rate |
| Unclear output | Confusion | Clarification requests |
| Non-recoverable failure | Blocked flow | Time-to-recovery, support contact |
| Use Case | Appropriate? | Why |
|---|---|---|
| Early concept brainstorming | WARNING: Supplement only | Generate edge cases, not validation |
| Scenario/edge case expansion | PASS Yes | Broaden coverage before real testing |
| Moderator training/practice | PASS Yes | Practice without participant burden |
| Hypothesis generation | PASS Yes | Explore directions to test with real users |
| Validation/go-no-go decisions | FAIL Never | Cannot substitute lived experience |
| Usability findings as evidence | FAIL Never | Real behavior required |
| Quotes in reports | FAIL Never | Fabricated quotes damage credibility |
| Domain | Industry Leaders to Check | Key Flows |
|---|---|---|
| Fintech/Banking | Wise, Revolut, Monzo, N26, Chime, Mercury | Onboarding/KYC, money transfer, card management, spend analytics |
| E-commerce | Shopify, Amazon, Stripe Checkout | Checkout, cart, product pages, returns |
| SaaS/B2B | Linear, Notion, Figma, Slack, Airtable | Onboarding, settings, collaboration, permissions |
| Developer Tools | Stripe, Vercel, GitHub, Supabase | Docs, API explorer, dashboard, CLI |
| Consumer Apps | Spotify, Airbnb, Uber, Instagram | Discovery, booking, feed, social |
| Healthcare | Oscar, One Medical, Calm, Headspace | Appointment booking, records, compliance flows |
| EdTech | Duolingo, Coursera, Khan Academy | Onboarding, progress, gamification |
"[domain] UX best practices 2026""[leader company] [flow type] UX""[leader company] app review UX" site:mobbin.com OR site:pageflows.com"[domain] onboarding flow examples"DOMAIN: Fintech (Money Transfer)
BENCHMARKED: Wise, Revolut
WISE PATTERNS:
- Upfront fee transparency (shows exact fee before recipient input)
- Mid-transfer rate lock (shows countdown timer)
- Delivery time estimate per payment method
- Recipient validation (bank account check before send)
REVOLUT PATTERNS:
- Instant send to Revolut users (P2P first)
- Currency conversion preview with rate comparison
- Scheduled/recurring transfers prominent
APPLY TO YOUR FLOW:
1. Add fee transparency at step 1 (not step 3)
2. Show delivery estimate per payment rail
3. Consider rate lock feature for FX transfers
DIFFERENTIATION OPPORTUNITY:
- Neither shows historical rate chart—add "is now a good time?" context"UX research trends 2026""UX research tools best practices 2026""[Maze/Hotjar/UserTesting] comparison 2026""AI in UX research 2026"