usability-tester

Conduct usability tests and identify UX issues through systematic observation. Use when testing user flows, validating designs, identifying friction points, running heuristic evaluations, or ensuring users can complete core tasks. Use for test planning, think-aloud protocol, task scenarios, severity rating, accessibility evaluation.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "usability-tester" with this command: pnpm dlx skills add oakoss/agent-skills

Usability Tester

Overview

Validates that users can successfully complete core tasks through systematic observation and expert evaluation. Covers moderated and unmoderated testing, heuristic evaluation, accessibility checks, and issue severity scoring. Not a substitute for analytics or A/B testing -- those measure what happens, usability testing reveals why.

When to use: Testing user flows, validating designs, identifying friction points, running heuristic evaluations, ensuring users can complete core tasks, planning and executing usability test sessions.

When NOT to use: Analytics or A/B test setup, visual design critique without task-based evaluation, automated UI testing (use a testing framework), performance benchmarking.

Quick Reference

MethodBest ForParticipantsWhen to Use
Moderated testingDeep insights, complex flows5-8 per personaDesign and prototyping stage
Unmoderated testingScale, quantitative data20-50+Pre-launch and post-launch
Guerrilla testingQuick validation, early concepts5-10 randomEarly concept stage
First-click testingNavigation, information architecture20-50Any stage, especially IA redesigns
Heuristic evaluationExpert review against principles3-5 evaluatorsBefore user testing, design audits
Cognitive walkthroughTask flow analysis2-3 evaluatorsEarly design, new feature review
Accessibility auditInclusive design validation3-5 users with disabilitiesPre-launch, compliance reviews
Synthetic user testingScalable task validation with AI agentsN/A (automated)Continuous, regression testing
AI-moderated sessionsAsync moderated testing at scale10-50+When moderator availability is limited

Core Metrics

MetricTargetWhat It Measures
Task success rate80% or higher for core tasksCan users complete the task
Time on taskSimple under 30s, medium 1-2m, complex 3-5mEfficiency
Error rateFewer than 2 per taskLearnability and clarity
Post-task satisfaction4.0 or higher on 5-point scaleSubjective ease
SUS score68+ (industry average), 80+ (excellent)Overall usability

Issue Severity

Severity equals Impact (1-3) multiplied by Frequency (1-3). Critical (8-9): fix before release. High (6-7): fix before release. Medium (4-5): next release. Low (1-3): backlog.

Common Mistakes

MistakeCorrect Pattern
Writing task scenarios with product jargon or step-by-step instructionsUse natural language with realistic context and clear goals to test discoverability
Helping users or explaining the UI during test sessionsObserve silently and let users struggle; confusion reveals real issues
Testing only the happy path with ideal inputsInclude error states, edge cases, and recovery scenarios in test plans
Using fewer than 5 participants per personaRecruit 5-8 users per persona; after 5, diminishing returns per Nielsen research
Collecting findings without severity scoring or follow-up actionsScore every issue by impact times frequency and assign fix priority
Skipping heuristic evaluation before user testingRun expert review first to catch obvious issues cheaply before spending on recruitment
Ignoring accessibility in test planningInclude participants with disabilities and test with assistive technologies
Running only one round of testingTest in waves: 5 users, fix issues, test 5 more until success rate meets targets

Delegation

  • Test plan and scenario creation: Use Task agent to draft objectives, research questions, and realistic task scenarios for each persona
  • Heuristic evaluation: Use Task agent to walk through interfaces against Nielsen's 10 heuristics and document violations with severity ratings
  • Metrics analysis and severity scoring: Use Task agent to calculate SUS scores, task success rates, and severity ratings from raw test data
  • Test report generation: Use Task agent to compile findings into structured reports with prioritized recommendations
  • Test infrastructure planning: Use Plan agent to select testing methods, define recruitment criteria, and schedule test waves

If the quality-auditor skill is available, delegate full codebase quality audits and scoring to it. Otherwise, recommend: pnpm dlx skills add oakoss/agent-skills -s quality-auditor -a claude-code -y

References

  • Test Planning -- objectives, research questions, recruitment, task scenario templates, screening criteria
  • Conducting Tests -- think-aloud protocol, facilitation rules, post-task questions, session structure
  • Testing Methods -- unmoderated, guerrilla, first-click, cognitive walkthrough, method selection by lifecycle stage
  • Heuristic Evaluation -- Nielsen's 10 heuristics, evaluation process, severity rating, combining with user testing
  • Metrics and Severity -- success rate, time on task, errors, satisfaction, SUS scoring, severity formula
  • Reporting -- test report template, key insights format, stakeholder presentation, recommended actions
  • Accessibility Testing -- inclusive recruitment, assistive technology testing, WCAG alignment, accessibility heuristics
  • Remote and Tools -- remote vs in-person comparison, testing tools, test frequency, checklists

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

playwright

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

ui-ux-polish

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

tanstack-form

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

find-skills

No summary provided by upstream source.

Repository SourceNeeds Review