test-analyze

Test Analysis and Fix Planning

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "test-analyze" with this command: npx skills add laurigates/claude-plugins/laurigates-claude-plugins-test-analyze

Test Analysis and Fix Planning

Analyzes test results from any testing framework, uses Zen planner to create a systematic fix strategy, and delegates fixes to appropriate subagents.

Usage

/test:analyze <results-path> [--type <test-type>] [--focus <area>]

Parameters

<results-path> : Path to test results directory or file (required)

  • Examples: ./test-results/ , ./coverage/ , pytest-report.xml

--type <test-type> : Type of tests (optional, auto-detected if omitted)

  • accessibility

  • Playwright a11y, axe-core

  • unit

  • Jest, pytest, cargo test

  • integration

  • API tests, database tests

  • e2e

  • Playwright, Cypress, Selenium

  • security

  • OWASP ZAP, Snyk, TruffleHog

  • performance

  • Lighthouse, k6, JMeter

--focus <area> : Specific area to focus on (optional)

  • Examples: authentication , api , ui-components , database

Examples

Analyze Playwright accessibility test results

/test:analyze ./test-results/ --type accessibility

Analyze unit test failures with focus on auth

/test:analyze ./coverage/junit.xml --type unit --focus authentication

Auto-detect test type and analyze all issues

/test:analyze ./test-output/

Analyze security scan results

/test:analyze ./security-report.json --type security

Command Flow

Analyze Test Results

  • Parse test result files (XML, JSON, HTML, text)

  • Extract failures, errors, warnings

  • Categorize issues by type and severity

  • Identify patterns and root causes

Plan Fixes with PAL Planner

  • Use mcp__pal__planner for systematic planning

  • Break down complex fixes into actionable steps

  • Identify dependencies between fixes

  • Estimate effort and priority

Delegate to Subagents

  • Accessibility issues → code-review agent (WCAG compliance)

  • Security vulnerabilities → security-audit agent

  • Performance problems → system-debugging agent

  • Code quality issues → code-refactoring agent

  • Test infrastructure → test-architecture agent

  • Integration failures → system-debugging agent

  • Documentation gaps → documentation agent

Execute Plan

  • Sequential execution based on dependencies

  • Verification after each fix

  • Re-run tests to confirm resolution

Subagent Selection Logic

The command uses this decision tree to delegate:

Accessibility violations (WCAG, ARIA, contrast) → code-review agent with accessibility focus

Security issues (XSS, SQLi, auth bypass) → security-audit agent with OWASP analysis

Performance bottlenecks (slow queries, memory leaks) → system-debugging agent with profiling

Code smells (duplicates, complexity, coupling) → code-refactoring agent with SOLID principles

Flaky tests (race conditions, timing issues) → test-architecture agent with stability analysis

Build/CI failures (pipeline errors, dependency issues) → cicd-pipelines agent with workflow optimization

Output

The command produces:

Summary Report

  • Total issues found

  • Breakdown by category/severity

  • Top priorities

Fix Plan (from PAL planner)

  • Step-by-step remediation strategy

  • Dependency graph

  • Effort estimates

Subagent Assignments

  • Which agent handles which issues

  • Rationale for delegation

  • Execution order

Actionable Next Steps

  • Commands to run

  • Files to modify

  • Verification steps

Notes

  • Works with any test framework that produces structured output

  • Auto-detects common test result formats (JUnit XML, JSON, TAP)

  • Preserves test evidence for debugging

  • Can be chained with /git:smartcommit for automated fixes

  • Respects TDD workflow (RED → GREEN → REFACTOR)

Related Commands

  • /test:run

  • Run tests with framework detection

  • /code:review

  • Manual code review for test files

  • /docs:update

  • Update test documentation

  • /git:smartcommit

  • Commit fixes with conventional messages

Prompt:

Analyze test results from {{ARG1}} and create a systematic fix plan.

{{#if ARG2}} Test type: {{ARG2}} {{else}} Auto-detect test type from file formats and content. {{/if}}

{{#if ARG3}} Focus area: {{ARG3}} {{/if}}

Step 1: Analyze Test Results

Read the test result files from {{ARG1}} and extract:

  • Failed tests with error messages

  • Warnings and deprecations

  • Performance metrics (if available)

  • Coverage gaps (if available)

  • Categorize by: severity (critical/high/medium/low), type (functional/security/performance/accessibility)

Step 2: Use PAL Planner

Call mcp__pal__planner with model "gemini-2.5-pro" to create a systematic fix plan:

  • Step 1: Summarize findings and identify root causes

  • Step 2: Prioritize issues (impact × effort matrix)

  • Step 3: Break down fixes into actionable tasks

  • Step 4: Identify dependencies between fixes

  • Step 5: Assign each fix category to appropriate subagent

  • Continue planning steps as needed for complex scenarios

Step 3: Subagent Delegation Strategy

Based on the issue categories, delegate to:

Accessibility violations (WCAG, ARIA, color contrast, keyboard nav) → Use Task tool with subagent_type: code-review

→ Focus: WCAG 2.1 compliance, semantic HTML, ARIA best practices

Security vulnerabilities (XSS, SQLi, CSRF, auth issues) → Use Task tool with subagent_type: security-audit

→ Focus: OWASP Top 10, input validation, authentication

Performance issues (slow tests, memory leaks, timeouts) → Use Task tool with subagent_type: system-debugging

→ Focus: Profiling, bottleneck identification, optimization

Code quality (duplicates, complexity, maintainability) → Use Task tool with subagent_type: code-refactoring

→ Focus: SOLID principles, DRY, code smells

Flaky/unreliable tests (race conditions, timing, dependencies) → Use Task tool with subagent_type: test-architecture

→ Focus: Test stability, isolation, determinism

CI/CD failures (build errors, pipeline issues) → Use Task tool with subagent_type: cicd-pipelines

→ Focus: GitHub Actions, dependency management, caching

Documentation gaps (missing docs, outdated examples) → Use Task tool with subagent_type: documentation

→ Focus: API docs, test documentation, migration guides

Step 4: Create Execution Plan

For each subagent assignment:

  • Context: What files/areas need attention

  • Objective: Specific fix goal

  • Success Criteria: How to verify the fix

  • Dependencies: What must be done first

  • Verification: Commands to re-run tests

Step 5: Present Summary

Provide:

  • 📊 Issue Breakdown: Count by category and severity

  • 🎯 Priorities: Top 3-5 issues to fix first

  • 🤖 Subagent Plan: Which agents will handle what

  • ✅ Next Steps: Concrete actions to take

  • 🔍 Verification: How to confirm fixes worked

{{#if ARG3}} Additional focus on {{ARG3}}: Prioritize issues related to this area and provide extra context for relevant subagents. {{/if}}

Documentation-First Reminder: Before implementing fixes, research relevant documentation using context7 to verify:

  • Test framework best practices

  • Accessibility standards (WCAG 2.1)

  • Security patterns (OWASP)

  • Performance optimization techniques

TDD Workflow: Follow RED → GREEN → REFACTOR:

  • Verify tests fail (RED) ✓ (already done)

  • Implement minimal fix (GREEN)

  • Refactor for quality

  • Re-run tests to confirm

Do you want me to proceed with the analysis and planning, or would you like to review the plan first?

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Research

binary analysis

No summary provided by upstream source.

Repository SourceNeeds Review
Research

docs-knowledge-graph

No summary provided by upstream source.

Repository SourceNeeds Review
Research

test quality analysis

No summary provided by upstream source.

Repository SourceNeeds Review
Research

documentation quality analysis

No summary provided by upstream source.

Repository SourceNeeds Review