test-analyze

安装量: 48
排名: #15521

安装

npx skills add https://github.com/laurigates/claude-plugins --skill test-analyze
Test Analysis and Fix Planning
Analyzes test results from any testing framework, uses Zen planner to create a systematic fix strategy, and delegates fixes to appropriate subagents.
Usage
/test:analyze
<
results-path
>
[
--type
<
test-type
>
]
[
--focus
<
area
>
]
Parameters
Path to test results directory or file (required)
Examples:
./test-results/
,
./coverage/
,
pytest-report.xml
--type
Type of tests (optional, auto-detected if omitted)
accessibility
- Playwright a11y, axe-core
unit
- Jest, pytest, cargo test
integration
- API tests, database tests
e2e
- Playwright, Cypress, Selenium
security
- OWASP ZAP, Snyk, TruffleHog
performance
- Lighthouse, k6, JMeter
--focus
Specific area to focus on (optional) Examples: authentication , api , ui-components , database Examples

Analyze Playwright accessibility test results

/test:analyze ./test-results/ --type accessibility

Analyze unit test failures with focus on auth

/test:analyze ./coverage/junit.xml --type unit --focus authentication

Auto-detect test type and analyze all issues

/test:analyze ./test-output/

Analyze security scan results

/test:analyze ./security-report.json
--type
security
Command Flow
Analyze Test Results
Parse test result files (XML, JSON, HTML, text)
Extract failures, errors, warnings
Categorize issues by type and severity
Identify patterns and root causes
Plan Fixes with PAL Planner
Use
mcp__pal__planner
for systematic planning
Break down complex fixes into actionable steps
Identify dependencies between fixes
Estimate effort and priority
Delegate to Subagents
Accessibility issues
code-review
agent (WCAG compliance)
Security vulnerabilities
security-audit
agent
Performance problems
system-debugging
agent
Code quality issues
code-refactoring
agent
Test infrastructure
test-architecture
agent
Integration failures
system-debugging
agent
Documentation gaps
documentation
agent
Execute Plan
Sequential execution based on dependencies
Verification after each fix
Re-run tests to confirm resolution
Subagent Selection Logic
The command uses this decision tree to delegate:
Accessibility violations
(WCAG, ARIA, contrast)
code-review
agent with accessibility focus
Security issues
(XSS, SQLi, auth bypass)
security-audit
agent with OWASP analysis
Performance bottlenecks
(slow queries, memory leaks)
system-debugging
agent with profiling
Code smells
(duplicates, complexity, coupling)
code-refactoring
agent with SOLID principles
Flaky tests
(race conditions, timing issues)
test-architecture
agent with stability analysis
Build/CI failures
(pipeline errors, dependency issues)
cicd-pipelines
agent with workflow optimization
Output
The command produces:
Summary Report
Total issues found
Breakdown by category/severity
Top priorities
Fix Plan
(from PAL planner)
Step-by-step remediation strategy
Dependency graph
Effort estimates
Subagent Assignments
Which agent handles which issues
Rationale for delegation
Execution order
Actionable Next Steps
Commands to run
Files to modify
Verification steps
Notes
Works with any test framework that produces structured output
Auto-detects common test result formats (JUnit XML, JSON, TAP)
Preserves test evidence for debugging
Can be chained with
/git:smartcommit
for automated fixes
Respects TDD workflow (RED → GREEN → REFACTOR)
Related Commands
/test:run
- Run tests with framework detection
/code:review
- Manual code review for test files
/docs:update
- Update test documentation
/git:smartcommit
- Commit fixes with conventional messages
Prompt:
Analyze test results from {{ARG1}} and create a systematic fix plan.
{{#if ARG2}}
Test type: {{ARG2}}
{{else}}
Auto-detect test type from file formats and content.
{{/if}}
{{#if ARG3}}
Focus area: {{ARG3}}
{{/if}}
Step 1: Analyze Test Results
Read the test result files from {{ARG1}} and extract:
Failed tests with error messages
Warnings and deprecations
Performance metrics (if available)
Coverage gaps (if available)
Categorize by: severity (critical/high/medium/low), type (functional/security/performance/accessibility)
Step 2: Use PAL Planner
Call
mcp__pal__planner
with model "gemini-2.5-pro" to create a systematic fix plan:
Step 1: Summarize findings and identify root causes
Step 2: Prioritize issues (impact × effort matrix)
Step 3: Break down fixes into actionable tasks
Step 4: Identify dependencies between fixes
Step 5: Assign each fix category to appropriate subagent
Continue planning steps as needed for complex scenarios
Step 3: Subagent Delegation Strategy
Based on the issue categories, delegate to:
Accessibility violations
(WCAG, ARIA, color contrast, keyboard nav)
→ Use
Task
tool with
subagent_type: code-review
→ Focus: WCAG 2.1 compliance, semantic HTML, ARIA best practices
Security vulnerabilities
(XSS, SQLi, CSRF, auth issues)
→ Use
Task
tool with
subagent_type: security-audit
→ Focus: OWASP Top 10, input validation, authentication
Performance issues
(slow tests, memory leaks, timeouts)
→ Use
Task
tool with
subagent_type: system-debugging
→ Focus: Profiling, bottleneck identification, optimization
Code quality
(duplicates, complexity, maintainability)
→ Use
Task
tool with
subagent_type: code-refactoring
→ Focus: SOLID principles, DRY, code smells
Flaky/unreliable tests
(race conditions, timing, dependencies)
→ Use
Task
tool with
subagent_type: test-architecture
→ Focus: Test stability, isolation, determinism
CI/CD failures
(build errors, pipeline issues)
→ Use
Task
tool with
subagent_type: cicd-pipelines
→ Focus: GitHub Actions, dependency management, caching
Documentation gaps
(missing docs, outdated examples)
→ Use
Task
tool with
subagent_type: documentation
→ Focus: API docs, test documentation, migration guides
Step 4: Create Execution Plan
For each subagent assignment:
Context
What files/areas need attention
Objective
Specific fix goal
Success Criteria
How to verify the fix
Dependencies
What must be done first
Verification
Commands to re-run tests
Step 5: Present Summary
Provide:
📊
Issue Breakdown
Count by category and severity
🎯
Priorities
Top 3-5 issues to fix first
🤖
Subagent Plan
Which agents will handle what
Next Steps
Concrete actions to take
🔍
Verification
How to confirm fixes worked
{{#if ARG3}}
Additional focus on {{ARG3}}
Prioritize issues related to this area and provide extra context for relevant subagents.
{{/if}}
Documentation-First Reminder
Before implementing fixes, research relevant documentation using context7 to verify:
Test framework best practices
Accessibility standards (WCAG 2.1)
Security patterns (OWASP)
Performance optimization techniques
TDD Workflow
Follow RED → GREEN → REFACTOR: Verify tests fail (RED) ✓ (already done) Implement minimal fix (GREEN) Refactor for quality Re-run tests to confirm Do you want me to proceed with the analysis and planning, or would you like to review the plan first?
返回排行榜