analytics-flow
Description: Project metrics, documentation analytics, progress tracking, and trend analysis
Category: Data Analysis & Reporting
Complexity: Medium (data aggregation + visualization)
Purpose
Provide visibility into project progress, documentation quality, requirement coverage, and development velocity. Generate actionable insights from SDD artifacts to identify bottlenecks and optimize workflows.
Capabilities 1. Documentation Coverage Metrics Total documents by type (BRD, PRD, REQ, BDD, etc.) Token count distribution Documentation completeness percentage Outdated document identification Documentation debt tracking 2. Requirement Tracking Total requirements by category Requirement status (draft, approved, implemented, tested) Requirement coverage (REQ → BDD → TASK mapping) Orphaned requirements Requirement age and staleness 3. Traceability Analysis Complete traceability chains (BRD → ... → Code) Broken traceability links Traceability coverage percentage Gap identification Circular dependency detection 4. Implementation Progress Tasks completed vs total Implementation phase progress Blocked vs active tasks Timeline adherence Milestone tracking 5. Velocity Tracking Requirements completed per sprint/week Implementation velocity Documentation velocity Test coverage velocity Trend analysis over time 6. Quality Metrics Code review scores over time Test coverage trends Security vulnerability trends Code complexity trends Documentation quality scores 7. Bottleneck Identification Stuck requirements (no progress) Blocked tasks Resource constraints Dependency chains blocking progress Review bottlenecks 8. Custom Report Generation Executive summaries Sprint/weekly reports Milestone reports Team performance reports Risk assessment reports Analytics Workflow graph TD A[SDD Artifacts] --> B[Data Collection] B --> C[Parse Documents] C --> D[Extract Metadata]
D --> E[Requirement Analysis]
D --> F[Traceability Analysis]
D --> G[Progress Analysis]
D --> H[Quality Analysis]
E --> I[Requirement Metrics]
F --> J[Traceability Matrix]
G --> K[Progress Dashboard]
H --> L[Quality Dashboard]
I --> M[Aggregate Data]
J --> M
K --> M
L --> M
M --> N[Calculate KPIs]
N --> O[Identify Trends]
O --> P[Detect Anomalies]
P --> Q[Generate Insights]
Q --> R{Report Type}
R -->|Executive| S[Executive Summary]
R -->|Technical| T[Technical Report]
R -->|Sprint| U[Sprint Report]
S --> V[Distribute Reports]
T --> V
U --> V
Usage Instructions Generate Project Dashboard analytics-flow dashboard --output reports/dashboard.html
Generated dashboard sections:
Project Overview Dashboard
Overall Health: 78/100 (Good)
Documentation Coverage: ├─ BRD Documents: 3/3 (100%) ✓ ├─ PRD Documents: 2/3 (67%) ⚠️ ├─ Requirements: 146/150 (97%) ✓ ├─ BDD Scenarios: 89/146 (61%) ⚠️ ├─ ADR Documents: 12/12 (100%) ✓ ├─ SPEC Documents: 5/8 (63%) ⚠️ └─ TASK Documents: 45/89 (51%) ❌
Implementation Progress: ├─ Phase 1: 100% (Completed) ✓ ├─ Phase 2: 75% (In Progress) → ├─ Phase 3: 0% (Not Started) ○ └─ Overall: 58% complete
Traceability Coverage: ├─ Forward Traceability: 94% ✓ ├─ Backward Traceability: 91% ✓ └─ Complete Chains: 87% ✓
Quality Metrics: ├─ Code Coverage: 87% ✓ ├─ Code Review Score: 82/100 ✓ ├─ Security Issues: 3 (1 Critical) ⚠️ └─ Documentation Quality: 89% ✓
Velocity (Last 4 Weeks): ├─ Requirements: 12/week (avg) ├─ Implementation: 8 tasks/week ├─ Tests Written: 45 tests/week └─ Trend: ↗ Improving
Bottlenecks Identified: ├─ REQ-AUTH-015: Stuck for 3 weeks ├─ BDD Scenarios: 39% gap vs requirements └─ Phase 2 Dependency: Blocked by external API
Generate Requirement Report analytics-flow requirements --format json --output reports/requirements.json
Output:
{ "summary": { "total_requirements": 146, "by_priority": { "MUST": 130, "SHOULD": 12, "MAY": 4 }, "by_status": { "draft": 5, "approved": 85, "implemented": 45, "tested": 38, "deployed": 30 }, "by_category": { "authentication": 15, "authorization": 12, "data_processing": 45, "api": 38, "ml_models": 32, "infrastructure": 4 } }, "coverage": { "with_bdd_scenarios": 89, "with_tasks": 78, "with_tests": 67, "complete_traceability": 127 }, "age_analysis": { "average_age_days": 45, "oldest_requirement": { "id": "REQ-AUTH-01", "age_days": 120, "status": "implemented" }, "stale_requirements": [ {"id": "REQ-ML-015", "age_days": 90, "status": "draft"} ] }, "orphaned_requirements": [ "REQ-AUTH-015", "REQ-DATA-023" ] }
Generate Traceability Matrix analytics-flow traceability --format mermaid --output reports/traceability.mmd
Generated Mermaid diagram:
graph LR BRD001[BRD-SYSTEM-001] --> PRD001[PRD-FEATURES-001] PRD001 --> REQ001[REQ-AUTH-01] PRD001 --> REQ002[REQ-AUTH-002]
REQ001 --> BDD001[BDD-LOGIN-001]
REQ001 --> BDD002[BDD-LOGIN-002]
REQ002 --> BDD003[BDD-LOGOUT-001]
BDD001 --> TASK001[TASK-AUTH-001]
BDD002 --> TASK002[TASK-AUTH-002]
BDD003 --> TASK003[TASK-AUTH-003]
TASK001 --> CODE001[auth/login.py]
TASK002 --> CODE001
TASK003 --> CODE002[auth/logout.py]
CODE001 --> TEST001[test_login.py]
CODE002 --> TEST002[test_logout.py]
style BRD001 fill:#e1f5ff
style PRD001 fill:#e8f5e9
style REQ001 fill:#fff3e0
style REQ002 fill:#fff3e0
style BDD001 fill:#f3e5f5
style BDD002 fill:#f3e5f5
style BDD003 fill:#f3e5f5
Generate Sprint Report analytics-flow sprint \ --start-date 2025-01-01 \ --end-date 2025-01-14 \ --output reports/sprint-report.md
Generated report:
Sprint Report: Jan 1-14, 2025
Sprint Summary
- Duration: 2 weeks
- Team Size: 5 developers
- Sprint Goal: Complete Phase 2 authentication features
Accomplishments
✓ Completed: 12 requirements ✓ Implemented: 18 tasks ✓ Tests Written: 67 tests ✓ Code Coverage: 87% → 91% (+4%)
Requirements Completed
- REQ-AUTH-01: User login with JWT ✓
- REQ-AUTH-002: User logout ✓
- REQ-AUTH-003: Password reset ✓
- REQ-AUTH-004: Email verification ✓ ... (8 more)
Velocity Metrics
- Requirements: 6/week (target: 5/week) ↗
- Tasks: 9/week (target: 8/week) ↗
- Tests: 33.5/week (target: 25/week) ↗
Quality Metrics
- Code Review Score: 85/100 (+3 from last sprint)
- Security Issues: 2 (down from 5)
- Test Coverage: 91% (up from 87%)
Blockers Resolved
- External API integration (blocked 5 days) → Resolved
- Database migration issue → Resolved
Current Blockers
- Waiting for design review on user profile page (2 days)
Risks
⚠️ Phase 3 dependency on external service (mitigation: mock implementation)
Next Sprint Goals
- Complete remaining 8 authentication requirements
- Begin authorization framework
- Improve test coverage to 93%
Track Velocity Trends analytics-flow velocity --weeks 12 --output reports/velocity.json
Output with trend analysis:
{ "period": "12 weeks", "metrics": [ { "week": "2025-W01", "requirements_completed": 8, "tasks_completed": 15, "tests_written": 42, "code_coverage": 85 }, { "week": "2025-W02", "requirements_completed": 12, "tasks_completed": 18, "tests_written": 67, "code_coverage": 91 } ], "trends": { "requirements": { "average": 10, "trend": "increasing", "velocity_change": "+15%" }, "tasks": { "average": 16.5, "trend": "stable", "velocity_change": "+2%" }, "tests": { "average": 54.5, "trend": "increasing", "velocity_change": "+25%" } }, "forecast": { "next_sprint_requirements": 13, "completion_date_phase_2": "2025-02-15", "confidence": "high" } }
Key Performance Indicators (KPIs) Documentation KPIs
Documentation Coverage
Coverage = (Completed Documents / Total Required Documents) × 100% Target: ≥95%
Documentation Debt
Debt = Outdated Documents + Missing Documents Target: <5% of total documents
Average Document Age
Age = Days since last update Target: <30 days for active documents
Requirement KPIs
Requirement Coverage
Coverage = (Requirements with Tests / Total Requirements) × 100% Target: 100% for MUST requirements
Requirement Cycle Time
Cycle Time = Days from draft to deployed Target: <14 days (2 sprints)
Orphaned Requirements
Orphaned = Requirements without forward/backward links Target: 0
Implementation KPIs
Implementation Velocity
Velocity = Tasks completed / Time period Track: Weekly, calculate 4-week moving average
Task Completion Rate
Completion = (Completed Tasks / Total Tasks) × 100% Target: ≥90% per sprint
Blocked Task Percentage
Blocked % = (Blocked Tasks / Active Tasks) × 100% Target: <10%
Traceability KPIs
Forward Traceability
Forward = (Requirements with downstream links / Total Requirements) × 100% Target: 100%
Backward Traceability
Backward = (Tasks with upstream links / Total Tasks) × 100% Target: 100%
Complete Traceability Chains
Complete = (Requirements with full BRD→Code chain / Total) × 100% Target: ≥95%
Quality KPIs
Code Coverage Trend
Track: Weekly, target: ≥90% Alert: If drops >5% week-over-week
Code Review Score
Track: Per PR, target: ≥80/100 Alert: If average <70 for 2 consecutive weeks
Security Vulnerability Density
Density = Critical vulnerabilities / KLOC Target: 0 critical vulnerabilities
Report Templates Executive Summary Template
Executive Project Summary
Date: {date} Period: {period}
Overall Status:
Progress
- Phase Completion: {phase_progress}%
- On Schedule: {schedule_status}
- Budget Status: {budget_status}
Key Achievements
- {achievement_1}
- {achievement_2}
- {achievement_3}
Risks & Mitigation
| Risk | Impact | Probability | Mitigation |
|------|--------|-------------|------------|
| {risk_1} | {impact} | {prob} | {mitigation} |
Next Milestone
- {milestone_name}: {milestone_date}
- {deliverables}
Resource Requirements
- {resource_needs}
Technical Report Template
Technical Project Report
Date: {date}
Architecture
- ADRs Approved: {adr_count}
- Major Decisions: {decisions_summary}
Implementation
- Requirements Completed: {req_completed}/{req_total}
- Code Coverage: {coverage}%
- Technical Debt: {tech_debt_score}
Quality Metrics
- Code Review Score: {review_score}/100
- Security Issues: {security_count}
- Performance Benchmarks: {performance_data}
Infrastructure
- Environments: {env_status}
- Deployments: {deployment_count}
- Uptime: {uptime}%
Traceability
- Forward Coverage: {forward_trace}%
- Backward Coverage: {backward_trace}%
- Gaps: {gap_count}
Data Sources Primary Sources All SDD artifact files (BRD, PRD, REQ, BDD, ADR, SPEC, TASK, CTR) Git commit history Test execution results Code coverage reports Security scan results CI/CD pipeline logs Derived Metrics Velocity calculations Trend analysis Predictive forecasting Anomaly detection Visualization Options Chart Types Burndown Chart: Sprint/project progress Velocity Chart: Weekly implementation velocity Coverage Trend: Test coverage over time Traceability Matrix: Requirement-to-code mapping Dependency Graph: Task dependencies Risk Heat Map: Risk probability vs impact Quality Dashboard: Multi-metric visualization Export Formats HTML dashboards (interactive) PDF reports (static) JSON data (for integration) Markdown reports (for Git) Mermaid diagrams (for documentation) CSV data (for spreadsheets) Tool Access
Required tools:
Read: Read all SDD artifacts Glob: Find all documents Grep: Search for patterns and IDs Bash: Execute analysis scripts
Required libraries:
pandas: Data analysis matplotlib/plotly: Visualization jinja2: Report templating pyyaml: YAML parsing Integration Points With doc-flow Extract metadata from generated documents Track documentation workflow progress Identify documentation gaps With test-automation Aggregate test coverage data Track test execution trends Correlate tests with requirements With code-review Track code quality trends Aggregate review scores Identify quality patterns With project-mngt Validate timeline adherence Track milestone progress Resource utilization analysis Best Practices Update frequency: Daily automated collection, weekly reports Baseline establishment: Set baseline metrics at project start Trend analysis: Focus on trends, not absolute numbers Actionable insights: Every metric should inform decisions Anomaly detection: Alert on unusual patterns Historical data: Maintain 6+ months of history Stakeholder alignment: Customize reports per audience Continuous improvement: Evolve metrics based on feedback Limitations Requires consistent SDD artifact structure Historical analysis limited to available data Predictive accuracy depends on data quality Cannot measure subjective quality factors Requires regular data collection for accuracy Success Criteria Real-time dashboard availability <5 minute report generation time 100% data accuracy All KPIs tracked automatically Trend analysis with 90% forecast accuracy Zero manual data collection required Stakeholder satisfaction ≥4/5 Notes Reports saved to reports/analytics/ Historical data stored in metrics/history.json Dashboard auto-refreshes every 15 minutes Alerts configured for KPI threshold breaches Custom metrics configurable via .analytics.yml