| title | analytics-flow: Analytics and data analysis workflow skill |
| name | analytics-flow |
| description | Analytics and data analysis workflow skill |
| tags | sdd-workflow, shared-architecture, domain-specific |
| custom_fields | [object Object] |
| version | 1.0.0 |
| document_control | [object Object] |
analytics-flow
Description: Project metrics, documentation analytics, progress tracking, and trend analysis
Category: Data Analysis & Reporting
Complexity: Medium (data aggregation + visualization)
Purpose
Provide visibility into project progress, documentation quality, requirement coverage, and development velocity. Generate actionable insights from SDD artifacts to identify bottlenecks and optimize workflows.
Capabilities
1. Documentation Coverage Metrics
- Total documents by type (BRD, PRD, REQ, BDD, etc.)
- Token count distribution
- Documentation completeness percentage
- Outdated document identification
- Documentation debt tracking
2. Requirement Tracking
- Total requirements by category
- Requirement status (draft, approved, implemented, tested)
- Requirement coverage (REQ → BDD → TASK mapping)
- Orphaned requirements
- Requirement age and staleness
3. Traceability Analysis
- Complete traceability chains (BRD → ... → Code)
- Broken traceability links
- Traceability coverage percentage
- Gap identification
- Circular dependency detection
4. Implementation Progress
- Tasks completed vs total
- Implementation phase progress
- Blocked vs active tasks
- Timeline adherence
- Milestone tracking
5. Velocity Tracking
- Requirements completed per sprint/week
- Implementation velocity
- Documentation velocity
- Test coverage velocity
- Trend analysis over time
6. Quality Metrics
- Code review scores over time
- Test coverage trends
- Security vulnerability trends
- Code complexity trends
- Documentation quality scores
7. Bottleneck Identification
- Stuck requirements (no progress)
- Blocked tasks
- Resource constraints
- Dependency chains blocking progress
- Review bottlenecks
8. Custom Report Generation
- Executive summaries
- Sprint/weekly reports
- Milestone reports
- Team performance reports
- Risk assessment reports
Analytics Workflow
graph TD
A[SDD Artifacts] --> B[Data Collection]
B --> C[Parse Documents]
C --> D[Extract Metadata]
D --> E[Requirement Analysis]
D --> F[Traceability Analysis]
D --> G[Progress Analysis]
D --> H[Quality Analysis]
E --> I[Requirement Metrics]
F --> J[Traceability Matrix]
G --> K[Progress Dashboard]
H --> L[Quality Dashboard]
I --> M[Aggregate Data]
J --> M
K --> M
L --> M
M --> N[Calculate KPIs]
N --> O[Identify Trends]
O --> P[Detect Anomalies]
P --> Q[Generate Insights]
Q --> R{Report Type}
R -->|Executive| S[Executive Summary]
R -->|Technical| T[Technical Report]
R -->|Sprint| U[Sprint Report]
S --> V[Distribute Reports]
T --> V
U --> V
Usage Instructions
Generate Project Dashboard
analytics-flow dashboard --output reports/dashboard.html
Generated dashboard sections:
Project Overview Dashboard
==========================
Overall Health: 78/100 (Good)
Documentation Coverage:
├─ BRD Documents: 3/3 (100%) ✓
├─ PRD Documents: 2/3 (67%) ⚠️
├─ Requirements: 146/150 (97%) ✓
├─ BDD Scenarios: 89/146 (61%) ⚠️
├─ ADR Documents: 12/12 (100%) ✓
├─ SPEC Documents: 5/8 (63%) ⚠️
└─ TASK Documents: 45/89 (51%) ❌
Implementation Progress:
├─ Phase 1: 100% (Completed) ✓
├─ Phase 2: 75% (In Progress) →
├─ Phase 3: 0% (Not Started) ○
└─ Overall: 58% complete
Traceability Coverage:
├─ Forward Traceability: 94% ✓
├─ Backward Traceability: 91% ✓
└─ Complete Chains: 87% ✓
Quality Metrics:
├─ Code Coverage: 87% ✓
├─ Code Review Score: 82/100 ✓
├─ Security Issues: 3 (1 Critical) ⚠️
└─ Documentation Quality: 89% ✓
Velocity (Last 4 Weeks):
├─ Requirements: 12/week (avg)
├─ Implementation: 8 tasks/week
├─ Tests Written: 45 tests/week
└─ Trend: ↗ Improving
Bottlenecks Identified:
├─ REQ-AUTH-015: Stuck for 3 weeks
├─ BDD Scenarios: 39% gap vs requirements
└─ Phase 2 Dependency: Blocked by external API
Generate Requirement Report
analytics-flow requirements --format json --output reports/requirements.json
Output:
{
"summary": {
"total_requirements": 146,
"by_priority": {
"MUST": 130,
"SHOULD": 12,
"MAY": 4
},
"by_status": {
"draft": 5,
"approved": 85,
"implemented": 45,
"tested": 38,
"deployed": 30
},
"by_category": {
"authentication": 15,
"authorization": 12,
"data_processing": 45,
"api": 38,
"ml_models": 32,
"infrastructure": 4
}
},
"coverage": {
"with_bdd_scenarios": 89,
"with_tasks": 78,
"with_tests": 67,
"complete_traceability": 127
},
"age_analysis": {
"average_age_days": 45,
"oldest_requirement": {
"id": "REQ-AUTH-01",
"age_days": 120,
"status": "implemented"
},
"stale_requirements": [
{"id": "REQ-ML-015", "age_days": 90, "status": "draft"}
]
},
"orphaned_requirements": [
"REQ-AUTH-015",
"REQ-DATA-023"
]
}
Generate Traceability Matrix
analytics-flow traceability --format mermaid --output reports/traceability.mmd
Generated Mermaid diagram:
graph LR
BRD001[BRD-SYSTEM-001] --> PRD001[PRD-FEATURES-001]
PRD001 --> REQ001[REQ-AUTH-01]
PRD001 --> REQ002[REQ-AUTH-002]
REQ001 --> BDD001[BDD-LOGIN-001]
REQ001 --> BDD002[BDD-LOGIN-002]
REQ002 --> BDD003[BDD-LOGOUT-001]
BDD001 --> TASK001[TASK-AUTH-001]
BDD002 --> TASK002[TASK-AUTH-002]
BDD003 --> TASK003[TASK-AUTH-003]
TASK001 --> CODE001[auth/login.py]
TASK002 --> CODE001
TASK003 --> CODE002[auth/logout.py]
CODE001 --> TEST001[test_login.py]
CODE002 --> TEST002[test_logout.py]
style BRD001 fill:#e1f5ff
style PRD001 fill:#e8f5e9
style REQ001 fill:#fff3e0
style REQ002 fill:#fff3e0
style BDD001 fill:#f3e5f5
style BDD002 fill:#f3e5f5
style BDD003 fill:#f3e5f5
Generate Sprint Report
analytics-flow sprint \
--start-date 2025-01-01 \
--end-date 2025-01-14 \
--output reports/sprint-report.md
Generated report:
# Sprint Report: Jan 1-14, 2025
## Sprint Summary
- Duration: 2 weeks
- Team Size: 5 developers
- Sprint Goal: Complete Phase 2 authentication features
## Accomplishments
✓ Completed: 12 requirements
✓ Implemented: 18 tasks
✓ Tests Written: 67 tests
✓ Code Coverage: 87% → 91% (+4%)
## Requirements Completed
1. REQ-AUTH-01: User login with JWT ✓
2. REQ-AUTH-002: User logout ✓
3. REQ-AUTH-003: Password reset ✓
4. REQ-AUTH-004: Email verification ✓
... (8 more)
## Velocity Metrics
- Requirements: 6/week (target: 5/week) ↗
- Tasks: 9/week (target: 8/week) ↗
- Tests: 33.5/week (target: 25/week) ↗
## Quality Metrics
- Code Review Score: 85/100 (+3 from last sprint)
- Security Issues: 2 (down from 5)
- Test Coverage: 91% (up from 87%)
## Blockers Resolved
1. External API integration (blocked 5 days) → Resolved
2. Database migration issue → Resolved
## Current Blockers
1. Waiting for design review on user profile page (2 days)
## Risks
⚠️ Phase 3 dependency on external service (mitigation: mock implementation)
## Next Sprint Goals
- Complete remaining 8 authentication requirements
- Begin authorization framework
- Improve test coverage to 93%
Track Velocity Trends
analytics-flow velocity --weeks 12 --output reports/velocity.json
Output with trend analysis:
{
"period": "12 weeks",
"metrics": [
{
"week": "2025-W01",
"requirements_completed": 8,
"tasks_completed": 15,
"tests_written": 42,
"code_coverage": 85
},
{
"week": "2025-W02",
"requirements_completed": 12,
"tasks_completed": 18,
"tests_written": 67,
"code_coverage": 91
}
],
"trends": {
"requirements": {
"average": 10,
"trend": "increasing",
"velocity_change": "+15%"
},
"tasks": {
"average": 16.5,
"trend": "stable",
"velocity_change": "+2%"
},
"tests": {
"average": 54.5,
"trend": "increasing",
"velocity_change": "+25%"
}
},
"forecast": {
"next_sprint_requirements": 13,
"completion_date_phase_2": "2025-02-15",
"confidence": "high"
}
}
Key Performance Indicators (KPIs)
Documentation KPIs
Documentation Coverage
Coverage = (Completed Documents / Total Required Documents) × 100% Target: ≥95%Documentation Debt
Debt = Outdated Documents + Missing Documents Target: <5% of total documentsAverage Document Age
Age = Days since last update Target: <30 days for active documents
Requirement KPIs
Requirement Coverage
Coverage = (Requirements with Tests / Total Requirements) × 100% Target: 100% for MUST requirementsRequirement Cycle Time
Cycle Time = Days from draft to deployed Target: <14 days (2 sprints)Orphaned Requirements
Orphaned = Requirements without forward/backward links Target: 0
Implementation KPIs
Implementation Velocity
Velocity = Tasks completed / Time period Track: Weekly, calculate 4-week moving averageTask Completion Rate
Completion = (Completed Tasks / Total Tasks) × 100% Target: ≥90% per sprintBlocked Task Percentage
Blocked % = (Blocked Tasks / Active Tasks) × 100% Target: <10%
Traceability KPIs
Forward Traceability
Forward = (Requirements with downstream links / Total Requirements) × 100% Target: 100%Backward Traceability
Backward = (Tasks with upstream links / Total Tasks) × 100% Target: 100%Complete Traceability Chains
Complete = (Requirements with full BRD→Code chain / Total) × 100% Target: ≥95%
Quality KPIs
Code Coverage Trend
Track: Weekly, target: ≥90% Alert: If drops >5% week-over-weekCode Review Score
Track: Per PR, target: ≥80/100 Alert: If average <70 for 2 consecutive weeksSecurity Vulnerability Density
Density = Critical vulnerabilities / KLOC Target: 0 critical vulnerabilities
Report Templates
Executive Summary Template
# Executive Project Summary
Date: {date}
Period: {period}
## Overall Status: {status_indicator}
### Progress
- Phase Completion: {phase_progress}%
- On Schedule: {schedule_status}
- Budget Status: {budget_status}
### Key Achievements
1. {achievement_1}
2. {achievement_2}
3. {achievement_3}
### Risks & Mitigation
| Risk | Impact | Probability | Mitigation |
|------|--------|-------------|------------|
| {risk_1} | {impact} | {prob} | {mitigation} |
### Next Milestone
- {milestone_name}: {milestone_date}
- {deliverables}
### Resource Requirements
- {resource_needs}
Technical Report Template
# Technical Project Report
Date: {date}
## Architecture
- ADRs Approved: {adr_count}
- Major Decisions: {decisions_summary}
## Implementation
- Requirements Completed: {req_completed}/{req_total}
- Code Coverage: {coverage}%
- Technical Debt: {tech_debt_score}
## Quality Metrics
- Code Review Score: {review_score}/100
- Security Issues: {security_count}
- Performance Benchmarks: {performance_data}
## Infrastructure
- Environments: {env_status}
- Deployments: {deployment_count}
- Uptime: {uptime}%
## Traceability
- Forward Coverage: {forward_trace}%
- Backward Coverage: {backward_trace}%
- Gaps: {gap_count}
Data Sources
Primary Sources
- All SDD artifact files (BRD, PRD, REQ, BDD, ADR, SPEC, TASK, CTR)
- Git commit history
- Test execution results
- Code coverage reports
- Security scan results
- CI/CD pipeline logs
Derived Metrics
- Velocity calculations
- Trend analysis
- Predictive forecasting
- Anomaly detection
Visualization Options
Chart Types
- Burndown Chart: Sprint/project progress
- Velocity Chart: Weekly implementation velocity
- Coverage Trend: Test coverage over time
- Traceability Matrix: Requirement-to-code mapping
- Dependency Graph: Task dependencies
- Risk Heat Map: Risk probability vs impact
- Quality Dashboard: Multi-metric visualization
Export Formats
- HTML dashboards (interactive)
- PDF reports (static)
- JSON data (for integration)
- Markdown reports (for Git)
- Mermaid diagrams (for documentation)
- CSV data (for spreadsheets)
Tool Access
Required tools:
Read: Read all SDD artifactsGlob: Find all documentsGrep: Search for patterns and IDsBash: Execute analysis scripts
Required libraries:
- pandas: Data analysis
- matplotlib/plotly: Visualization
- jinja2: Report templating
- pyyaml: YAML parsing
Integration Points
With doc-flow
- Extract metadata from generated documents
- Track documentation workflow progress
- Identify documentation gaps
With test-automation
- Aggregate test coverage data
- Track test execution trends
- Correlate tests with requirements
With code-review
- Track code quality trends
- Aggregate review scores
- Identify quality patterns
With project-mngt
- Validate timeline adherence
- Track milestone progress
- Resource utilization analysis
Best Practices
- Update frequency: Daily automated collection, weekly reports
- Baseline establishment: Set baseline metrics at project start
- Trend analysis: Focus on trends, not absolute numbers
- Actionable insights: Every metric should inform decisions
- Anomaly detection: Alert on unusual patterns
- Historical data: Maintain 6+ months of history
- Stakeholder alignment: Customize reports per audience
- Continuous improvement: Evolve metrics based on feedback
Limitations
- Requires consistent SDD artifact structure
- Historical analysis limited to available data
- Predictive accuracy depends on data quality
- Cannot measure subjective quality factors
- Requires regular data collection for accuracy
Success Criteria
- Real-time dashboard availability
- <5 minute report generation time
- 100% data accuracy
- All KPIs tracked automatically
- Trend analysis with 90% forecast accuracy
- Zero manual data collection required
- Stakeholder satisfaction ≥4/5
Notes
- Reports saved to
reports/analytics/ - Historical data stored in
metrics/history.json - Dashboard auto-refreshes every 15 minutes
- Alerts configured for KPI threshold breaches
- Custom metrics configurable via
.analytics.yml