--- description: > Calculate a productivity score using actual Agent Monitor metrics — session completion rates, cache efficiency (cache_read vs input), compaction pressure (baseline tokens), turn velocity (turn_count / total_turn_duration_ms), tool success ratio (PreToolUse vs PostToolUse), and the workflow intelligence API's complexity and effectiveness scores. --- # Productivity Score Calculate a productivity scorecard from the Agent Monitor's real data. ## Input The user provides: **$ARGUMENTS** Options: "today", "this week", "last 30 days", a session ID, or "compare" for period comparison. ## Data Sources | Endpoint | Returns | |----------|---------| | `GET /api/analytics` | Token totals (`total_input`, `total_output`, `total_cache_read`, `total_cache_write` — baselines pre-summed), tool_usage top 20, daily_events/sessions, event_types, sessions_by_status, agents_by_status, avg_events_per_session, total_subagents | | `GET /api/sessions?limit=100` | Sessions with metadata JSON: `thinking_blocks`, `turn_count`, `total_turn_duration_ms`, `usage_extras` (service_tier, speed, inference_geo) | | `GET /api/pricing/cost` | Total cost with per-model breakdown | | `GET /api/workflows/{sessionId}` | 11 workflow datasets: stats, orchestration, toolFlow, effectiveness, patterns, modelDelegation, errorPropagation, concurrency, complexity, compaction, cooccurrence | ## Score Components (each 0–100) ### 1. Completion Rate (20% weight) From `sessions_by_status`: - `completed / (completed + error + abandoned) × 100` - Bonus for high completed-to-active ratio - Penalty for abandoned sessions (wasted work) ### 2. Token Efficiency (20% weight) From analytics `tokens` (baselines are pre-summed into totals): - **Cache hit rate**: `total_cache_read / (total_cache_read + total_input) × 100` - Above 60% = excellent, below 30% = poor - **Output concentration**: `total_output / total_input` — 0.3–0.8 is balanced ### 3. Tool Effectiveness (20% weight) From `event_types`: - **Success ratio**: Count `PostToolUse` / Count `PreToolUse` — should be ~1.0; gap = tool failures - **API error rate**: Count `APIError` / total events — should be near 0 - From workflow `effectiveness` data: subagent completion rates, task success per type ### 4. Velocity (20% weight) From session metadata: - **Turns per session**: average `turn_count` across sessions - **Turn speed**: average `total_turn_duration_ms / turn_count` — lower = faster - **Events per session**: from `avg_events_per_session` in analytics overview - **Thinking depth**: average `thinking_blocks` — more thinking = more thorough (neutral metric) ### 5. Cost Efficiency (20% weight) From pricing: - **Cost per completed session**: `total_cost / completed_sessions` - **Cost trend**: comparing current period to previous (decreasing = improving) - **Model optimization**: sessions using expensive models (Opus) for tasks subagents handle with Haiku/Sonnet ## Overall Score Weighted sum → letter grade: - **A+** (95-100), **A** (90-94), **B+** (85-89), **B** (80-84), **C+** (75-79), **C** (70-74), **D** (60-69), **F** (<60) ## Output Format ``` ═══════════════════════════════════════ PRODUCTIVITY SCORE: 87/100 (B+) ═══════════════════════════════════════ Completion Rate ████████░░ 80/100 Token Efficiency █████████░ 92/100 Tool Effectiveness████████░░ 85/100 Velocity █████████░ 88/100 Cost Efficiency █████████░ 90/100 ═══════════════════════════════════════ ``` Then: top 3 strengths, top 3 improvement areas with actionable steps, and period comparison if available.