--- name: doc-tasks-autopilot description: Automated TASKS generation from SPEC/TSPEC - generates AI-structured TODO tasks for implementation with CODE-Ready scoring tags: - sdd-workflow - layer-11-artifact - automation-workflow - shared-architecture custom_fields: layer: 11 artifact_type: TASKS architecture_approaches: [ai-agent-based] priority: primary development_status: active skill_category: automation-workflow upstream_artifacts: [BRD, PRD, EARS, BDD, ADR, SYS, REQ, CTR, SPEC, TSPEC] downstream_artifacts: [Code] version: "2.4" last_updated: "2026-02-10T15:00:00" --- # doc-tasks-autopilot ## Purpose Automated **Task Breakdown (TASKS)** generation pipeline that processes SPEC and TSPEC documents to generate AI-structured TODO tasks for implementation with CODE-Ready scoring. **Layer**: 11 (Final documentation layer before code) **Upstream**: SPEC (Layer 9), TSPEC (Layer 10) **Downstream**: Code (Layer 12 - execution layer) --- ## Skill Dependencies | Skill | Purpose | Phase | |-------|---------|-------| | `doc-naming` | Element ID format (TASKS.NN.TT.SS, codes 18, 30) | All Phases | | `doc-spec-validator` | Validate SPEC TASKS-Ready score | Phase 2 | | `doc-tspec-validator` | Validate TSPEC TASKS-Ready score | Phase 2 | | `doc-tasks` | TASKS creation rules, TODO format | Phase 3 | | `quality-advisor` | Real-time quality feedback | Phase 3 | | `doc-tasks-validator` | Validation with CODE-Ready scoring | Phase 4 | | `doc-tasks-audit` | Unified validator + reviewer wrapper output | Phase 5: Audit | | `doc-tasks-reviewer` | Content review, link validation, quality scoring | Phase 5: Review | | `doc-tasks-fixer` | Apply fixes from audit/review report, create missing files | Phase 5: Fix | --- ## Smart Document Detection The autopilot automatically determines the action based on the input document type. ### Input Type Recognition (Multiple Upstreams) TASKS can be derived from SPEC and/or TSPEC: | Input | Detected As | Action | |-------|-------------|--------| | `TASKS-NN` | Self type | Review existing TASKS document | | `SPEC-NN` | Primary upstream | Generate if missing, review if exists | | `TSPEC-NN` | Alternative upstream | Generate if missing, review if exists | ### Detection Algorithm ``` 1. Parse input: Extract TYPE and NN from "{TYPE}-{NN}" 2. Determine action: - IF TYPE == "TASKS": Review Mode - ELSE IF TYPE in ["SPEC", "TSPEC"]: Generate/Find Mode - ELSE: Error (invalid type for this autopilot) 3. For Generate/Find Mode: - Check: Does TASKS-{NN} exist in docs/11_TASKS/? - IF exists: Switch to Review Mode for TASKS-{NN} - ELSE: Proceed with Generation from {TYPE}-{NN} ``` ### File Existence Check ```bash # Check for nested folder structure (mandatory) ls docs/11_TASKS/TASKS-{NN}_*/ ``` ### Examples ```bash # Review mode (same type - TASKS input) /doc-tasks-autopilot TASKS-01 # Reviews existing TASKS-01 # Generate/Find mode (upstream types) /doc-tasks-autopilot SPEC-01 # Generates TASKS-01 if missing, or reviews existing TASKS-01 /doc-tasks-autopilot TSPEC-01 # Generates TASKS-01 if missing, or reviews existing TASKS-01 # Multiple inputs /doc-tasks-autopilot SPEC-01,SPEC-02 # Generates/reviews TASKS-01 and TASKS-02 /doc-tasks-autopilot TASKS-01,TASKS-02 # Reviews TASKS-01 and TASKS-02 ``` ### Action Determination Output ``` Input: SPEC-01 ├── Detected Type: SPEC (primary upstream) ├── Expected TASKS: TASKS-01 ├── TASKS Exists: Yes → docs/11_TASKS/TASKS-01_f1_iam/ └── Action: REVIEW MODE - Running doc-tasks-reviewer on TASKS-01 Input: TSPEC-05 ├── Detected Type: TSPEC (alternative upstream) ├── Expected TASKS: TASKS-05 ├── TASKS Exists: No └── Action: GENERATE MODE - Creating TASKS-05 from TSPEC-05 Input: TASKS-03 ├── Detected Type: TASKS (self) └── Action: REVIEW MODE - Running doc-tasks-reviewer on TASKS-03 ``` --- ## Workflow Overview ```mermaid flowchart TD subgraph Phase1["Phase 1: SPEC/TSPEC Analysis"] A[Start] --> B[Read SPEC Documents] B --> C[Read TSPEC Documents] C --> D[Extract Implementation Units] D --> E[Extract Test Requirements] E --> F[Catalog Work Items] end subgraph Phase2["Phase 2: Task Decomposition Planning"] F --> G[Check SPEC TASKS-Ready Score] G --> H[Check TSPEC TASKS-Ready Score] H --> I{Both >= 90%?} I -->|No| J[Flag Issues] J --> K{Auto-Fixable?} K -->|Yes| L[Fix Issues] L --> G K -->|No| M[Abort - Manual Fix Required] I -->|Yes| N[Plan Task Decomposition] N --> O[Identify Dependencies] end subgraph Phase3["Phase 3: TASKS Generation"] O --> P[Generate Implementation Tasks] P --> Q[Generate Test Tasks] Q --> R[Build Dependency Graph] R --> S[Assign Priorities] S --> T[quality-advisor: Real-time Feedback] T --> U[Generate Implementation Contracts] U --> V[Add Traceability Tags] V --> W[Write TASKS File] end subgraph Phase4["Phase 4: TASKS Validation"] W --> X[Run doc-tasks-validator] X --> Y{CODE-Ready >= 90%?} Y -->|No| Z[Auto-Fix TASKS Issues] Z --> AA[Re-validate TASKS] AA --> Y Y -->|Yes| AB[Mark TASKS Validated] end subgraph Phase5["Phase 5: Review & Fix Cycle"] AB --> AC[Run doc-tasks-reviewer] AC --> AC2{Score >= 90?} AC2 -->|No| AC3[Run doc-tasks-fixer] AC3 --> AC4{Iteration < Max?} AC4 -->|Yes| AC AC4 -->|No| AC5[Flag Manual Review] AC2 -->|Yes| AD[Verify Quality Checks] AC5 --> AD AD --> AE[Update Traceability Matrix] AE --> AF[Generate Summary Report] end AF --> AG[Complete] M --> AI[Exit with Error] ``` --- ## Task Structure ### Element ID Format | Element Type | Code | Pattern | Example | |--------------|------|---------|---------| | Task | 18 | TASKS.NN.18.SS | TASKS.01.18.01 | | Task Item | 30 | TASKS.NN.30.SS | TASKS.01.30.01 | ### Priority Levels | Priority | Code | Description | |----------|------|-------------| | P0 | Critical | Must complete first, blocks all others | | P1 | High | Core functionality | | P2 | Medium | Important but not blocking | | P3 | Low | Nice to have, can defer | --- ## TASKS Document Structure ```markdown # TASKS-01: Component Implementation ## Document Control | Field | Value | |-------|-------| | Status | Draft | | CODE-Ready Score | 95% | ## 1. Task Overview Summary of implementation tasks derived from SPEC-01. ## 2. Dependency Graph ```mermaid graph TD T1[TASKS.01.18.01] --> T2[TASKS.01.18.02] T1 --> T3[TASKS.01.18.03] T2 --> T4[TASKS.01.18.04] T3 --> T4 ``` ## 3. Implementation Tasks ### TASKS.01.18.01: Setup Project Structure - **Priority**: P0 - **Upstream**: None - **Downstream**: TASKS.01.18.02, TASKS.01.18.03 - **SPEC Reference**: SPEC.01.28.01 - **Estimated Complexity**: 2/5 ### TASKS.01.18.02: Implement Core Logic - **Priority**: P1 - **Upstream**: TASKS.01.18.01 - **Downstream**: TASKS.01.18.04 - **SPEC Reference**: SPEC.01.28.02 ## 7. Implementation Contracts [Protocol interfaces, exception hierarchies, state machine contracts] ## 8. Traceability @brd: BRD.01.01.01 @prd: PRD.01.09.01 ... ``` --- ## Implementation Contracts Per `IMPLEMENTATION_CONTRACTS_GUIDE.md`, generate contracts when: - TASKS file has 3+ downstream dependencies - Shared interfaces across multiple sessions - Complex state machines or exception hierarchies **Contract Types**: 1. Protocol Interfaces 2. Exception Hierarchies 3. State Machine Contracts 4. Data Models 5. Dependency Injection Interfaces --- ## Phase 5: Review & Fix Cycle (v2.3) Iterative review and fix cycle to ensure TASKS quality before completion. ```mermaid flowchart TD A[Phase 5 Start] --> B[Run doc-tasks-reviewer] B --> C[Generate Review Report] C --> D{Review Score >= 90?} D -->|Yes| E[PASS - Proceed to Phase 6] D -->|No| F{Iteration < Max?} F -->|Yes| G[Run doc-tasks-fixer] G --> H[Apply Fixes] H --> I[Generate Fix Report] I --> J[Increment Iteration] J --> B F -->|No| K[Flag for Manual Review] K --> L[Generate Final Report with Remaining Issues] L --> E ``` ### 5.1 Initial Review Run `doc-tasks-reviewer` to identify issues. ```bash /doc-tasks-reviewer TASKS-NN ``` **Output**: `TASKS-NN.R_review_report_v001.md` ### 5.2 Fix Cycle If review score < 90%, invoke `doc-tasks-fixer`. ```bash /doc-tasks-fixer TASKS-NN --revalidate ``` **Fix Categories**: | Category | Fixes Applied | |----------|---------------| | Missing Sections | Add missing task overview, contracts sections | | Element IDs | Convert legacy patterns (TASK-XXX, TODO-XXX) | | Dependency Graph | Regenerate Mermaid diagram from task references | | Priority Format | Normalize to P0-P3 format | | Implementation Contracts | Generate when 3+ dependencies exist | | Traceability | Update cumulative tags (9 layers) | **Output**: `TASKS-NN.F_fix_report_v001.md` ### 5.3 Re-Review After fixes, automatically re-run reviewer. ```bash /doc-tasks-reviewer TASKS-NN ``` **Output**: `TASKS-NN.R_review_report_v002.md` ### 5.4 Iteration Control | Parameter | Default | Description | |-----------|---------|-------------| | `max_iterations` | 3 | Maximum fix-review cycles | | `target_score` | 90 | Minimum passing score | | `stop_on_manual` | false | Stop if only manual issues remain | **Iteration Example**: ``` Iteration 1: Review v001: Score 75 (3 errors, 5 warnings) Fix v001: Fixed 6 issues, regenerated dependency graph Iteration 2: Review v002: Score 92 (0 errors, 2 warnings) Status: PASS (score >= 90) ``` ### 5.5 Quality Checks (Post-Fix) After passing the fix cycle: 1. **Task Structure Completeness**: - All required task fields present (Priority, Upstream, Downstream) - SPEC references valid and linked - Complexity estimates provided 2. **Dependency Graph Validity**: - Valid DAG (no circular dependencies) - All task references resolved - Mermaid diagram renders correctly 3. **Element ID Compliance** (per `doc-naming` skill): - All IDs use TASKS.NN.TT.SS format - Element type codes valid for TASKS (18, 30) - No legacy patterns (TASK-XXX, TODO-XXX, TI-XXX) 4. **CODE-Ready Report**: ``` CODE-Ready Score Breakdown ========================== Task Completeness: 24/25 (all required fields) Dependency Graph: 20/20 (valid DAG) Priority Assignments: 15/15 (P0-P3 assigned) Implementation Contracts: 14/15 (contracts where needed) Traceability Tags: 15/15 (all 9 tags present) SPEC/TSPEC Alignment: 10/10 (references valid) ---------------------------- Total CODE-Ready Score: 98/100 (Target: >= 90) Status: READY FOR CODE IMPLEMENTATION ``` 5. **Traceability Matrix Update**: ```bash # Update TASKS traceability python ai_dev_flow/scripts/update_traceability_matrix.py \ --tasks docs/11_TASKS/TASKS-NN_{slug}.md \ --matrix docs/11_TASKS/TASKS-00_TRACEABILITY_MATRIX.md ``` --- ## Cumulative Tags (9 Required) ```markdown @brd: BRD.NN.TT.SS @prd: PRD.NN.TT.SS @ears: EARS.NN.TT.SS @bdd: BDD.NN.TT.SS @adr: ADR-NN @sys: SYS.NN.TT.SS @req: REQ.NN.TT.SS @spec: SPEC.NN.TT.SS @tspec: TSPEC.NN.TT.SS @ctr: CTR.NN.TT.SS # Optional ``` --- ## Configuration ### Default Configuration ```yaml tasks_autopilot: version: "1.0" scoring: tasks_ready_min: 90 code_ready_min: 90 strict_mode: false execution: max_parallel: 3 # HARD LIMIT - do not exceed chunk_size: 3 # Documents per chunk pause_between_chunks: true auto_fix: true continue_on_error: false timeout_per_spec: 180 # seconds output: report_format: markdown validation: skip_validation: false fix_iterations_max: 3 contracts: generate_when_dependencies_gte: 3 include_protocol_interfaces: true include_exception_hierarchies: true include_state_machines: true ``` --- ## Execution Modes ### Mode 1: Generate Mode (Default) Standard TASKS generation from SPEC/TSPEC documents (see Workflow Overview above). ### Mode 2: Review Mode (v2.1) Validate existing TASKS documents without modification. Generates quality report with actionable recommendations. **Command**: ```bash # Review single TASKS /doc-tasks-autopilot TASKS-01 --review # Review all TASKS in directory /doc-tasks-autopilot docs/11_TASKS/ --review --all # Review with detailed report /doc-tasks-autopilot TASKS-01 --review --verbose ``` **Review Process**: ```mermaid flowchart TD A[Input: TASKS Document] --> B[Load TASKS Content] B --> C[Validate Task Structure] C --> D[Check CODE-Ready Score] D --> E[Verify Dependency Graph] E --> F[Validate Element IDs] F --> G[Check Implementation Contracts] G --> H[Verify Cumulative Tags] H --> I[Generate Review Report] I --> J{Issues Found?} J -->|Yes| K[Categorize Issues] K --> L[Generate Fix Recommendations] L --> M[Output Review Report] J -->|No| M ``` **Review Report Template**: ```markdown # TASKS Review Report: TASKS-NN_{slug} ## Summary - **CODE-Ready Score**: NN% (✅/🟡/❌) - **Total Issues**: N (E errors, W warnings) - **Auto-Fixable**: N issues - **Manual Review**: N issues ## Score Breakdown | Category | Score | Max | Status | |----------|-------|-----|--------| | Task Completeness | NN | 25 | ✅/🟡/❌ | | Dependency Graph | NN | 20 | ✅/🟡/❌ | | Priority Assignments | NN | 15 | ✅/🟡/❌ | | Implementation Contracts | NN | 15 | ✅/🟡/❌ | | Traceability Tags | NN | 15 | ✅/🟡/❌ | | SPEC/TSPEC Alignment | NN | 10 | ✅/🟡/❌ | ## Issues by Category ### Auto-Fixable Issues | Issue | Location | Fix Action | |-------|----------|------------| | Legacy ID pattern | Line 45 | Convert TASK-001 → TASKS.01.18.01 | | Missing cumulative tag | Traceability | Add @sys: SYS.01.01.01 | ### Manual Review Required | Issue | Location | Recommendation | |-------|----------|----------------| | Circular dependency | T3 ↔ T5 | Restructure task dependencies | | Missing complexity | TASKS.01.18.04 | Add Estimated Complexity | ``` **Score Indicators**: - ✅ Green (>=90%): CODE-Ready - 🟡 Yellow (70-89%): Needs improvement - ❌ Red (<70%): Significant issues **Review Configuration**: ```yaml review_mode: enabled: true checks: - task_structure # All required task fields present - dependency_graph # Valid DAG, no cycles - priority_assignments # P0-P3 properly assigned - implementation_contracts # Contracts when 3+ dependencies - cumulative_tags # 9 required tags present - element_ids # TASKS.NN.18.SS, TASKS.NN.30.SS format - spec_alignment # Tasks trace to SPEC/TSPEC output: format: markdown # markdown, json, html include_recommendations: true include_fix_commands: true ``` ### Mode 3: Fix Mode (v2.1) Auto-repair existing TASKS documents with backup and content preservation. **Command**: ```bash # Fix single TASKS /doc-tasks-autopilot TASKS-01 --fix # Fix with backup /doc-tasks-autopilot TASKS-01 --fix --backup # Fix all TASKS /doc-tasks-autopilot docs/11_TASKS/ --fix --all # Fix specific categories only /doc-tasks-autopilot TASKS-01 --fix --only element_ids,tags # Dry-run fix (preview changes) /doc-tasks-autopilot TASKS-01 --fix --dry-run ``` **Fix Process**: ```mermaid flowchart TD A[Input: TASKS Document] --> B[Create Backup] B --> C[Load TASKS Content] C --> D[Run Review Analysis] D --> E{Auto-Fixable Issues?} E -->|No| F[Report: No Fixes Needed] E -->|Yes| G[Apply Fixes by Category] subgraph FixCategories["Fix Categories"] G --> H[Fix Element IDs] H --> I[Fix Priority Format] I --> J[Regenerate Dependency Graph] J --> K[Add Missing Tags] K --> L[Fix Section Structure] L --> M[Add Implementation Contracts] end M --> N[Validate Fixed Document] N --> O{Validation Passed?} O -->|No| P[Rollback to Backup] P --> Q[Report: Fix Failed] O -->|Yes| R[Save Fixed Document] R --> S[Generate Fix Report] ``` **TASKS-Specific Fix Categories**: | Category | Description | Auto-Fix Actions | |----------|-------------|------------------| | `element_ids` | Element ID format | Convert legacy patterns to TASKS.NN.TT.SS | | `priorities` | Priority format | Normalize to P0-P3 format | | `dependency_graph` | Task dependencies | Regenerate Mermaid graph from task refs | | `cumulative_tags` | Traceability tags | Add missing 9 upstream tags | | `sections` | Section structure | Add missing Overview, Contracts sections | | `contracts` | Implementation contracts | Generate when 3+ dependencies exist | | `circular_deps` | Circular dependencies | Flag only (manual restructure required) | **Element ID Migration** (Layer 11): | Legacy Pattern | New Format | Example | |----------------|------------|---------| | TASK-NNN | TASKS.NN.18.SS | TASK-001 → TASKS.01.18.01 | | TODO-NNN | TASKS.NN.18.SS | TODO-005 → TASKS.01.18.05 | | TI-NNN | TASKS.NN.30.SS | TI-001 → TASKS.01.30.01 | | ITEM-NNN | TASKS.NN.30.SS | ITEM-010 → TASKS.01.30.10 | **Content Preservation Rules**: | Content Type | Preservation Rule | |--------------|-------------------| | Custom task descriptions | Never delete, only enhance metadata | | Dependency relationships | Preserve logic, update format | | Priority assignments | Preserve values, normalize format | | Implementation contracts | Preserve all contract code | | Complexity estimates | Preserve values, add if missing | | SPEC/TSPEC references | Validate and update format only | **Fix Configuration**: ```yaml fix_mode: enabled: true backup: enabled: true location: "tmp/backups/" timestamp: true fix_categories: element_ids: true # Convert legacy ID patterns priorities: true # Normalize P0-P3 format dependency_graph: true # Regenerate Mermaid diagram cumulative_tags: true # Add 9 required tags sections: true # Add missing sections contracts: true # Generate when criteria met circular_deps: false # Manual only (flag but don't auto-fix) validation: post_fix: true # Validate after fixes rollback_on_fail: true # Restore backup if validation fails preserve: task_descriptions: true dependency_logic: true priority_values: true contract_code: true ``` **Fix Report Template**: ```markdown # TASKS Fix Report: TASKS-NN_{slug} ## Summary - **Backup Created**: tmp/backups/TASKS-NN_{slug}_20260209_143022.md - **Issues Fixed**: N of M auto-fixable issues - **Manual Review**: N issues flagged ## Fixes Applied ### Element ID Migration | Original | Fixed | Location | |----------|-------|----------| | TASK-001 | TASKS.01.18.01 | Line 45 | | TI-001 | TASKS.01.30.01 | Line 78 | ### Dependency Graph Regenerated - Updated Mermaid diagram with corrected task references - Total tasks: 12, Dependencies: 18 ### Cumulative Tags Added - @ears: EARS.01.25.001 (added) - @sys: SYS.01.01.01 (added) - @req: REQ.01.01.01 (added) ### Implementation Contracts Generated - Protocol interface: `IDataProcessor` (TASKS.01.18.03 has 4 dependencies) - Exception hierarchy: `TaskExecutionError` family ## Manual Review Required ### Circular Dependencies | Cycle | Tasks Involved | Recommendation | |-------|----------------|----------------| | Cycle 1 | T3 → T5 → T3 | Restructure: T5 should not depend on T3 | ## Validation Results - **CODE-Ready Score**: Before: 72% → After: 94% - **Validation Errors**: Before: 8 → After: 0 - **Status**: ✅ All auto-fixes validated ``` **Command Line Options** (Review/Fix Modes): | Option | Default | Description | |--------|---------|-------------| | `--review` | false | Run review mode only | | `--fix` | false | Run fix mode | | `--backup` | true | Create backup before fixing | | `--dry-run` | false | Preview fixes without applying | | `--only` | all | Comma-separated fix categories | | `--verbose` | false | Detailed output | | `--all` | false | Process all TASKS in directory | | `--output-format` | markdown | Report format (markdown, json) | | `--regenerate-contracts` | false | Force regenerate implementation contracts | --- ## Context Management ### Chunked Parallel Execution (MANDATORY) **CRITICAL**: To prevent conversation context overflow errors ("Prompt is too long", "Conversation too long"), all autopilot operations MUST follow chunked execution rules: **Chunk Size Limit**: Maximum 3 documents per chunk **Chunking Rules**: 1. **Chunk Formation**: Group SPEC/TSPEC-derived TASKS documents into chunks of maximum 3 at a time 2. **Sequential Chunk Processing**: Process one chunk at a time, completing all documents in a chunk before starting the next 3. **Context Pause**: After completing each chunk, provide a summary and pause for user acknowledgment 4. **Progress Tracking**: Display chunk progress (e.g., "Chunk 2/4: Processing TASKS-04, TASKS-05, TASKS-06...") **Why Chunking is Required**: - Prevents "Conversation too long" errors during batch processing - Allows context compaction between chunks - Enables recovery from failures without losing all progress - Provides natural checkpoints for user review **Chunk Completion Template**: ```markdown ## Chunk N/M Complete Generated: - TASKS-XX: CODE-Ready Score 94% - TASKS-YY: CODE-Ready Score 92% - TASKS-ZZ: CODE-Ready Score 95% Proceeding to next chunk... ``` --- ## Related Resources - **TASKS Skill**: `.claude/skills/doc-tasks/SKILL.md` - **TASKS Validator**: `.claude/skills/doc-tasks-validator/SKILL.md` - **Naming Standards**: `.claude/skills/doc-naming/SKILL.md` - **Quality Advisor**: `.claude/skills/quality-advisor/SKILL.md` - **TASKS Template**: `ai_dev_flow/11_TASKS/TASKS-TEMPLATE.md` - **Implementation Contracts**: `ai_dev_flow/11_TASKS/IMPLEMENTATION_CONTRACTS_GUIDE.md` --- ## Review Document Standards (v2.2) Review reports generated by this skill are formal project documents and MUST comply with shared standards. **Reference**: See `REVIEW_DOCUMENT_STANDARDS.md` in the skills directory for complete requirements. **Key Requirements**: 1. **Storage Location**: Same folder as the reviewed TASKS document 2. **File Naming**: `TASKS-NN.R_review_report_vNNN.md` (legacy) or `TASKS-NN.A_audit_report_vNNN.md` (preferred) 3. **YAML Frontmatter**: Required with `artifact_type: TASKS-REVIEW`, `layer: 11` 4. **Score Field**: `code_ready_score_claimed` / `code_ready_score_validated` 5. **Parent Reference**: Must link to parent TASKS document **Example Location** (ALWAYS use nested folders): ``` docs/11_TASKS/TASKS-03_f3_observability/ ├── TASKS-03_f3_observability.md # ← Main document ├── TASKS-03.R_review_report_v001.md # ← Review report ├── TASKS-03.F_fix_report_v001.md # ← Fix report └── .drift_cache.json # ← Drift cache ``` **Nested Folder Rule**: ALL TASKS use nested folders (`TASKS-NN_{slug}/`) regardless of size. This keeps companion files (review reports, fix reports, drift cache) organized with their parent document. --- ## Version History | Version | Date | Changes | |---------|------|---------| | 2.4 | 2026-02-11 | **Smart Document Detection**: Added automatic document type recognition; Self-type input (TASKS-NN) triggers review mode; Multiple upstream-type inputs (SPEC/TSPEC-NN) trigger generate-if-missing or find-and-review; Updated input patterns table with type-based actions | | 2.3 | 2026-02-10 | **Review & Fix Cycle**: Replaced Phase 5 with iterative Review -> Fix cycle using `doc-tasks-reviewer` and `doc-tasks-fixer`; Added `doc-tasks-fixer` skill dependency; Added iteration control (max 3 cycles); Added quality checks (task structure completeness, dependency graph validity, element ID compliance, CODE-Ready report); Added traceability matrix update step | | 2.2 | 2026-02-10 | Added Review Document Standards section; Review reports now stored alongside reviewed documents with proper YAML frontmatter and parent references | | 2.1 | 2026-02-09 | Added Mode 2: Review Mode for validation-only analysis with visual score indicators; Added Mode 3: Fix Mode for auto-repair with backup and content preservation; Element ID migration (TASK-NNN→TASKS.NN.18.SS, TI-NNN→TASKS.NN.30.SS); Implementation contracts auto-generation | | 1.0 | 2026-02-08 | Initial skill creation with 5-phase workflow; Integrated doc-naming, doc-tasks, quality-advisor, doc-tasks-validator; Added implementation contracts support |