You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
feat(debrief): scaffold dogfood debriefs from deterministic artifacts
Closes the debrief asymmetry flagged in docs/v4-release/
dogfood-shade-20260413.md §6: failed dogfood runs prompt findings-doc
authorship, successful runs leave only artifacts. The Apr 13 atlas-shade
run was the canonical example — ran cleanly end-to-end but went
undocumented for a full day until a later session dug the artifacts out
of untracked paths and a volatile /tmp tmux snapshot.
New subcommand: `script.py debrief --from-tasks` reads tasks.json +
complexity-report + PRD and emits a markdown skeleton to
docs/v4-release/dogfood-<slug>-<YYYYMMDD>.md. Mechanical sections are
filled deterministically (task count, dep coverage %, priority
distribution, complexity stats, top-5 complexity table, goal extracted
from PRD executive summary). Judgment sections (what worked / what
broke / comparison / meta) stay as explicit TODO placeholders — those
require authorship, not automation.
Validated on the real atlas-shade artifacts: emits 12 tasks, 8 with deps
(66.7%), avg complexity 4.58, top-3 matching the earlier manual audit
(#12=8, #10=7, #7=6). Goal extraction pulls the Executive Summary
opener, which required scanning past the frontmatter metadata block
(** Version, ** Date, ** Sentinel) instead of stopping at the first ##.
Flags:
--from-tasks / --complexity-report / --prd (default taskmaster paths)
--slug (required — filename anchor)
--output-dir (default docs/v4-release)
--grade (optional validate-prd grade to embed)
--force (overwrite existing output)
Tests: 3 new cases covering happy path (mechanical sections filled +
TODO placeholders present), missing-tasks-json error, and --force
overwrite semantics. Full suite: 225 passed / 1 skip.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
TODO — one or two paragraphs. Look for: anything the user / operator had
1144
+
to work around, hook conflicts, session retirement, flags that didn't
1145
+
work, naming mismatches. If the run was clean, say so explicitly and
1146
+
note what you tried to break.
1147
+
1148
+
## 5. Comparison / context
1149
+
1150
+
TODO — reference related runs (self-dogfood, sibling projects). What does
1151
+
this run tell you that the others didn't?
1152
+
1153
+
## 6. Meta-findings
1154
+
1155
+
TODO — patterns worth capturing for next time. Leave blank if none.
1156
+
1157
+
---
1158
+
1159
+
_Scaffolded by `script.py debrief --from-tasks` on {datetime.now().strftime("%Y-%m-%d %H:%M")}. TODO sections need human authorship — the mechanical sections are load-bearing; the judgment sections are what you actually have to write._
0 commit comments