--- name: content-refinement-agent description: Step 5 of the PaperOrchestra pipeline (arXiv:2604.05018). Iteratively refine drafts/paper.tex by simulating peer review and applying targeted revisions, with strict accept/revert halt rules. Maintains a worklog and snapshots each iteration so revert is real, not symbolic. TRIGGER when the orchestrator delegates Step 5 or when the user asks to "refine the draft", "iterate on the paper", or "run peer review on this paper". --- # Content Refinement Agent (Step 5) Faithful implementation of the Content Refinement Agent from PaperOrchestra (Song et al., 2026, arXiv:2604.05018, §4 Step 5, App. F.1 pp. 49–51). **Cost: ~5–7 LLM calls** (App. B), typically ~3 refinement iterations, each consisting of one reviewer call and one revision call. The paper highlights this step as one of the largest contributors to overall quality: refinement alone accounts for +19% (CVPR) and +22% (ICLR) absolute acceptance-rate improvement (Fig. 4). Get this step right. ## Inputs - `workspace/drafts/paper.tex` — output of Step 4 - `workspace/inputs/conference_guidelines.md` - `workspace/inputs/experimental_log.md` — used as ground truth for the hallucination check - `workspace/citation_pool.json` / `workspace/refs.bib` — the allowed bibliography ## Outputs - `workspace/refinement/iter1/`, `iter2/`, `iter3/` — per-iteration snapshots containing `paper.tex`, `paper.pdf`, `review.json`, `score.json` - `workspace/refinement/worklog.json` — append-only history of decisions - `workspace/final/paper.tex` and `workspace/final/paper.pdf` — copy of the best accepted snapshot ## The refinement loop ``` prev_score = score(paper.tex) # baseline from initial draft snapshot iter0/ for iter in 1..ITER_CAP (default 3): 1. simulate_review(paper.tex) → review.json (uses `references/reviewer-rubric.md` rubric) 2. apply_revision(paper.tex, review.json) → new_paper.tex (uses verbatim Refinement Agent prompt at `references/prompt.md`) 3. snapshot iter/ with new_paper.tex, review.json latexmk -pdf new_paper.tex → iter/paper.pdf 4. score(new_paper.tex) → curr_score 5. decide via score_delta.py: - if curr.overall > prev.overall: ACCEPT - elif curr.overall == prev.overall and net_subaxis ≥0: ACCEPT - else: REVERT 6. apply_worklog.py to append the decision 7. if REVERT or no actionable weaknesses or iter == ITER_CAP: HALT paper.tex ← new_paper.tex (only on ACCEPT) prev_score ← curr_score cp /paper.tex → workspace/final/paper.tex ``` The "best" snapshot at HALT is the one with the highest accepted overall score. On a REVERT halt, the best is the iteration immediately before the revert. ## Step-by-step ### 0. Snapshot the initial draft ```bash python skills/content-refinement-agent/scripts/snapshot.py \ --src workspace/drafts/paper.tex \ --dst workspace/refinement/iter0/ ``` This creates `iter0/paper.tex`. Then compile to `iter0/paper.pdf`: ```bash cd workspace/refinement/iter0/ && latexmk -pdf -interaction=nonstopmode paper.tex ``` Score it (see Step 1 below) → `iter0/score.json`. ### 1. Simulate peer review For each iteration N starting from 1: Load `references/reviewer-rubric.md` as the system prompt for the simulated reviewer call. The reviewer reads `iter/paper.pdf` (or `paper.tex` if your host LLM lacks PDF input) and produces a JSON of strengths, weaknesses, questions, and per-axis scores. The rubric is structured to mimic AgentReview (Jin et al., 2024) — the paper's chosen evaluator. We ship a faithful rubric in the references directory; the host agent's LLM does the actual reviewing. Save to `workspace/refinement/iter/review.json`. ### 2. Score the draft The reviewer call produces both qualitative feedback and a per-axis score: ```json { "axis_scores": { "scientific_depth": {"score": 65, "justification": "..."}, "technical_execution": {"score": 70, "justification": "..."}, "logical_flow": {"score": 60, "justification": "..."}, "writing_clarity": {"score": 55, "justification": "..."}, "evidence_presentation":{"score": 72, "justification": "..."}, "academic_style": {"score": 68, "justification": "..."} }, "overall_score": 64.5, "strengths": [...], "weaknesses": [...], "questions": [...] } ``` Save to `iter/score.json`. (Combined with `review.json` if your host emits one document; the schemas overlap.) ### 3. Apply revision Load the **verbatim Content Refinement Agent prompt** at `references/prompt.md`. Prepend the Anti-Leakage Prompt. Inputs: - `paper.tex` — current draft - `paper.pdf` — compiled PDF (multimodal context if available) - `conference_guidelines.md` - `experimental_log.md` — ground truth for numeric claims - `worklog.json` — history of previous changes - `citation_pool.json` — the allowed bibliography - `reviewer_feedback` — the JSON from Step 1 The prompt instructs the model to address weaknesses, integrate question answers, and emit two output blocks: 1. A worklog JSON `{addressed_weaknesses[], integrated_answers[], actions_taken[]}` 2. The full revised LaTeX code Save the revised LaTeX as `iter/paper.tex`. Append the worklog JSON to `workspace/refinement/worklog.json` via `apply_worklog.py`. ### 4. Compile and re-score ```bash cd workspace/refinement/iter/ && latexmk -pdf -interaction=nonstopmode paper.tex ``` Then re-run the simulated review on the new draft → updated `score.json` for the new iteration. (This is the "re-score after revision" call.) ### 5. Apply the accept/revert decision The calling loop must track `CONSECUTIVE_SMALL` (starts at 0) and pass it on each call so `score_delta.py` can detect the plateau: ```bash python skills/content-refinement-agent/scripts/score_delta.py \ --prev workspace/refinement/iter/score.json \ --curr workspace/refinement/iter/score.json \ --plateau-threshold 1.0 \ --plateau-streak 3 \ --consecutive-small $CONSECUTIVE_SMALL \ > workspace/refinement/iter/delta.json EXIT=$? # Update streak for next iteration: CONSECUTIVE_SMALL=$(python3 -c " import json d = json.load(open('workspace/refinement/iter/delta.json')) print(d['consecutive_small']) ") ``` Exit codes: - `0` — ACCEPT (overall improved or tied with non-negative net sub-axis, no plateau) - `1` — REVERT (overall decreased) - `2` — REVERT (tied overall, but net sub-axis change negative) - `4` — HALT_PLATEAU (accepted but N consecutive iterations below threshold — stop early) Behavior: - **ACCEPT (exit 0)**: keep `iter/paper.tex` as the new best. Continue to iter N+1. - **REVERT (exit 1 or 2)**: copy `iter/paper.tex` back as canonical, halt. - **HALT_PLATEAU (exit 4)**: keep current (it was accepted), but stop — further iterations are unlikely to yield meaningful gains. In practice ~85% of refinement gain comes in iteration 1; the plateau fires when subsequent iterations improve by less than 1 point for 3 consecutive rounds. Always log the decision via `apply_worklog.py --decision ...`. ### 6. Halt rules Halt the loop when ANY of these is true: 1. Iteration count reaches `ITER_CAP` (default 3). 2. `score_delta.py` returned exit code 1 or 2 (REVERT). 3. The simulated reviewer's `weaknesses` list is empty (no actionable feedback to apply). 4. `score_delta.py` returned exit code 4 (HALT_PLATEAU — plateau early-stop). ### 7. Promote the best snapshot Identify the iteration with the highest accepted `overall_score` (this may be the latest accepted iteration, OR an earlier one if a later iteration was reverted). Copy: ```bash cp workspace/refinement/iter/paper.tex workspace/final/paper.tex cp workspace/refinement/iter/paper.pdf workspace/final/paper.pdf ``` Then in the final report, tell the user: - How many iterations were run - The final overall score - The score trajectory (e.g., "iter0 64.5 → iter1 67.3 (accept) → iter2 69.1 (accept) → iter3 68.9 (revert, halt)") - Which iteration was promoted ## Critical safety constraints (App. F.1 page 50–51) The paper explicitly notes that early versions of the Refinement Agent "exploited the automated reviewer's scoring function by superficially listing missing baselines as limitations to artificially inflate acceptance scores." The verbatim prompt forbids this. **You must honor it:** - **Ignore reviewer requests for new experiments, ablations, or baselines.** The Refinement Agent's job is presentation, not new science. If the reviewer asks for missing data, simply skip those points — do NOT add fabricated experiments, do NOT add a "future work" item promising them. - **Never explicitly state a limitation.** The phrase "we acknowledge as a limitation that..." is forbidden. The model can address weaknesses through clearer explanation, but must not game the evaluator by listing them defensively. - **All numeric claims MUST be verified against `experimental_log.md`.** The agent cannot introduce new numbers, only re-present existing ones. These rules prevent reward hacking and keep the refinement loop honest. ## Resources - `references/prompt.md` — verbatim Content Refinement Agent prompt from App. F.1 - `references/reviewer-rubric.md` — AgentReview-style scoring rubric (6 axes) - `references/halt-rules.md` — accept/revert/halt logic in formal pseudocode - `references/safe-revision-rules.md` — anti-reward-hack constraints - `scripts/score_delta.py` — accept/revert decision from two score JSONs - `scripts/apply_worklog.py` — append iteration entries to worklog.json - `scripts/snapshot.py` — copy paper.tex/paper.pdf into iter/ for rollback