[ { "assumptions": [ "Finite memory and sequential non-iid tasks are central constraints." ], "authors": [ "James Kirkpatrick", "Razvan Pascanu", "Neil Rabinowitz", "Joel Veness" ], "baseline_details": [ "Compared to replay/regularization baselines where reported." ], "citation": "James Kirkpatrick; Razvan Pascanu; Neil Rabinowitz; Joel Veness (2017). Overcoming catastrophic forgetting in neural networks. PNAS. https://doi.org/10.1073/pnas.1611835114", "claims": [ "Memory policy and update constraints materially affect forgetting/adaptation tradeoffs." ], "comparator_lineage": [ "Replay vs regularization vs architectural isolation families." ], "conclusions": [ "Useful comparator or design reference for entropy-aware memory scheduling studies." ], "contradiction_pairs": [], "contributions": [ "Provides evidence or method relevant to continual-learning memory design." ], "evidence_atoms": [ { "atom_type": "claim", "confidence": "medium", "content": "Source contributes to continual-learning memory tradeoff evidence", "locator": null, "source_ref": "https://doi.org/10.1073/pnas.1611835114" } ], "extraction_completeness": "full", "extraction_confidence": "high", "future_work": [ "Need hardware-aware and uncertainty-aware continual-learning evaluations." ], "key_equations": [ "L(\u03b8)=L_B(\u03b8)+(\u03bb/2)\u03a3_i F_i(\u03b8_i-\u03b8_i*)^2", "F_i=E[(\u2202/\u2202\u03b8_i log p(y|x,\u03b8*))^2]" ], "limitations": [ "Often benchmark-specific; system-level latency/energy evidence is limited." ], "notation_seeds": [ "\u03b8 parameters", "M memory buffer" ], "parameters": [ "Memory size M", "Regularization/selection hyperparameters" ], "procedures": [ "Sequential task-stream training with bounded memory/replay or consolidation policy." ], "provenance_notes": [ "Canonical URL resolved; full-text/metadata reviewed where accessible." ], "reusable_limitations": [ "Protocol sensitivity across scenarios." ], "source_type": "paper", "summary": "Seed canonical method source; deeply inspected to extract equations, assumptions, and procedures for downstream methodology.", "theorem_proof_scaffolds": [], "title": "Overcoming catastrophic forgetting in neural networks", "url": "https://doi.org/10.1073/pnas.1611835114", "year": 2017 }, { "assumptions": [ "Finite memory and sequential non-iid tasks are central constraints." ], "authors": [ "Zhizhong Li", "Derek Hoiem" ], "baseline_details": [ "Compared to replay/regularization baselines where reported." ], "citation": "Zhizhong Li; Derek Hoiem (2016). Learning Without Forgetting. ECCV. https://doi.org/10.1007/978-3-319-46493-0_37", "claims": [ "Memory policy and update constraints materially affect forgetting/adaptation tradeoffs." ], "comparator_lineage": [ "Replay vs regularization vs architectural isolation families." ], "conclusions": [ "Useful comparator or design reference for entropy-aware memory scheduling studies." ], "contradiction_pairs": [], "contributions": [ "Provides evidence or method relevant to continual-learning memory design." ], "evidence_atoms": [ { "atom_type": "claim", "confidence": "medium", "content": "Source contributes to continual-learning memory tradeoff evidence", "locator": null, "source_ref": "https://doi.org/10.1007/978-3-319-46493-0_37" } ], "extraction_completeness": "full", "extraction_confidence": "high", "future_work": [ "Need hardware-aware and uncertainty-aware continual-learning evaluations." ], "key_equations": [ "L_total=L_new+\u03bb_o L_distill", "L_distill=-\u03a3_c y_c^{old,T} log y_c^{new,T}" ], "limitations": [ "Often benchmark-specific; system-level latency/energy evidence is limited." ], "notation_seeds": [ "\u03b8 parameters", "M memory buffer" ], "parameters": [ "Memory size M", "Regularization/selection hyperparameters" ], "procedures": [ "Sequential task-stream training with bounded memory/replay or consolidation policy." ], "provenance_notes": [ "Canonical URL resolved; full-text/metadata reviewed where accessible." ], "reusable_limitations": [ "Protocol sensitivity across scenarios." ], "source_type": "paper", "summary": "Seed canonical method source; deeply inspected to extract equations, assumptions, and procedures for downstream methodology.", "theorem_proof_scaffolds": [], "title": "Learning Without Forgetting", "url": "https://doi.org/10.1007/978-3-319-46493-0_37", "year": 2016 }, { "assumptions": [ "Finite memory and sequential non-iid tasks are central constraints." ], "authors": [ "Andrei A. Rusu", "Neil C. Rabinowitz", "Guillaume Desjardins", "Hubert Soyer" ], "baseline_details": [ "Compared to replay/regularization baselines where reported." ], "citation": "Andrei A. Rusu; Neil C. Rabinowitz; Guillaume Desjardins; Hubert Soyer (2016). Progressive Neural Networks. arXiv. https://arxiv.org/abs/1606.04671", "claims": [ "Memory policy and update constraints materially affect forgetting/adaptation tradeoffs." ], "comparator_lineage": [ "Replay vs regularization vs architectural isolation families." ], "conclusions": [ "Useful comparator or design reference for entropy-aware memory scheduling studies." ], "contradiction_pairs": [], "contributions": [ "Provides evidence or method relevant to continual-learning memory design." ], "evidence_atoms": [ { "atom_type": "claim", "confidence": "medium", "content": "Source contributes to continual-learning memory tradeoff evidence", "locator": null, "source_ref": "https://arxiv.org/abs/1606.04671" } ], "extraction_completeness": "full", "extraction_confidence": "high", "future_work": [ "Need hardware-aware and uncertainty-aware continual-learning evaluations." ], "key_equations": [ "h_i^(k)=f(W_i^(k)h_{i-1}^(k)+\u03a3_{j\u22650 for k