% This file was adapted from ICLR2022_conference.tex example provided for the ICLR conference \documentclass{article} % For LaTeX2e \usepackage{conference,times} \usepackage{easyReview} \usepackage{algorithm} \usepackage{algorithmic} % Optional math commands from https://github.com/goodfeli/dlbook_notation. \input{math_commands.tex} \usepackage{amsthm,amssymb} \newtheorem{theorem}{Theorem}[section] \newtheorem{corollary}{Corollary}[theorem] \newtheorem{lemma}[theorem]{Lemma} \newtheorem{definition}[theorem]{Definition} % Please leave these options as they are \usepackage{hyperref} \hypersetup{ colorlinks=true, linkcolor=red, filecolor=magenta, urlcolor=blue, citecolor=purple, pdftitle={Entropy-Aware Memory Systems for Continual Learning: Balancing Neuroplasticity and Stability Under Stochastic Workloads}, pdfpagemode=FullScreen, } \title{Entropy-Aware Memory Systems for Continual Learning:\\Balancing Neuroplasticity and Stability Under Stochastic Workloads} \author{Marius-Constantin Dinu \\ Independent Systems Research \\ \texttt{marius.dinu@research.invalid}} \begin{document} \maketitle \begin{abstract} Continual learning systems are increasingly limited by memory behavior rather than arithmetic throughput: the same memory substrate must support stable recall and adaptive updates while respecting strict latency and energy constraints. We study this bottleneck through an entropy-parameterized access model that treats deterministic retrieval as a limit case of stochastic replay, then couple that model to an entropy-conditioned projection rule that recovers classical A-GEM behavior when slack is zero. The manuscript provides a formal derivation of the limit identity and shifted projection closed form, followed by protocol-locked validation evidence under a CPU-only experimental contract. Across surrogate continual-learning streams, entropy-aware replay improves frontier-area metrics against static replay with a reported 95\% confidence interval of approximately $[0.0533,0.0598]$, while entropy-conditioned projection shows a positive aggregate crossover interval (95\% confidence interval approximately $[0.028,0.194]$) relative to fixed A-GEM in validated operating regions. Symbolic theorem audits pass all recorded checks in the current run, and protocol compliance remains high under matched-memory controls. The resulting contribution is a hybrid formal-and-empirical framework that clarifies where entropy-aware control is supported, where it is conditional, and how to translate those boundaries into implementable continual-learning system design. \end{abstract} \section{Introduction} Catastrophic forgetting remains the central technical barrier to reliable continual learning, but the practical constraint profile has shifted. In many deployment settings, especially edge and shared-server systems, the dominant bottleneck is no longer only model capacity or gradient noise; it is the memory system that mediates retrieval, replay, bookkeeping, and measurement under a bounded compute envelope. Classical regularization methods penalize parameter drift to preserve previous tasks \citep{ewc_2017,si_2017,mas_2018,rwalk_2018}, distillation approaches stabilize outputs across tasks \citep{lwf_2016}, and replay-based families preserve evidence directly in memory \citep{er_2019,gem_2017,agem_2019,derpp_2020,gdumb_2020}. Yet these families are often discussed as if they solve orthogonal problems, even though all of them implicitly choose a memory-access policy. This framing gap matters for systems work. A deterministic retrieval path that always fetches one exemplar, a stochastic replay path that samples from a buffer, and a constrained-projection path that uses replay gradients are usually treated as separate algorithmic motifs. In reality, each motif induces a distribution over memory operations and therefore a measurable entropy profile, with direct consequences for retention, plasticity, per-step latency, and energy draw. Existing benchmark infrastructure has improved reproducibility \citep{scenario_2019,delange_2021,avalanche_2023,mammoth_2024,baselines_repo}, but literature still provides limited standardized evidence on CPU latency and energy outcomes, particularly for newer replay variants \citep{amr_2024,sprint_2024,saliency_2024,adiabatic_2024,prototype_2025,meta_replay_2025,latent_2025}. The result is a recurring claim-evidence mismatch: methods are presented as efficient while systems-level evidence remains sparse or protocol-inconsistent. This paper addresses that mismatch by coupling a formal access-distribution model with evidence assembly that is explicit about support and caveats. We retain the strongest comparator lineage from continual-learning practice (ER, EWC, LwF, GEM/A-GEM, DER++, and related baselines) \citep{ewc_2017,lwf_2016,gem_2017,agem_2019,er_2019,derpp_2020,gdumb_2020}. We then make two central design moves. First, we define an entropy-parameterized memory-access process that unifies deterministic retrieval and stochastic replay. Second, we define an entropy-conditioned shifted projection operator that extends A-GEM geometry while recovering A-GEM exactly at zero slack. Both moves are anchored to formal derivations and then evaluated through protocol-locked evidence that includes theorem-audit artifacts, aggregate tables, and explicit negative-result logs. Beyond a single algorithmic proposal, the contribution is methodological: we separate what is inherited from prior work, what is newly defined in this manuscript, and what remains conditionally supported. This distinction is crucial for trustworthy scientific reporting in open-question settings where formal closure and empirical closure can progress at different rates. \paragraph{Contributions.} \begin{itemize} \item We define a unified memory-access formalism in which deterministic retrieval is the entropy-zero limit of stochastic replay and provide a complete proof of the limit identity under explicit assumptions. \item We derive an entropy-conditioned shifted projection rule that includes a closed-form solution, exact A-GEM recovery at zero slack, and a monotonic correction-norm property. \item We present a protocol-locked validation assembly that links each major claim to specific quantitative artifacts, symbolic checks, and caveats rather than reporting aggregate outcomes without provenance. \item We provide a systems-oriented interpretation of positive results, counterexample slices, and failure boundaries to guide practical continual-learning deployment under CPU memory, latency, and energy constraints. \end{itemize} The broader relevance of this framing extends beyond the specific continual-learning benchmarks used in this run. Any adaptive system that alternates between deterministic cache hits and stochastic refresh operations faces a similar tension: concentrating access can stabilize behavior but may reduce exploration, while diversifying access can improve adaptation but may increase variance and operating cost. This includes online recommendation systems with replay buffers, stream-processing pipelines with bounded state stores, and hardware-aware training loops that dynamically budget memory traffic. By explicitly treating access entropy as a control variable rather than an incidental statistic, we obtain a language that is both mathematically analyzable and operationally measurable. That dual property is the reason we emphasize claim-evidence closure in addition to formal derivation: the contribution should remain inspectable by both theorists and systems practitioners. A second motivation is reproducible disagreement resolution. The continual-learning literature contains recurrent contradictions that are difficult to resolve from aggregate accuracy plots alone: replay versus regularization sufficiency, constrained projection versus simple rehearsal efficiency, and compression gains versus collapse risk. If those contradictions are framed only as leaderboards, one can report improvement without clarifying whether the gain is causal, protocol-specific, or boundary-limited. We therefore position this manuscript as a contribution to scientific structure: each major claim is paired with a derivation dependency, comparator lineage, executable check path, and caveat condition. The resulting narrative is intentionally stricter than typical benchmark reporting because the target problem is a systems bottleneck where hidden assumptions are often the dominant source of irreproducibility. \section{Related Work and Novelty Boundary} \subsection{Regularization and Distillation as Memory-Light Stability Controls} Parameter-consolidation and output-preservation methods remain core references for continual learning because they operate with minimal episodic memory \citep{ewc_2017,lwf_2016,si_2017,mas_2018,rwalk_2018}. EWC-style Fisher penalties define importance-weighted resistance to parameter drift \citep{ewc_2017}, while LwF-style distillation regularizes current outputs toward previous-task behavior \citep{lwf_2016}. These methods are compelling when explicit replay storage is costly, but they can underperform in long heterogeneous streams where update restriction trades away adaptation. The literature increasingly recognizes this regime dependence \citep{delange_2021,scenario_2019}, but consistent systems-level comparisons to replay methods remain incomplete. From a novelty-boundary perspective, regularization and distillation methods provide two important inherited ideas for this manuscript. First, they clarify that retention can be represented as a continuously weighted penalty rather than a binary constraint; this informs our use of weighted objective terms in \eqref{eq:control_objective}. Second, they highlight the risk of overly rigid stabilization under nonstationary streams, which motivates entropy-conditioned rather than fixed-strength control. We therefore reuse these families as scientific anchors and controls, not as strawman baselines. The relevant comparison question is not whether replay or regularization is universally superior, but whether an entropy-aware controller can outperform fixed policies under matched system budgets while preserving explicit stability boundaries. \subsection{Replay, Constraint Geometry, and Retrieval Policy} Replay families expose a rich geometry of constraints and sampling policies. GEM projects updates against multiple task constraints \citep{gem_2017}, whereas A-GEM uses a single reference gradient for efficiency \citep{agem_2019}. ER and derivatives emphasize buffer maintenance and sample utility \citep{er_2019,mir_2019,derpp_2020,gdumb_2020,rehearsal_2021,rainbow_2021}. Recent variants introduce saliency, curriculum, or adaptive scheduling \citep{amr_2024,sprint_2024,saliency_2024,enhanced_er_2023,mgser_2024,coord_2023}. The central contradiction is unresolved: stronger geometric constraints can protect old tasks but may increase projection overhead, while simpler replay can be computationally cheaper but more sensitive to sampling quality. This contradiction motivates our entropy-conditioned projection analysis and explicit overhead-focused evidence reporting. This contradiction is especially relevant under CPU-only constraints because projection cost and memory-traffic cost can interact nonlinearly. A method that appears superior in GPU-heavy settings may lose its advantage when gradient projection and data movement dominate wall-clock behavior. The shifted projection in \eqref{eq:shifted_projection_closed_form} is therefore chosen to preserve an interpretable relation to A-GEM while exposing a tunable slack variable tied to measured entropy context. The goal is not to maximize one metric in isolation; it is to preserve retention control while reducing the risk that constraint handling itself becomes the bottleneck. \subsection{Memory Representation and Dual-Memory Lines} A parallel thread explores memory representation rather than only sampling policy. Generative replay, dual-memory strategies, and compressed or prototype memories seek better retention-per-byte under fixed budgets \citep{dgr_2017,brain_replay_2020,hippocampal_2023,chameleon_2023,snn_2023,sketch_2024,prototype_2025,meta_replay_2025,latent_2025}. Modular-expansion approaches reduce interference by architectural isolation \citep{progressive_2016,three_types_2022,icarl_2017}, but can raise memory or routing overhead in deployment. In practice, system designers often need both perspectives: representation efficiency and access control under non-ideal hardware constraints. The representation thread also exposes a practical lesson that motivates our mixed-support reporting policy. Compression and latent replay can produce impressive average gains while simultaneously creating brittle failure regions under shift or class imbalance. Such failure regions are often underreported when papers focus on average metrics and omit stress slices. By carrying complementary representation results as a mixed-support claim rather than forcing a binary success narrative, we preserve scientific information that is valuable for deployment decisions. This is why the manuscript includes explicit discussion of collapse boundaries and does not elevate complementary results to primary claims. \subsection{Benchmark Infrastructure, Comparability, and Gaps} The field now has stronger tooling for reproducibility and scenario standardization \citep{scenario_2019,delange_2021,avalanche_2023,mammoth_2024,baselines_repo}, yet two gaps still block decisive systems claims. First, CPU latency and energy are under-reported relative to accuracy and forgetting metrics, especially in newer papers with heterogeneous protocol choices. Second, deterministic retrieval and stochastic replay are usually narrated separately instead of within one formal access model. Our novelty boundary follows directly: we do not claim universal superiority across all continual-learning methods; instead, we contribute a defensible boundary in which entropy-aware access and entropy-conditioned projection are formally derived, empirically audited, and explicitly caveated. This boundary is intentionally conservative. We avoid claiming that entropy alone explains continual-learning outcomes, that surrogate evidence replaces real-data closure, or that one parameterization subsumes all replay paradigms. Instead, we claim that an entropy-aware formulation creates a common axis on which deterministic and stochastic access can be analyzed jointly, and that this joint axis enables cleaner alignment between formal proofs, algorithm design, and systems-level measurements. The value of the contribution is therefore coherence under scrutiny: when a claim is strong, evidence is explicit; when a claim is conditional, caveats are explicit; when evidence is missing, unresolved gaps are explicit. This emphasis on explicit evidence termination also supports downstream peer review because each argument can be traced to a derivation, an experiment summary, a symbolic check, or a declared caveat without ambiguity. \section{Problem Setting and Notation} We consider a sequential task stream $\{\train_t\}_{t=1}^{T}$ with samples $(x_t,y_t)$, model parameters $\vtheta_t\in\mathbb{R}^d$, and bounded episodic memory $B_t$ with $|B_t|\le M$. The objective is to maintain old-task performance (retention) while learning new tasks (plasticity) under system constraints. \begin{definition}[Entropy-Parameterized Access Process] For step $t$, let $i_t^*$ denote a deterministic retrieval index and $q_t$ a stochastic replay distribution over non-selected memory elements. We define \begin{equation} \label{eq:access_mixture} p_t(i)=\alpha_t\,\delta_{i=i_t^*} + (1-\alpha_t)\,q_t(i),\qquad 0\le \alpha_t\le 1, \end{equation} where $\delta$ is the Kronecker mass and $q_t(i_t^*)=0$ under exact-decomposition conditions. \end{definition} The access entropy is \begin{equation} \label{eq:entropy_def} H_t=-\sum_{i\in B_t} p_t(i)\log p_t(i), \end{equation} so larger $H_t$ corresponds to broader stochastic access pressure and smaller $H_t$ corresponds to deterministic concentration. We use a composite systems-aware objective, introduced in this work, to align learning and operating costs: \begin{equation} \label{eq:control_objective} \max_{\{\alpha_t,r_t\}_{t=1}^T}\sum_{t=1}^{T}\left(w_R R_t + w_P P_t - w_L\,\mathrm{Lat}_t - w_E\,\mathrm{En}_t\right), \end{equation} subject to $|B_t|\le M$ and replay-ratio constraints $r_t\in[r_{\min},r_{\max}]$. Here $R_t$ and $P_t$ are retention and plasticity summaries, while $\mathrm{Lat}_t$ and $\mathrm{En}_t$ are per-step latency and energy measurements under a fixed protocol. The optimization objects are therefore explicit: decision variables are $\{\alpha_t,r_t\}_{t=1}^T$ and, when projection is active, the corrected gradient $\tilde{\bm g}_t$; the feasible set is defined by memory budget, replay-ratio bounds, and projection constraints; and the optimality criterion is maximization of the weighted objective in \eqref{eq:control_objective} over the stream horizon. In this manuscript, we evaluate practical optimality through Pareto-improvement evidence and constraint-compliance checks rather than claiming closed-form global optimal solutions for the full nonstationary control problem. This choice is consistent with the hybrid emphasis: rigorous local derivations where tractable, empirical dominance checks where global analytic closure is not available. For constrained updates we adopt the A-GEM/GEM lineage \citep{gem_2017,agem_2019} and define a manuscript-specific entropy slack $\kappa(H_t)\ge 0$: \begin{equation} \label{eq:shifted_projection_problem} \min_{\tilde{\bm g}_t}\frac{1}{2}\|\tilde{\bm g}_t-\bm g_t\|_2^2\quad\text{s.t.}\quad\langle \tilde{\bm g}_t,\bm g_{\mathrm{ref},t}\rangle + \kappa(H_t)\ge 0. \end{equation} \paragraph{Assumptions.} Our main derivations use: (A1) exact decomposition condition $q_t(i_t^*)=0$ unless corrected form is used; (A2) deterministic tie-breaking for $i_t^*$; (A3) measurable latency/energy instrumentation under matched protocol; (A4) matched memory budget across comparators. For projection analysis we also use (B1) $\bm g_{\mathrm{ref},t}\neq \bm 0$ when projection is active; (B2) measurable nonnegative monotone $\kappa(H)$; (B3) forgetting increment bound with stream-specific $L_F>0$; (B4) matched comparator protocol. These conditions are inherited from formal handoff artifacts and are audited explicitly in symbolic and empirical checks. \section{Entropy-Aware Method} \subsection{Entropy-Conditioned Replay Control} The first method component selects replay intensity and access concentration jointly. The control variable $\alpha_t$ in \eqref{eq:access_mixture} controls deterministic concentration, while replay ratio $r_t$ controls data-mix pressure in \eqref{eq:control_objective}. We use entropy regimes as control context: low entropy favors deterministic retrieval, medium entropy uses mixed access, and high entropy increases replay diversity while constraining latency overhead. This design is motivated by replay scheduling studies \citep{mir_2019,rainbow_2021,saliency_2024,sprint_2024,leitner_2024} and by systems constraints emphasized in benchmark infrastructure \citep{avalanche_2023,mammoth_2024}. Importantly, this manuscript does not claim entropy is the only useful control signal. Instead, entropy is used as a measurable summary that connects formal access analysis with practical scheduler implementation. In the validation assembly, entropy targets are swept across multiple regimes, and support is judged by confidence intervals and failure logs, not by single-run best points. Component choices are motivated by both prior literature and operational constraints. We select entropy as the principal context variable because it is directly tied to the formal access model in \eqref{eq:access_mixture}--\eqref{eq:entropy_def}, unlike ad hoc heuristics that are difficult to interpret analytically. We retain replay-ratio bounds because fixed memory and bounded latency are hard constraints, not soft preferences. We include retention and plasticity jointly in \eqref{eq:control_objective} to prevent one-sided optimization that appears strong on final accuracy while hiding forward-transfer or forgetting costs. Finally, we keep the controller modular so it can be attached to established replay baselines without rewriting full training pipelines. \subsection{Entropy-Conditioned Shifted Projection} The second component extends A-GEM-style constraint handling \citep{agem_2019}. Solving \eqref{eq:shifted_projection_problem} yields the closed-form update \begin{equation} \label{eq:shifted_projection_closed_form} \tilde{\bm g}_t = \bm g_t - \max\!\left(0,\frac{-\kappa_t-\langle \bm g_t,\bm g_{\mathrm{ref},t}\rangle}{\|\bm g_{\mathrm{ref},t}\|_2^2}\right)\bm g_{\mathrm{ref},t},\quad \kappa_t=\kappa(H_t). \end{equation} When $\kappa_t=0$, \eqref{eq:shifted_projection_closed_form} reduces exactly to A-GEM. This recovery property is central: the extension is conservative with respect to a well-established baseline, rather than replacing it with an unrelated mechanism. Under a Lipschitz-style forgetting envelope, the slack-dependent term induces a conditional admissibility bound \begin{equation} \label{eq:forgetting_bound} \Delta F_t \le \frac{L_F\,\kappa_t}{\|\bm g_{\mathrm{ref},t}\|_2^2}, \end{equation} which is interpreted as a calibration condition, not a universal guarantee. \subsection{Module Responsibilities and Complexity} The architecture has four modules relevant to reproducibility and systems interpretation. \textbf{Access controller.} Produces $(\alpha_t,r_t)$ from entropy statistics and protocol limits. Its complexity is constant in buffer size for control updates, excluding metric collection. \textbf{Replay sampler.} Instantiates $p_t$ and retrieves samples from $B_t$. Complexity scales with the selected replay batch size and any optional weighting. \textbf{Projection operator.} Applies \eqref{eq:shifted_projection_closed_form} when gradient conflict is active. Relative to fixed A-GEM, overhead is dominated by one additional scalar slack term and monitoring of $H_t$. \textbf{Protocol auditor.} Logs latency, energy, memory-budget compliance, and theorem-check hooks. This module is necessary for claim-evidence closure under systems constraints. The modular decomposition also supports ablation clarity. If gains are observed, we can test whether they come from access scheduling, projection geometry, or protocol artifacts by disabling one module at a time under the same instrumentation contract. If gains disappear under module removal, causal interpretation is strengthened; if gains persist, the model may be over-parameterized and can be simplified. This ablation-first logic is one reason we treat architecture description as part of scientific method rather than implementation trivia. \subsection{Algorithmic Workflow} \begin{algorithm}[t] \caption{Entropy-Aware Continual Update with Shifted Projection} \label{alg:entropy_update} \begin{algorithmic}[1] \STATE \textbf{Input:} Stream sample $(x_t,y_t)$, memory $B_t$, parameters $\vtheta_t$, reference gradient $\bm g_{\mathrm{ref},t}$ \STATE Estimate entropy context $H_t$ from current access statistics \STATE Compute controller outputs $(\alpha_t,r_t)$ and build $p_t$ via \eqref{eq:access_mixture} \STATE Draw replay batch under $p_t$ and form task-plus-replay gradient $\bm g_t$ \STATE Set $\kappa_t\leftarrow\kappa(H_t)$ and compute projected gradient with \eqref{eq:shifted_projection_closed_form} \STATE Apply parameter update $\vtheta_{t+1}\leftarrow\vtheta_t-\eta\tilde{\bm g}_t$ \STATE Update memory subject to $|B_t|\le M$ and replay-ratio bounds \STATE Log $R_t,P_t,\mathrm{Lat}_t,\mathrm{En}_t$ and protocol compliance for audit \end{algorithmic} \end{algorithm} Algorithm~\ref{alg:entropy_update} highlights that all method choices are measurable under the same protocol contract. This point is critical in a hybrid formal-empirical manuscript: without matched instrumentation, neither support nor refutation of systems claims is interpretable. \section{Formal Analysis} \subsection{Deterministic Retrieval as an Entropy Limit} \begin{theorem}[Deterministic Limit of Entropy-Parameterized Access] \label{thm:det_limit} Assume \eqref{eq:access_mixture} with $q_t(i_t^*)=0$ and deterministic tie-breaking for $i_t^*$. Then as $\alpha_t\to 1$, the access entropy $H_t$ in \eqref{eq:entropy_def} converges to zero and $p_t$ converges in total variation to the deterministic mass at $i_t^*$. \end{theorem} \begin{proof} Substituting \eqref{eq:access_mixture} into \eqref{eq:entropy_def} gives \[ H_t = -\alpha_t\log\alpha_t -(1-\alpha_t)\sum_{i\neq i_t^*}q_t(i)\big(\log(1-\alpha_t)+\log q_t(i)\big), \] which simplifies to \[ H_t = h_2(\alpha_t)+(1-\alpha_t)H(q_t), \] where $h_2$ is the binary entropy and $H(q_t)$ is the entropy of $q_t$. Since $0\le H(q_t)\le \log(|B_t|-1)$ and $h_2(\alpha_t)\to0$ as $\alpha_t\to1$, both terms converge to zero, so $H_t\to0$. For total variation, note $p_t(i_t^*)=\alpha_t$ and $p_t(i)=(1-\alpha_t)q_t(i)$ for $i\neq i_t^*$. Therefore, \[ \|p_t-\delta_{i_t^*}\|_{\mathrm{TV}}=\tfrac{1}{2}\Big(|\alpha_t-1|+\sum_{i\neq i_t^*}(1-\alpha_t)q_t(i)\Big)=1-\alpha_t\to0. \] Hence deterministic retrieval is the entropy-zero limit of the unified access process. \end{proof} \subsection{Shifted Projection with A-GEM Recovery} \begin{lemma}[Closed Form for Entropy-Conditioned Shifted Projection] \label{lem:shifted_proj} Assume $\bm g_{\mathrm{ref},t}\neq\bm 0$. The optimizer of \eqref{eq:shifted_projection_problem} is exactly \eqref{eq:shifted_projection_closed_form}. \end{lemma} \begin{proof} The objective in \eqref{eq:shifted_projection_problem} is strictly convex and the feasible set is a closed half-space, so a unique minimizer exists. With Lagrange multiplier $\lambda\ge0$, \[ \mathcal{L}(\tilde{\bm g},\lambda)=\tfrac12\|\tilde{\bm g}-\bm g_t\|_2^2-\lambda\big(\langle\tilde{\bm g},\bm g_{\mathrm{ref},t}\rangle+\kappa_t\big). \] Stationarity gives $\tilde{\bm g}=\bm g_t+\lambda\bm g_{\mathrm{ref},t}$. Complementary slackness yields two cases. If $\langle\bm g_t,\bm g_{\mathrm{ref},t}\rangle\ge-\kappa_t$, then $\lambda=0$ and $\tilde{\bm g}=\bm g_t$. Otherwise, \[ \lambda=\frac{-\kappa_t-\langle\bm g_t,\bm g_{\mathrm{ref},t}\rangle}{\|\bm g_{\mathrm{ref},t}\|_2^2}>0, \] which substituted back gives \eqref{eq:shifted_projection_closed_form}. \end{proof} \begin{corollary}[Exact A-GEM Recovery] \label{cor:agem_recovery} Setting $\kappa_t=0$ in \eqref{eq:shifted_projection_closed_form} recovers the A-GEM correction rule \citep{agem_2019} exactly. \end{corollary} \begin{proof} With $\kappa_t=0$, the max term in \eqref{eq:shifted_projection_closed_form} becomes $\max\left(0,\frac{-\langle\bm g_t,\bm g_{\mathrm{ref},t}\rangle}{\|\bm g_{\mathrm{ref},t}\|_2^2}\right)$, which is precisely the A-GEM half-space correction factor. \end{proof} \subsection{Boundary Sensitivity and Conditional Guarantees} \begin{lemma}[Monotonicity of Correction Norm in Slack] \label{lem:monotonicity} In the active-constraint regime, the correction norm in \eqref{eq:shifted_projection_closed_form} is non-increasing with respect to $\kappa_t$. \end{lemma} \begin{proof} In the active regime, the correction norm equals \[ \left\|\tilde{\bm g}_t-\bm g_t\right\|_2=\frac{-\kappa_t-\langle\bm g_t,\bm g_{\mathrm{ref},t}\rangle}{\|\bm g_{\mathrm{ref},t}\|_2}. \] Differentiating with respect to $\kappa_t$ gives derivative $-1/\|\bm g_{\mathrm{ref},t}\|_2\le0$. \end{proof} The implication is practical: larger slack can reduce projection aggressiveness and overhead, but only within admissible forgetting regions. Thus \eqref{eq:forgetting_bound} is interpreted as a calibration constraint tied to $L_F$ estimation quality rather than a global theorem without caveats. \section{Validation Protocol and Evidence Assembly} \subsection{Protocol Lock, Comparators, and Metrics} Validation follows a CPU-only, matched-budget protocol with fixed memory envelopes and reproducibility checks inherited from benchmark infrastructure \citep{scenario_2019,avalanche_2023,mammoth_2024}. Comparator families include replay and regularization anchors: ER, EWC, LwF, A-GEM, DER++, and GDumb \citep{er_2019,ewc_2017,lwf_2016,agem_2019,derpp_2020,gdumb_2020}. Supporting analyses include constrained-replay and uncertainty-aware lines when interpreting contradiction boundaries \citep{gem_2017,mir_2019,rainbow_2021,traj_2023,amr_2024,sprint_2024}. Metrics include retention/plasticity summaries, forgetting deltas, protocol compliance, and systems costs. Uncertainty reporting is based on confidence intervals in summary tables and claim-specific caveat tracking. The current evidence package reports symbolic checks, protocol compliance diagnostics, quantitative comparison tables, and negative-result ledgers. The evidence assembly is designed to terminate claims in concrete artifacts. Formal claims terminate in theorem statements, proofs, and symbolic audit rows; benchmark claims terminate in figures, tabular effect summaries, and uncertainty intervals; systems claims terminate in protocol-compliance and instrumentation summaries. This explicit termination structure prevents accidental inflation of claims when some evidence channels are stronger than others. In the present run, formal channels are comparatively strong and empirical channels are directionally positive but still caveated by surrogate data and calibration limits. A related benefit is contradiction management. Prior literature synthesis identified recurring disagreements about replay sufficiency, projection overhead, and compression stability. In this manuscript, those disagreements are not resolved by rhetoric; they are mapped onto specific testable slices. For example, replay-versus-regularization disagreements are reflected in matched-comparator controls, projection-overhead disagreements are reflected in crossover diagnostics tied to \eqref{eq:shifted_projection_closed_form}, and compression-stability disagreements are reflected in complementary stress logs. This mapping preserves scientific continuity across the study design and allows later evidence updates to strengthen or weaken claims without rewriting the methodological core. \subsection{Claim-Evidence Mapping} Table~\ref{tab:claim_map} provides the manuscript-level claim assembly. Each major claim is tied to at least one figure or table and accompanied by caveats. This avoids unsupported narrative extrapolation and preserves a clear bridge from method to evidence. \begin{table}[t] \centering \small \renewcommand{\arraystretch}{1.1} \setlength{\tabcolsep}{4pt} \caption{Claim-evidence closure used in this manuscript. Each row states the evidence source, support level, and explicit caveat boundary. The table is designed to make support conditionality visible rather than implicit.} \label{tab:claim_map} \begin{tabular}{p{0.17\linewidth}p{0.27\linewidth}p{0.14\linewidth}p{0.34\linewidth}} \hline Claim focus & Primary evidence & Status & Caveat boundary \\ \hline Entropy-aware replay improves frontier behavior against static replay & \Figref{fig:h1h2}, Table~\ref{tab:quant_summary} & Supported & Current evidence uses protocol-locked surrogate streams; high-entropy degradation slices are logged. \\ Entropy-conditioned slack improves crossover against fixed projection in validated regions & \Figref{fig:h1h2}, Table~\ref{tab:quant_summary} & Supported (conditional) & Counterexample slices appear under abrupt drift with high slack and uncertain $L_F$. \\ Adaptive memory mixture improves efficiency in part of the regime & \Figref{fig:h3protocol}, Table~\ref{tab:quant_summary} & Mixed & Compression-collapse boundaries appear under stress and reduce global support strength. \\ \hline \end{tabular} \end{table} \section{Results} \subsection{Primary Hybrid Evidence} \begin{figure}[t] \centering \includegraphics[width=0.68\linewidth]{figures/h1_h2_tradeoff_panels.pdf} \caption{Multi-panel evidence for the two primary claims. Panel (a) reports frontier-gain behavior as entropy targets vary, comparing entropy-aware replay against static replay across benchmark slices; the horizontal axis is entropy target and the vertical axis summarizes frontier improvement under matched budgets. Panel (b) reports forgetting-overhead crossover as projection slack changes, with a guardrail reference indicating acceptable forgetting inflation relative to fixed A-GEM. Together, the panels show positive gains in validated regions while preserving visibility of boundary regimes where gains narrow or reverse.} \label{fig:h1h2} \end{figure} Quantitatively, the frontier-gain deltas between entropy-aware replay and static replay are positive across three reported dataset slices, with pairwise mean differences of approximately $0.056$, $0.060$, and $0.054$ from the summary table. The aggregate support interval reported for this claim is 95\% confidence interval $\approx [0.0533,0.0598]$. These findings are consistent with the first claim in Table~\ref{tab:claim_map} and align with the limit-based intuition from \eqref{eq:access_mixture} and \eqref{eq:entropy_def}: controlled stochasticity can improve tradeoff surfaces when protocol constraints are respected. For the projection claim, crossover improvements over fixed A-GEM are positive in all three reported slices (approximately $0.066$, $0.021$, and $0.080$ in forgetting-delta comparisons). The aggregate support interval is positive but wider (95\% confidence interval $\approx [0.028,0.194]$), so this claim remains explicitly conditional. This trend is consistent with the shifted-projection derivation in \eqref{eq:shifted_projection_closed_form} and the conditional nature of \eqref{eq:forgetting_bound}. However, support remains region-dependent: high-slack abrupt-drift slices produce failures that are explicitly retained in the negative-result ledger. A useful interpretation is that the projection extension behaves as a controlled relaxation mechanism. In moderate regimes, nonzero slack reduces unnecessary correction and can improve throughput-retention balance. In adversarial drift regimes, the same relaxation can become too permissive unless calibration terms are conservative. This behavior is expected from \eqref{eq:forgetting_bound}: admissibility is proportional to both slack and the stream-specific sensitivity term. The empirical pattern therefore matches the formal structure instead of contradicting it, but only when the caveat is respected. \subsection{Complementary Efficiency and Protocol Behavior} \begin{figure}[t] \centering \includegraphics[width=0.68\linewidth]{figures/h3_protocol_panels.pdf} \caption{Complementary evidence on memory-composition efficiency and protocol robustness. Panel (a) compares retention-per-byte trends for adaptive mixture versus exemplar replay as compression pressure increases, showing modest gains in some regimes and collapse tendencies in stress regimes. Panel (b) summarizes protocol-compliance rates under a fixed CPU contract across methods, indicating that the instrumentation and budget checks remain stable during the reported runs. The figure supports nuanced interpretation: efficiency gains are real but not uniform, and protocol fidelity is high enough to make caveated comparisons meaningful.} \label{fig:h3protocol} \end{figure} The complementary memory-mixture claim is mixed by design in this iteration. Reported deltas in retention-per-byte are positive in two slices but near zero or negative in one stress slice, and failure logs are substantial under extreme compression and abrupt drift. This pattern is scientifically useful because it sharpens boundaries instead of masking them. It suggests that representation mixing should be interpreted as a conditional optimization tool rather than a universally dominant strategy. Protocol evidence remains strong: mean compliance is approximately $0.960$ with a range near $[0.950,0.967]$ across reported checks. This reduces ambiguity about whether observed differences are instrumentation artifacts. The protocol row in Table~\ref{tab:quant_summary} is not a secondary detail; it is a precondition for interpretability. Without high compliance, apparent method differences could be explained by instrumentation drift, logging loss, or budget mismatch. By keeping compliance high across reported slices, we can interpret positive and negative trends as properties of method behavior rather than measurement instability. This distinction is essential for any subsequent revision that might tighten or rebut current claims. \begin{table}[t] \centering \small \renewcommand{\arraystretch}{1.1} \setlength{\tabcolsep}{4pt} \caption{Quantitative summary of main evidence streams. Values are drawn from exported validation tables and support a hybrid interpretation: strong formal alignment with positive but caveated empirical trends. The final row summarizes theorem-audit pass counts from symbolic checks used to validate derivation obligations.} \label{tab:quant_summary} \begin{tabular}{p{0.33\linewidth}p{0.16\linewidth}p{0.16\linewidth}p{0.24\linewidth}} \hline Metric summary & Mean effect & Support label & Evidence role \\ \hline Frontier-gain delta (entropy-aware replay minus static replay) & $+0.056$ to $+0.060$ & Supported & Tests replay-control claim with matched budget \\ Forgetting crossover improvement (fixed A-GEM minus entropy-slack projection) & $+0.021$ to $+0.080$ & Supported (conditional) & Tests projection-slack claim in valid regions \\ Retention-per-byte delta (adaptive mixture minus exemplar replay) & $-0.001$ to $+0.016$ & Mixed & Tests complementary efficiency claim and failure boundary \\ Protocol compliance rate & $0.960$ (range $0.950$--$0.967$) & Supported & Confirms measurement-contract stability \\ Symbolic theorem checks & $7/7$ pass & Supported & Validates formal obligations for limit and projection identities \\ \hline \end{tabular} \end{table} \subsection{Claim-Level Interpretation} The manuscript presents three claim classes with different confidence tiers. First, the entropy-aware replay claim is currently supported by both quantitative trends and consistency with the formal limit analysis. Second, the entropy-conditioned projection claim is supported in bounded regions and theoretically anchored, but depends on calibration assumptions and drift regime. Third, the memory-mixture claim is explicitly complementary and mixed, with substantial failure evidence that informs boundary-aware deployment. This tiered interpretation is deliberate. It avoids binary framing and instead maps each claim to support status, caveat structure, and actionable next experiments. \subsection{Uncertainty, Robustness, and Contradiction Closure} The uncertainty profile of the current evidence is informative on its own. The first primary claim shows a relatively tight positive interval in the summarized support note, whereas the second claim shows a wider interval that remains positive in aggregate but overlaps more with regime-dependent behavior. This difference is consistent with mechanism complexity: replay-frontier effects are driven mainly by access scheduling, while projection effects depend on interaction among entropy context, slack magnitude, and gradient geometry. The manuscript therefore treats the second claim as conditionally supported even though the aggregate direction is favorable. Robustness is further interpreted through explicit contradiction closure rather than through average scores alone. The contradiction map from prior literature synthesis highlighted replay-versus-regularization sufficiency and constrained-projection overhead as unresolved debates. In this manuscript, those debates are narrowed by evidence channels that directly target them: matched comparator controls for replay sufficiency, crossover diagnostics for projection overhead, and failure-ledger accounting for stress regimes. This does not eliminate all uncertainty, but it changes uncertainty from a vague concern into a structured object that can be tested in follow-up experiments. A practical advantage of this structure is that revisions can be local and auditable. If future real-data sweeps weaken a currently supported claim, the corresponding row in the claim-evidence map can be downgraded without invalidating unrelated derivations. If future calibration strengthens a conditional claim, the caveat language can be narrowed with direct reference to updated bounds and protocol checks. In other words, the manuscript is written to be revision-friendly under scientific pressure: claims are modular, dependencies are explicit, and evidence provenance is preserved. This robustness posture is especially important for open-question work where methods evolve faster than benchmark standards. By preserving both support and contradiction signals in the main narrative, we reduce the chance that downstream readers overfit to headline outcomes and miss the operating conditions that actually govern deployment success. \section{Discussion} \subsection{What the Hybrid Evidence Changes} The combined formal and empirical evidence supports a systems interpretation of continual learning: memory access should be treated as a programmable control surface rather than a fixed implementation detail. The deterministic-limit theorem and shifted-projection derivation contribute formal structure that ties algorithmic behavior to measurable operating variables. The protocol-locked evidence then constrains interpretation by showing where improvements occur and where they fail. For practitioners, the key implication is design modularity. An entropy-conditioned controller can be deployed incrementally on top of established replay and projection baselines, provided matched-budget auditing is preserved. For researchers, the implication is methodological: reporting should include claim-evidence closure tables and negative-result handling, especially when system efficiency claims are central. An additional implication concerns evaluation culture. Continual-learning papers often combine many moving parts, making it tempting to optimize for aggregate leaderboards while leaving assumption provenance implicit. The present manuscript argues for the opposite strategy in systems-sensitive settings: make assumptions explicit, define novelty boundaries narrowly, and preserve contradictory evidence in the main narrative rather than relegating it to untracked notes. This does not reduce ambition; it increases transferability of results because downstream users can see where mechanisms are expected to hold. \subsection{Negative Results and Unsupported Edges} Negative evidence is substantial for stress regimes and should shape conclusions. Counterexample slices appear for high-slack abrupt-drift settings in projection experiments, and high-compression mixture experiments produce frequent collapse events. These outcomes do not invalidate the primary formal contributions, but they constrain where empirical claims are justified. They also explain why the manuscript distinguishes supported, conditionally supported, and mixed claims. The most important unsupported edge remains full real-data closure under the exact same protocol contract. Current evidence is sufficient for hybrid inference but not for universal empirical ranking statements. \subsection{Deployment-Oriented Interpretation Across Regimes} To make the results actionable, it is useful to interpret operating regimes rather than only aggregate effects. In low-entropy regimes, deterministic concentration dominates and the model behaves close to stable retrieval with limited replay traffic. In medium-entropy regimes, mixed access can improve adaptation without sharply increasing projection burden, and this is where most positive effects in the current evidence appear. In high-entropy regimes, both replay traffic and projection sensitivity can rise; here the controller must trade off adaptation against explicit latency and energy constraints, and boundary failures become more likely if slack or compression is too aggressive. This regime view suggests practical control policies. A conservative deployment can begin with low-to-moderate entropy targets, monitor compliance and forgetting drift, and only escalate entropy or slack when guardrail metrics remain stable. Conversely, if compliance drops or forgetting spikes under drift, the controller can automatically retreat toward lower entropy and lower slack, effectively moving behavior closer to the A-GEM recovery regime. The important point is not the exact heuristic threshold but the principle that control adaptation should be tied to measurable guardrails and not only to performance objectives. The same logic applies to representation strategy. Adaptive mixture can provide efficiency gains when drift and compression remain within calibrated bounds, but aggressive compression under abrupt shift can negate those gains. Therefore representation adaptation should be coupled to failure predictors and bounded by explicit rollback rules. In systems terms, this is a reliability-first interpretation: local gains are valuable, but only when accompanied by mechanisms that prevent silent degradation outside validated regions. \section{Limitations and Future Work} \subsection{Current Limitations} Three limitations are material. First, the current benchmark run relies on deterministic surrogate streams for core tables and figures; this preserves protocol control but limits external validity for real datasets. Second, the forgetting-sensitivity constant $L_F$ is treated through symbolic checks and bounded reasoning, not yet through fully calibrated empirical intervals across all stream families. Third, assumption naming consistency between symbolic and experimental layers can be improved to reduce audit friction. These limitations directly affect interpretation: conclusions about formal identities are strong, conclusions about trend direction are moderate, and conclusions about universal deployment superiority are intentionally withheld. The limitations also indicate concrete failure pathways that should be tested before stronger claims are made. Surrogate-to-real transfer may fail if real dataset entropy dynamics differ from synthetic schedules. Calibration may fail if $L_F$ varies sharply across stream families or time horizons. Assumption-tag mismatch may fail audit reproducibility when future contributors cannot align theorem premises with instrumentation checks. Each of these pathways has a direct remediation experiment, which is why we list them as first-order follow-ups rather than generic future-work placeholders. \subsection{Future Work} Immediate follow-up experiments are clear: execute full real-data sweeps under the same matched-budget protocol lock, calibrate $L_F$ empirically with stream-wise confidence intervals, and rerun theorem-boundary audits with measured rather than symbolic-only calibration terms. A second direction is regime-adaptive slack governance that can automatically retreat toward zero slack in detected drift spikes, preserving A-GEM-like behavior when boundary risk rises. A third direction is extending memory-mixture controls with explicit collapse predictors so representation compression can be scheduled safely rather than statically. \section{Conclusion} This paper develops and evaluates an entropy-aware memory-systems perspective for continual learning under practical CPU constraints. The formal side shows that deterministic retrieval is a limit case of stochastic replay and that entropy-conditioned shifted projection has a closed form with exact A-GEM recovery. The evidence side provides protocol-locked support for primary claims, explicit caveats for conditional regions, and transparent reporting of negative results for complementary claims. The central takeaway is not that one mechanism universally wins, but that claim-evidence closure and boundary-aware control produce more reliable scientific and engineering conclusions for continual-learning systems. \clearpage\phantomsection\label{sec:end_of_main} \bibliographystyle{conference} \bibliography{references} \appendix \clearpage\phantomsection\label{sec:appendix_start} \section{Extended Proofs and Derivation Details} \subsection{Extended Derivation for Entropy Decomposition} We expand \eqref{eq:entropy_def} directly to make provenance explicit. Using \eqref{eq:access_mixture}, for $i\neq i_t^*$, \[ p_t(i)=(1-\alpha_t)q_t(i),\qquad p_t(i_t^*)=\alpha_t. \] Therefore, \[ H_t=-\alpha_t\log\alpha_t-\sum_{i\neq i_t^*}(1-\alpha_t)q_t(i)\log\left((1-\alpha_t)q_t(i)\right). \] Splitting logarithms yields \[ H_t=-\alpha_t\log\alpha_t-(1-\alpha_t)\log(1-\alpha_t)\sum_{i\neq i_t^*}q_t(i)-(1-\alpha_t)\sum_{i\neq i_t^*}q_t(i)\log q_t(i), \] and because $\sum_{i\neq i_t^*}q_t(i)=1$, this becomes \[ H_t=h_2(\alpha_t)+(1-\alpha_t)H(q_t). \] This is the identity checked in symbolic audits and used in \Secref{sec:formal_references_note} for reproducibility traceability. \subsection{Extended KKT Derivation for Shifted Projection} For completeness, we restate the constrained problem in \eqref{eq:shifted_projection_problem} with primal variable $\tilde{\bm g}$ and dual variable $\lambda\ge 0$. Stationarity gives \[ \tilde{\bm g}=\bm g_t+\lambda\bm g_{\mathrm{ref},t}. \] Feasibility requires \[ \langle\bm g_t+\lambda\bm g_{\mathrm{ref},t},\bm g_{\mathrm{ref},t}\rangle+\kappa_t\ge 0. \] Complementary slackness requires \[ \lambda\left(\langle\bm g_t+\lambda\bm g_{\mathrm{ref},t},\bm g_{\mathrm{ref},t}\rangle+\kappa_t\right)=0. \] Inactive branch gives $\lambda=0$. Active branch gives \[ \lambda=\frac{-\kappa_t-\langle\bm g_t,\bm g_{\mathrm{ref},t}\rangle}{\|\bm g_{\mathrm{ref},t}\|_2^2}, \] which is valid when numerator is positive. Substituting into stationarity yields \eqref{eq:shifted_projection_closed_form} exactly. \subsection{Boundary Conditions and Caveat Mapping} Three boundary cases are central for interpretation: (i) overlap-correction conditions where $q_t(i_t^*)>0$ invalidate exact decomposition unless corrected, (ii) degenerate projection conditions where $\bm g_{\mathrm{ref},t}=\bm 0$, and (iii) over-large slack under abrupt drift. The symbolic audit marks these as surfaced caveats, not silent failures. This mapping is essential for keeping formal claims and empirical caveats synchronized. \subsection{Assumption-Label Bridge for Reproducibility} To reduce audit ambiguity between the formal and empirical layers, Table~\ref{tab:assumption_bridge} maps each assumption label from Section~3 to the theorem obligations and validation evidence used in this manuscript. \begin{table}[h] \centering \footnotesize \renewcommand{\arraystretch}{1.1} \setlength{\tabcolsep}{4pt} \caption{Assumption-label bridge used for reproducibility and caveat tracking. Each assumption is linked to the exact formal or empirical evidence channel that checks it.} \label{tab:assumption_bridge} \begin{tabular}{p{0.10\linewidth}p{0.28\linewidth}p{0.50\linewidth}} \hline Label & Statement focus & Evidence channel in this manuscript \\ \hline A1 & Exact access decomposition ($q_t(i_t^*)=0$) & Theorem~\ref{thm:det_limit} plus boundary-case caveat mapping in this appendix. \\ A2 & Deterministic tie-breaking for retrieval index & Theorem~\ref{thm:det_limit} and deterministic-limit derivation details in Appendix A.1. \\ A3 & Measurable latency/energy protocol & Protocol-compliance summaries in Table~\ref{tab:quant_summary} and Table~\ref{tab:audit_summary}. \\ A4 & Matched memory budget across comparators & Matched-budget validation contract in Section~5 and protocol-compliance summaries. \\ B1 & Nonzero reference gradient in active projection & Lemma~\ref{lem:shifted_proj} conditions and boundary tagging in Appendix A.3. \\ B2 & Nonnegative monotone slack function $\kappa(H_t)$ & Lemma~\ref{lem:monotonicity} and shifted-projection derivation in Section~4.2--4.3. \\ B3 & Stream-dependent forgetting-sensitivity bound ($L_F$) & Conditional interpretation around \eqref{eq:forgetting_bound} and caveat text in Sections~5--7. \\ B4 & Comparator protocol parity for projection claims & CPU-only matched-budget protocol and compliance statistics reported in Section~5 and Table~\ref{tab:quant_summary}. \\ \hline \end{tabular} \end{table} \section{Notation and Symbol Glossary} \begin{table}[h] \centering \footnotesize \renewcommand{\arraystretch}{1.1} \setlength{\tabcolsep}{4pt} \caption{Core notation used across methods, proofs, and evidence interpretation. Symbols are defined near first use in the main text and collected here for compact reference.} \label{tab:symbols} \begin{tabular}{p{0.20\linewidth}p{0.72\linewidth}} \hline Symbol & Meaning \\ \hline $\train_t$ & Task-specific stream partition observed at step $t$ \\ $B_t, M$ & Episodic memory and hard memory budget \\ $\alpha_t$ & Deterministic concentration factor in access mixture \\ $q_t$ & Stochastic replay distribution over non-selected memory indices \\ $p_t$ & Combined access distribution defined in \eqref{eq:access_mixture} \\ $H_t$ & Access entropy from \eqref{eq:entropy_def} \\ $R_t, P_t$ & Retention and plasticity summary metrics \\ $\mathrm{Lat}_t,\mathrm{En}_t$ & Per-step latency and energy summaries under fixed protocol \\ $\bm g_t, \bm g_{\mathrm{ref},t}$ & Current gradient and replay/reference gradient \\ $\kappa_t$ & Entropy-conditioned slack term for projection \\ $L_F$ & Stream-specific forgetting sensitivity constant \\ \hline \end{tabular} \end{table} \section{Reproducibility and Implementation Details} \subsection{Execution Envelope} The reported validation run uses CPU-only execution with matched-memory controls and protocol-compliance auditing. Seeds, comparator lists, and sweep structures are fixed by configuration artifacts and reused across claim evaluations to avoid selective reporting. The generated outputs include two multi-panel figures, five summary tables, dataset-level run logs, symbolic check logs, and negative-result ledgers. \subsection{Uncertainty and Aggregation} Uncertainty is summarized through confidence intervals in summary tables and claim-level support notes. For example, frontier-gain and crossover effects are reported with interval bounds rather than only point estimates. This supports calibrated interpretation: a claim is treated as supported only when both trend direction and uncertainty behavior are aligned with method assumptions. \subsection{Theorem-Check Reproducibility} Symbolic obligations are reproduced through an explicit check suite that verifies entropy decomposition, deterministic limit behavior, shifted-projection reduction to A-GEM, correction monotonicity, and boundary-case tagging. All seven theorem-check rows in the audit table are recorded as passing in the current run. These checks do not replace empirical validation, but they prevent algebraic drift between derivation and implementation. \section{Extended Evidence and Failure Analysis} \subsection{Protocol and Theorem Audit Summaries} \begin{table}[h] \centering \footnotesize \renewcommand{\arraystretch}{1.1} \setlength{\tabcolsep}{4pt} \caption{Supplementary audit summary for protocol fidelity and theorem checks. The first row reports protocol compliance behavior across method-dataset slices, while the second row reports symbolic theorem obligations. These quantities contextualize the reliability of claim interpretation by separating measurement-contract quality from method effect size.} \label{tab:audit_summary} \begin{tabular}{p{0.32\linewidth}p{0.58\linewidth}} \hline Audit dimension & Summary \\ \hline Protocol compliance & Mean compliance approximately $0.960$ across 27 reported method-dataset slices, with observed range near $[0.950,0.967]$. \\ Theorem obligations & Seven of seven symbolic theorem checks reported as pass, including entropy-limit, A-GEM recovery, monotonicity, and boundary tagging obligations. \\ Negative-result burden & Zero logged failures for the first primary claim, seven for the second under abrupt drift/high slack, and seventy-three for complementary mixture stress slices. \\ \hline \end{tabular} \end{table} \subsection{Interpretive Guidance for Practitioners} For deployment-oriented users, the practical recommendation is to treat entropy-aware replay and entropy-conditioned projection as controllable modules with runtime guards, not as static replacements for existing baselines. In moderate entropy regimes with stable calibration, the reported trends justify adoption trials. In abrupt-drift or high-compression regimes, fail-safe behavior should default toward conservative slack and lower compression until real-data calibration confirms safe operation. \subsection{Limits of Generalization} Because the current manuscript reports hybrid evidence with surrogate-stream empirical closure, strong claims should be interpreted as regime-bounded. The formal results generalize under their assumptions; empirical gains should be treated as validated trends requiring real-data replication for broader external validity. \section{Equation Provenance Note} \label{sec:formal_references_note} The optimization geometry in \eqref{eq:shifted_projection_problem} and the recovery property in \eqref{eq:shifted_projection_closed_form} inherit constrained-gradient lineage from GEM and A-GEM \citep{gem_2017,agem_2019}. The regularization and distillation terms that motivate composite retention-plasticity objectives connect to EWC and LwF \citep{ewc_2017,lwf_2016}. The entropy-parameterized access process and systems-aware control objective are introduced in this manuscript as the primary formal extensions. \end{document}