[ { "assumptions": [ "Output-layer-only training with fixed reservoir dynamics." ], "citation": "S01 (2025)", "claims": [ "Entanglement-correlated regimes improve embedding quality without proving asymptotic quantum advantage." ], "conclusions": [ "Representational gains can coexist with classically simulable operating points." ], "contributions": [ "Quantifies accuracy transition versus evolution time on image tasks." ], "future_work": [ "Test harder datasets and deeper/nonintegrable dynamics." ], "key_equations": [ "H=(1/2)\u2211i(\u03c3x(i)\u03c3x(i+1)+\u03c3y(i)\u03c3y(i+1))" ], "limitations": [ "Moderate system sizes and benchmark-scale datasets." ], "source_type": "paper", "summary": "QELM image-study linking an entanglement onset regime to improved separability while noting broad classical simulability.", "title": "Entanglement and Classical Simulability in Quantum Extreme Learning Machines", "url": "https://arxiv.org/abs/2509.06873", "year": 2025 }, { "assumptions": [ "PCA/AE compression preserves task signal." ], "citation": "S02 (2025)", "claims": [ "Encoding choice strongly affects final accuracy." ], "conclusions": [ "Preprocessing, encoding, and measurement must be co-designed." ], "contributions": [ "Compares encoding and Hamiltonian families under matched readouts." ], "future_work": [ "Hardware execution with calibrated measurement constraints." ], "key_equations": [ "|\u03c8\u27e9=cos(\u03b8/2)|0\u27e9+ei\u03c6sin(\u03b8/2)|1\u27e9" ], "limitations": [ "Primarily simulation-oriented assessment." ], "source_type": "paper", "summary": "End-to-end QELM image pipeline with encoding and Hamiltonian ablations.", "title": "Harnessing Quantum Extreme Learning Machines for image classification", "url": "https://arxiv.org/abs/2409.00998", "year": 2025 }, { "assumptions": [ "Fixed reservoir map, optimized readout dominates gains." ], "citation": "S03 (2026)", "claims": [ "Kernel-space optimization improves efficiency and accuracy versus naive readout search." ], "conclusions": [ "Measurement design is a primary practical lever." ], "contributions": [ "Derives dual/kernel readout optimization for QELM and recurrent QRC." ], "future_work": [ "Add hardware-aware regularization and non-MSE losses." ], "key_equations": [ "K(x,x\u2032)=tr(\u03c1(x)\u03c1(x\u2032))" ], "limitations": [ "Hardware-observable constraints remain." ], "source_type": "paper", "summary": "Casts QRC observable optimization as kernel regression for fixed reservoirs.", "title": "Kernel-based optimization of measurement operators for quantum reservoir computers", "url": "https://arxiv.org/abs/2602.14677", "year": 2026 }, { "assumptions": [ "Data encoding implemented via exp(-ixH)-style operators." ], "citation": "S04 (2021)", "claims": [ "Encoding is a first-order determinant of model expressivity." ], "conclusions": [ "Feature-map design is central to QML performance." ], "contributions": [ "Formal expressive-power characterization through encoding-induced frequencies." ], "future_work": [ "Finite-sample/noisy-device expressivity studies." ], "key_equations": [ "f\u03b8(x)=\u2211\u03c9\u2208\u03a9 c\u03c9(\u03b8)e^{i\u03c9x}" ], "limitations": [ "Theory-heavy; limited hardware-noise treatment." ], "source_type": "paper", "summary": "Fourier-spectrum analysis showing encoding operators determine reachable function classes.", "title": "The effect of data encoding on the expressive power of variational quantum machine learning models", "url": "https://arxiv.org/abs/2008.08605", "year": 2021 }, { "assumptions": [ "Quantum models are state-preparation plus measurement pipelines." ], "citation": "S05 (2021)", "claims": [ "Data encoding largely determines inductive bias via kernel choice." ], "conclusions": [ "Kernel view gives clearer optimization and baseline design." ], "contributions": [ "Operator-space/kernel interpretation for supervised QML." ], "future_work": [ "Scalable kernel-estimation and complexity-separation studies." ], "key_equations": [ "\u03ba(x,x\u2032)=tr(\u03c1(x)\u03c1(x\u2032))" ], "limitations": [ "Kernel computation can be costly at scale." ], "source_type": "paper", "summary": "Unifies supervised QML with kernel methods and optimal measurement expansions.", "title": "Supervised quantum machine learning models are kernel methods", "url": "https://arxiv.org/abs/2101.11020", "year": 2021 }, { "assumptions": [ "Finite-shot statistics govern practical performance." ], "citation": "S06 (2024)", "claims": [ "Competitive performance is feasible but shot-limited." ], "conclusions": [ "Measurement budget is a core deployment bottleneck." ], "contributions": [ "Characterizes shot-budget and runtime-performance tradeoffs." ], "future_work": [ "Larger arrays and stronger multiclass protocols." ], "key_equations": [ "y=Wout\u00b7z where z are measured reservoir features" ], "limitations": [ "Benchmark scale and hardware drift constraints." ], "source_type": "paper", "summary": "Neutral-atom QRC workflow with finite-shot tradeoff analysis on practical tasks.", "title": "Practical Quantum Reservoir Computing in Rydberg Atom Arrays", "url": "https://arxiv.org/abs/2407.02553", "year": 2024 }, { "assumptions": [ "Feature selection and reduced-dimensional encodings retain chemical signal." ], "citation": "S07 (2024)", "claims": [ "QRC features degrade more slowly as training size decreases." ], "conclusions": [ "Reservoir embeddings can help small/noisy scientific datasets." ], "contributions": [ "Integrates SHAP feature filtering with QRC embedding and downstream regressors." ], "future_work": [ "Broader molecular benchmarks and hardware trials." ], "key_equations": [ "H(t)=\u03a9/2\u2211j(|gj\u27e9\u27e8rj|+|rj\u27e9\u27e8gj|)+\u2211j400-mode GBS and correlation-feature gains.", "title": "Large-scale quantum reservoir computing using a Gaussian Boson Sampler", "url": "https://arxiv.org/abs/2505.13695", "year": 2025 }, { "assumptions": [ "Finite-shot statistics can induce concentration bottlenecks." ], "citation": "S22 (2025)", "claims": [ "Hamiltonian symmetries can suppress concentration effects." ], "conclusions": [ "Sample complexity is a central scaling bottleneck." ], "contributions": [ "Characterizes concentration risk and symmetry mitigation strategy." ], "future_work": [ "Platform-specific concentration diagnostics and mitigation." ], "key_equations": [ "Var[feature]\u21920 exponentially under concentration regimes" ], "limitations": [ "Model assumptions may not cover all hardware regimes." ], "source_type": "paper", "summary": "Analyzes concentration effects and symmetry-based mitigation in QRC sample complexity.", "title": "Exponential concentration and symmetries in Quantum Reservoir Computing", "url": "https://arxiv.org/abs/2505.10062", "year": 2025 }, { "assumptions": [ "Chaos diagnostics transfer to QRC design choices." ], "citation": "S23 (2025)", "claims": [ "Two characteristic edges correlate with improved QRC performance." ], "conclusions": [ "Chaos diagnostics can guide operating-point selection." ], "contributions": [ "Introduces quantum edge-of-chaos operating-point analysis." ], "future_work": [ "Validate criteria on experimentally accessible reservoirs." ], "key_equations": [ "performance peak near Thouless-time related boundary" ], "limitations": [ "SYK-centered analysis scope." ], "source_type": "paper", "summary": "Identifies QRC performance peaks near many-body chaos boundaries.", "title": "Edge of Many-Body Quantum Chaos in Quantum Reservoir Computing", "url": "https://arxiv.org/abs/2506.17547", "year": 2025 }, { "assumptions": [ "PTM is suitable for open-system memory diagnostics." ], "citation": "S24 (2024)", "claims": [ "Coherence influx is necessary for non-stationary ESP in studied settings." ], "conclusions": [ "Open-system coherence flow is a key QRC design variable." ], "contributions": [ "Derives non-stationary ESP conditions tied to coherence influx." ], "future_work": [ "Empirical PTM diagnostics on noisy hardware traces." ], "key_equations": [ "\u03c1(PTM) characterizes fading memory" ], "limitations": [ "Theory-heavy with limited hardware calibration detail." ], "source_type": "paper", "summary": "PTM-based analysis showing coherence influx and spectral radius govern non-stationary ESP behavior.", "title": "Coherence influx is indispensable for quantum reservoir computing", "url": "https://arxiv.org/abs/2409.12693", "year": 2024 }, { "citation": "S25 (2018)", "source_type": "article", "summary": "Baseline QML framework used to contextualize output-layer-only QRC/QELM alternatives.", "title": "Quantum circuit learning", "url": "https://doi.org/10.1103/PhysRevA.98.032309", "year": 2018 }, { "citation": "S26 (2019)", "source_type": "article", "summary": "Feature-space perspective used for kernel-baseline framing.", "title": "Quantum machine learning in feature Hilbert spaces", "url": "https://doi.org/10.1103/PhysRevLett.122.040504", "year": 2019 }, { "citation": "S27 (2019)", "source_type": "article", "summary": "Quantum-kernel benchmark reference for matched baseline comparisons.", "title": "Supervised learning with quantum-enhanced feature spaces", "url": "https://doi.org/10.1038/s41586-019-0980-2", "year": 2019 }, { "citation": "S28 (2021)", "source_type": "report", "summary": "Survey context for trainability and optimization tradeoffs in quantum ML.", "title": "Variational quantum algorithms", "url": "https://doi.org/10.1038/s42254-021-00348-9", "year": 2021 }, { "citation": "S29 (2018)", "source_type": "article", "summary": "Trainability limitation reference motivating fixed-reservoir output-only learning.", "title": "Barren plateaus in quantum neural network training landscapes", "url": "https://doi.org/10.1038/s41467-018-07090-4", "year": 2018 }, { "citation": "S30 (2001)", "source_type": "report", "summary": "Foundational echo-state framework used in RC/QRC methodology.", "title": "The echo state approach to analysing and training recurrent neural networks", "url": "https://www.ai.rug.nl/minds/uploads/EchoStatesTechRep.pdf", "year": 2001 }, { "citation": "S31 (2002)", "source_type": "article", "summary": "Liquid-state computing foundation for reservoir-style computation.", "title": "Real-time computing without stable states", "url": "https://doi.org/10.1162/089976602760407955", "year": 2002 }, { "citation": "S32 (2009)", "source_type": "report", "summary": "Classical RC survey for baseline construction and evaluation norms.", "title": "Reservoir computing approaches to recurrent neural network training", "url": "https://doi.org/10.1016/j.cosrev.2009.03.005", "year": 2009 }, { "citation": "S33 (1998)", "source_type": "dataset", "summary": "Canonical image benchmark reference used in many QRC/QELM studies.", "title": "MNIST dataset reference", "url": "https://doi.org/10.1109/5.726791", "year": 1998 }, { "citation": "S34 (2017)", "source_type": "dataset", "summary": "Harder drop-in MNIST replacement for benchmark robustness checks.", "title": "Fashion-MNIST", "url": "https://arxiv.org/abs/1708.07747", "year": 2017 }, { "citation": "S35 (2009)", "source_type": "dataset", "summary": "Higher-complexity image benchmark for stress-testing representations.", "title": "CIFAR-10", "url": "https://www.cs.toronto.edu/~kriz/cifar.html", "year": 2009 } ]