Papers
Topics
Authors
Recent
2000 character limit reached

Quantum Boltzmann Machines

Updated 24 December 2025
  • Quantum Boltzmann Machines are energy-based models that generalize classical Boltzmann machines by using quantum Hamiltonians to encode both classical and quantum correlations.
  • They employ diverse architectures and sampling techniques—including quantum annealing and hybrid quantum-classical methods—to manage the challenges posed by non-commuting Hamiltonians.
  • QBMs find practical use in generative modeling, state tomography, reinforcement learning, and financial or medical data analysis, despite hardware and scalability constraints.

Quantum Boltzmann Machines (QBMs) are energy-based models that generalize classical Boltzmann machines by encoding statistical dependencies via quantum Hamiltonians. This framework allows both classical and quantum correlations to be represented, with applications extending from generative modeling to quantum state tomography and quantum-enhanced supervised learning. QBMs take advantage of quantum statistical mechanics, using quantum Gibbs states governed by non-commuting Hamiltonians, and employ a variety of sampling, training, and inference schemes—often hybridizing quantum resources with classical optimization and data processing.

1. Mathematical Structure and Model Classes

A Quantum Boltzmann Machine is formally defined in terms of a parametric Hamiltonian H(θ)H(\theta) acting on a Hilbert space of nn qubits (partitioned into visible and possibly hidden units), with the quantum Gibbs state

ρ(θ)=eH(θ)Z(θ),Z(θ)=Tr[eH(θ)].\rho(\theta) = \frac{e^{-H(\theta)}}{Z(\theta)}, \quad Z(\theta) = \operatorname{Tr}[e^{-H(\theta)}].

Marginal distributions over visible units are extracted as

p(v;θ)=Tr[Λvρ(θ)],p(v;\theta) = \operatorname{Tr}[\Lambda_v \rho(\theta)],

where Λv=vvIhidden\Lambda_v = |v\rangle \langle v| \otimes I_{\mathrm{hidden}} projects onto a classical configuration of the visibles.

Model topologies include:

  • Fully-Visible QBMs: No hidden units; parameterized by arbitrary Hermitian operators on the visible subspace. Convex optimization in parameter space is possible and trainability is empirically favorable, avoiding "barren plateaus" (Tüysüz et al., 21 Oct 2024, Coopmans et al., 2023).
  • Restricted/Deep QBMs: Bipartite or multi-layer architectures with hidden qubits, typically with transverse-field or more general non-commuting interactions (Amin et al., 2016, Demidik et al., 24 Feb 2025). Hidden–hidden and visible–hidden couplings can be quantum or classical (commuting).
  • Semi-Quantum Restricted Boltzmann Machines (sqRBM): Hard restriction to commuting visible–visible terms with non-commuting hidden-sector Hamiltonians, permitting closed-form gradients and a favorable scaling in hidden-unit count compared to classical RBMs (Demidik et al., 24 Feb 2025, Kimura et al., 29 Jul 2025).
  • Continuous-Variable QBMs: Extension to bosonic modes, where the thermal state is defined over a non-commuting Hamiltonian on continuous variables and learning is performed via variational quantum imaginary-time evolution (Bangar et al., 10 May 2024).
  • Evolved QBMs: The state is obtained by applying parameterized real-time unitary evolution to a quantum Gibbs state, thereby augmenting expressivity (Minervini et al., 6 Jan 2025, Wilde, 2 Dec 2025).

Typical quantum Hamiltonians include:

H=a<bwabσazσbzabaσazaΓaσax+Hqm,H = -\sum_{a<b} w_{ab} \sigma_a^z \sigma_b^z - \sum_a b_a \sigma_a^z - \sum_a \Gamma_a \sigma_a^x + H_{\mathrm{qm}},

where HqmH_{\mathrm{qm}} contains additional (possibly non-stoquastic) quantum terms, e.g., XX-YY couplings, or more general interactions.

2. Quantum Sampling and Annealing Techniques

Sampling from the quantum Gibbs distribution, required for negative-phase gradient estimates, is computationally intractable in the general, non-commuting case (QMA-hard). Several hardware and algorithmic solutions are leveraged:

  • Quantum Annealing (QA): Physical sampling is performed by evolving a transverse-field Ising Hamiltonian on devices such as D-Wave. Annealing schedule H(t)=A(t)HD+B(t)HPH(t) = A(t) H_D + B(t) H_P with HD=iσixH_D = -\sum_{i} \sigma_{i}^x and HPH_P encoding the Ising energy. Devices sample approximately from eβHPe^{-\beta H_P} in a restricted regime (Perot, 2023, Schuman et al., 18 Jul 2025).
  • Simulated Quantum Annealing (SQA): Suzuki–Trotter mapping is used to reduce the quantum partition function to a classical one in higher dimensions, enabling efficient Monte Carlo sampling for transverse-field models (Müller et al., 2021, Crawford et al., 2016).
  • Parallel Quantum Annealing (PQA): Partition hardware into parallel subgraphs to extract multiple independent samples in a single anneal, yielding empirical 69.65% QPU runtime reduction for medical image classification (Schuman et al., 18 Jul 2025).
  • Eigenstate Thermalization Hypothesis (ETH): For k-local Hamiltonians, expectation values of local observables following a quantum quench converge rapidly to their thermal values, allowing efficient "thermalization-by-quench" as a heuristic sampling protocol (Anschuetz et al., 2019).
  • Quantum Metropolis, Gibbs-state Preparation: For fully-visible or small models, quantum algorithms such as Szegedy walks, fluctuation-theorem approaches, and classical shadow tomography can be employed (Tüysüz et al., 21 Oct 2024).
  • Classical–Quantum Hybrid Schemes: QA initialization combined with classical Markov chains accelerates convergence in classical RBMs and can be generalized to QBM training (Kālis et al., 2023).

3. Training Algorithms for Quantum Boltzmann Machines

QBM training is formulated as minimizing a discrepancy (typically quantum relative entropy) between the model and data distributions:

L(θ)=vPvdatalogp(v;θ)\mathcal{L}(\theta) = -\sum_{v} P_v^{\mathrm{data}} \log p(v; \theta)

or, for quantum state targets,

S(ρσ(θ))=Tr[ρ(logρlogσ(θ))]S(\rho||\sigma(\theta)) = \operatorname{Tr}[ \rho (\log \rho - \log \sigma(\theta))]

Key algorithmic strategies:

  • Gradient-based methods: For general non-commuting Hamiltonians, the gradient is formally

θL=vPvdata[Tr[ΛvθeβH]Tr[ΛveβH]Tr[θeβH]Z]\nabla_\theta \mathcal{L} = \sum_v P^{\mathrm{data}}_v \left[ \frac{\operatorname{Tr}[ \Lambda_v \partial_\theta e^{-\beta H} ]}{\operatorname{Tr}[ \Lambda_v e^{-\beta H} ]} - \frac{ \operatorname{Tr}[ \partial_\theta e^{-\beta H} ] }{ Z } \right ]

The negative-phase term simplifies to an expectation value, but the positive phase remains generally intractable. Golden–Thompson bounds or commutator expansions are employed to produce trainable surrogates (Amin et al., 2016, Kieferova et al., 2016).

  • EM Information-Geometric Optimization: The semi-quantum RBM supports an EM-style quantum "em" algorithm, alternating ee-projections (E-steps: updating the hidden distribution for fixed visibles) with mm-projections (M-steps: parameter updates minimizing quantum relative entropy). Each step is analytically tractable and guarantees monotonic loss descent for qRBMs with commuting visible terms (Kimura et al., 29 Jul 2025).
  • Hybrid Newton Methods: When gradients and (approximate) Hessians can be estimated, Newton–Raphson updates accelerate convergence, especially when coupled with batch QA sampling and regularization strategies (Srivastava et al., 2020).
  • Variational Quantum Eigensolvers (VQE) and Low-Rank β-VQE: For general QBM expectation estimation, variational or low-rank sampling methods and nested optimization are utilized, exploiting parameter-shift rules, warm-starts, and classical models for purification (Huijgen et al., 2023).
  • Natural Gradient and Fisher Information Matrices: For evolved QBMs, the Fisher–Bures, Wigner–Yanase, and Kubo–Mori information matrices are computed to enable geometry-aware gradient descent; their matrix ordering properties ensure interchangeability up to constant factors (Minervini et al., 6 Jan 2025, Wilde, 2 Dec 2025).

Empirically, pre-training simpler submodels (mean-field, Gaussian Fermionic, or geometric-local Hamiltonians) can reduce convergence time and final relative entropy loss (Coopmans et al., 2023).

4. Expressive Power, Tractability, and Comparison to Classical Models

The expressiveness and trainability of various QBM classes are summarized as follows:

Model Type Expressivity Trainability (Gradient) Resource Demand
Fully-Visible QBM Non-commuting Pauli terms: strict generalization of classical BM Polynomial; convex loss (Tüysüz et al., 21 Oct 2024) Only visible qubits; no hidden layer
Restricted QBM / QRBM Hidden units, unrestricted Hamiltonians Intractable; commutator expansions needed—can be QMA-hard (Demidik et al., 24 Feb 2025) Qubits scale with hidden layer size
Semi-Quantum RBM (sqRBM) Same as classical RBM given scaling mc=Wmm_c = |W|\,m (factor 3\sim3 savings); injects quantum hidden-layer correlations Analytical gradients, convex updates (Demidik et al., 24 Feb 2025) Fewer hidden qubits for same expressivity
Continuous-Variable QBM Continuous observables, non-commuting Gaussian dynamics Trained via variational QITE; cost linear in cutoff Experimental: squeezed photons, PNR detectors (Bangar et al., 10 May 2024)

Significant results include:

  • Low-dimensional, fully-visible QBMs capture distributions with correlations of nominally much higher dimensionality, e.g., $1$D chains learning $2$D statistics (Tüysüz et al., 21 Oct 2024).
  • sqRBMs match classical RBMs in total parameter count but require three times fewer hidden units for equivalent expressivity in the typical operator pool W={X,Y,Z}W = \{X, Y, Z\} (Demidik et al., 24 Feb 2025).
  • Deep and evolved QBMs (interleaving imaginary and real-time evolution) can, in principle, represent any quantum computation output state via a 2-local Hamiltonian, ensuring universality for quantum generative modeling tasks (Wu et al., 2020, Minervini et al., 6 Jan 2025).

5. Applications and Empirical Performance

QBMs have been applied in diverse domains:

  • Quantum-Enhanced Reinforcement Learning: Multi-agent Q-learning architectures leveraging DBM-style QBMs as Q-value approximators demonstrate reduced time-steps to convergence and increased policy fidelity compared to classical deep RL or RBM-based methods (Müller et al., 2021, Crawford et al., 2016).
  • Medical Image Classification: Supervised learning with QBMs and parallel quantum annealing achieves 69.65% QPU time reduction, matching CNNs' accuracy (within 3%\sim 3\%) but requiring far fewer epochs. QBM convergence is robust for both large and small datasets; scalability is maintained by encoding large input dimensions as biases rather than logical qubits (Schuman et al., 18 Jul 2025).
  • Financial Modeling: D-Wave-based QBMs trained on foreign exchange market data generate plausible synthetic samples but currently underperform classical RBMs by \simone order of magnitude in KL divergence, reflecting sampling noise and hardware limitations. Empirical effective temperatures \sim100 mK, well above cryostat temperature, suggest a practical limit for true quantum-enhanced sampling until future hardware generations (Perot, 2023).
  • Continuous Data Generation: CVQBMs achieve high fidelity (F0.995F \sim 0.995) and low KL divergence (<102<10^{-2}) for both synthetic and real-world continuous distributions, such as SAR images, leveraging bosonic encodings and variational QITE (Bangar et al., 10 May 2024).
  • State Tomography and Hamiltonian Learning: QBMs admit efficient relative-entropy-based training, achieving quantum state reconstruction and generative modeling with parameter-efficient architectures (Kieferova et al., 2016, Huijgen et al., 2023).
  • Generative Modeling in High Energy Physics: Fully-visible QBMs learn high-dimensional jet event distributions, matching complex higher-order correlations where sparse classical BMs fail (Tüysüz et al., 21 Oct 2024).

6. Hardware Implementation, Limitations, and Scalability

Deployment of QBMs on current or near-term platforms faces several constraints:

  • QPU Topologies and Minor-Embedding: The size and connectivity of quantum annealers (e.g., D-Wave Pegasus) limit embeddable model size, with 20\leq 20 hidden units for reliable sampling when running parallel anneals (Schuman et al., 18 Jul 2025). Chain breaks and qubit noise reduce effective capacity.
  • Temperature Control and Sampling Fidelity: The mapping from physical control parameters to Hamiltonian coefficients is indirect; effective sampling temperatures are significantly higher than device base temperatures (Perot, 2023, Srivastava et al., 2020).
  • Sampling Overhead: Classical Gibbs sampling is orders of magnitude slower than block-parallel QA; quantum speedup is realized when PQA or similar partitioning is feasible (Schuman et al., 18 Jul 2025).
  • Trainability and Barren Plateaus: Restricted QBMs and general non-commuting architectures suffer from vanishing gradients in large systems. Fully-visible QBMs and sqRBMs are empirically free from this pathology, supporting sample-efficient learning (Coopmans et al., 2023, Demidik et al., 24 Feb 2025).
  • Integration with Classical Optimization: All practical QBM implementations are hybrid, with quantum sampling/expectation estimation feeding into classical optimizers for parameter updates (Newton, Adam, natural gradient, EM) (Srivastava et al., 2020, Kimura et al., 29 Jul 2025).

7. Outlook, Open Problems, and Future Directions

  • Universality and Quantum Learning: QBMs with sufficient hidden or evolved structure, e.g., interleaved imaginary/real time, have been proven universal for quantum computation, supporting arbitrary state preparation within ϵ\epsilon accuracy (Wu et al., 2020, Minervini et al., 6 Jan 2025).
  • Hybrid and Information-Geometric Approaches: The em algorithm and extensions of the quantum Arimoto–Blahut algorithm have established a framework for robust, monotonic learning in QBMs, suggesting a pathway around non-convexity and barren plateaus (Kimura et al., 29 Jul 2025).
  • Scalability to High-Dimensional and Quantum Data: Pre-training strategies, parameter pruning, and problem-adapted connectivity (graph-imposed Hamiltonian terms) reduce resource cost and accelerate learning for realistic systems (Tüysüz et al., 21 Oct 2024, Coopmans et al., 2023).
  • Hardware Advances and Algorithmic Co-Design: Improved annealing schedules, increased connectivity (Pegasus \to Zephyr), active error-mitigation, and classical–quantum co-design promise further scalability and possible quantum advantage in generative learning or discriminative modeling (Schuman et al., 18 Jul 2025, Kālis et al., 2023).
  • Extension to Continuous Variables and Bosonic Systems: CVQBMs, currently demonstrated on single-mode distributions, could, with improved squeezing and detection hardware, naturally model complex continuous distributions for applications in remote sensing, finance, or quantum state engineering (Bangar et al., 10 May 2024).
  • Benchmarks and Applications: QBMs are now established as competitive with leading classical models (RBMs, CNNs) in various benchmark tasks, while offering new capabilities unique to quantum statistics. However, realization of robust, scalable quantum generative modeling awaits further advances in both theory and hardware.

Key open challenges involve hardware-limited sampling fidelity, effective scaling of general non-commuting architectures, deployment strategies for large quantum data sets, and integration of quantum-adapted model selection and inference protocols. These directions form the core of ongoing QBM research.

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Quantum Boltzmann Machines.