Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 70 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 37 tok/s Pro
GPT-5 High 34 tok/s Pro
GPT-4o 21 tok/s Pro
Kimi K2 191 tok/s Pro
GPT OSS 120B 448 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Quantum Diffusion Models (2311.15444v1)

Published 26 Nov 2023 in quant-ph

Abstract: We propose a quantum version of a generative diffusion model. In this algorithm, artificial neural networks are replaced with parameterized quantum circuits, in order to directly generate quantum states. We present both a full quantum and a latent quantum version of the algorithm; we also present a conditioned version of these models. The models' performances have been evaluated using quantitative metrics complemented by qualitative assessments. An implementation of a simplified version of the algorithm has been executed on real NISQ quantum hardware.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (51)
  1. “Deep unsupervised learning using nonequilibrium thermodynamics” In International conference on machine learning, 2015, pp. 2256–2265 PMLR
  2. Jonathan Ho, Ajay Jain and Pieter Abbeel “Denoising diffusion probabilistic models” In Advances in neural information processing systems 33, 2020, pp. 6840–6851
  3. “Diffusion Models in Vision: A Survey” In IEEE Transactions on Pattern Analysis and Machine Intelligence 45.9, 2023, pp. 10850–10869 DOI: 10.1109/TPAMI.2023.3261988
  4. Jiaming Song, Chenlin Meng and Stefano Ermon “Denoising Diffusion Implicit Models”, 2022 arXiv:2010.02502 [cs.LG]
  5. Diederik P Kingma and Max Welling “Auto-encoding variational bayes” In arXiv preprint arXiv:1312.6114, 2013
  6. “Generative adversarial networks” In Communications of the ACM 63.11 ACM New York, NY, USA, 2020, pp. 139–144
  7. “Surface codes: Towards practical large-scale quantum computation” In Phys. Rev. A 86 American Physical Society, 2012, pp. 032324 DOI: 10.1103/PhysRevA.86.032324
  8. Savvas Varsamopoulos, Ben Criger and Koen Bertels “Decoding small surface codes with feedforward neural networks” In Quantum Science and Technology 3.1 IOP Publishing, 2017, pp. 015004 DOI: 10.1088/2058-9565/aa955a
  9. “Convolutional neural network based decoders for surface codes” In Quantum Information Processing 22.3, 2023, pp. 151 DOI: 10.1007/s11128-023-03898-2
  10. E Knill “Quantum computing with realistically noisy devices” In Nature 434, 2005, pp. 39–44 DOI: 10.1038/nature03350
  11. “High-Threshold Fault-Tolerant Quantum Computation with Analog Quantum Error Correction” In Phys. Rev. X 8 American Physical Society, 2018, pp. 021054 DOI: 10.1103/PhysRevX.8.021054
  12. John Preskill “Quantum computing in the NISQ era and beyond” In Quantum 2 Verein zur Förderung des Open Access Publizierens in den Quantenwissenschaften, 2018, pp. 79
  13. Peter W. Shor “Polynomial-Time Algorithms for Prime Factorization and Discrete Logarithms on a Quantum Computer” In SIAM Journal on Computing 26.5 Society for Industrial & Applied Mathematics (SIAM), 1997, pp. 1484–1509 DOI: 10.1137/s0097539795293172
  14. “Quantum machine learning” In Nature 549.7671 Springer ScienceBusiness Media LLC, 2017, pp. 195–202 DOI: 10.1038/nature23474
  15. Sau Lan Wu and Shinjae Yoo “Challenges and opportunities in quantum machine learning for high-energy physics” In Nature Reviews Physics 4.3, 2022, pp. 143–144 DOI: 10.1038/s42254-022-00425-7
  16. “Variational quantum algorithms” In Nature Reviews Physics 3.9 Springer ScienceBusiness Media LLC, 2021, pp. 625–644 DOI: 10.1038/s42254-021-00348-9
  17. “A study of the pulse-based variational quantum eigensolver on cross-resonance based hardware”, 2023 arXiv:2303.02410 [quant-ph]
  18. “Long-Lived Particles Anomaly Detection with Parametrized Quantum Circuits” In Particles 6.1, 2023, pp. 297–311 DOI: 10.3390/particles6010016
  19. Matteo Robbiati, Juan M. Cruz-Martinez and Stefano Carrazza “Determining probability density functions with adiabatic quantum computing” 7 pages, 3 figures, 2023 arXiv: http://cds.cern.ch/record/2853183
  20. “A quantum analytical Adam descent through parameter shift rule using Qibo”, 2022 arXiv:2210.10787 [quant-ph]
  21. Juan M. Cruz-Martinez, Matteo Robbiati and Stefano Carrazza “Multi-variable integration with a variational quantum circuit”, 2023 arXiv:2308.05657 [quant-ph]
  22. “A Quantum Convolutional Neural Network on NISQ Devices”, 2021 arXiv:2104.06918 [quant-ph]
  23. “Simulating quench dynamics on a digital quantum computer with data-driven error mitigation” In Quantum Science and Technology 6.4 IOP Publishing, 2021, pp. 045003 DOI: 10.1088/2058-9565/ac0e7a
  24. “Learning-Based Quantum Error Mitigation” In PRX Quantum 2 American Physical Society, 2021, pp. 040330 DOI: 10.1103/PRXQuantum.2.040330
  25. “Scalable Mitigation of Measurement Errors on Quantum Computers” In PRX Quantum 2 American Physical Society, 2021, pp. 040326 DOI: 10.1103/PRXQuantum.2.040326
  26. “Measurement error mitigation in quantum computers through classical bit-flip correction” In Phys. Rev. A 105 American Physical Society, 2022, pp. 062404 DOI: 10.1103/PhysRevA.105.062404
  27. “Quantum generative adversarial networks” In Physical Review A 98.1 APS, 2018, pp. 012324
  28. “Style-based quantum generative adversarial networks for Monte Carlo events” In Quantum 6 Verein zur Forderung des Open Access Publizierens in den Quantenwissenschaften, 2022, pp. 777 DOI: 10.22331/q-2022-08-17-777
  29. “Quantum variational autoencoder” In Quantum Science and Technology 4.1 IOP Publishing, 2018, pp. 014001
  30. Marco Parigi, Stefano Martina and Filippo Caruso “Quantum-Noise-driven Generative Diffusion Models” In arXiv preprint arXiv:2308.12013, 2023
  31. “Generative quantum machine learning via denoising diffusion probabilistic models” In arXiv preprint arXiv:2310.05866, 2023
  32. “Parameterized quantum circuits as machine learning models” In Quantum Science and Technology 4.4 IOP Publishing, 2019, pp. 043001
  33. “Circuit-centric quantum classifiers” In Physical Review A 101.3 APS, 2020, pp. 032308
  34. “Quantum circuit learning” In Physical Review A 98.3 APS, 2018, pp. 032309
  35. “Evaluating analytic gradients on quantum hardware” In Physical Review A 99.3 APS, 2019, pp. 032331
  36. “Pennylane: Automatic differentiation of hybrid quantum-classical computations” In arXiv preprint arXiv:1811.04968, 2018
  37. “Machine learning with quantum computers” Springer, 2021
  38. Maria Schuld, Ryan Sweke and Johannes Jakob Meyer “Effect of data encoding on the expressive power of variational quantum-machine-learning models” In Phys. Rev. A 103 American Physical Society, 2021, pp. 032430 DOI: 10.1103/PhysRevA.103.032430
  39. “An introduction to deep generative modeling” In GAMM-Mitteilungen 44.2 Wiley Online Library, 2021, pp. e202100008
  40. “Deep Unsupervised Learning using Nonequilibrium Thermodynamics”, 2015 arXiv:1503.03585 [cs.LG]
  41. John A. Cortese and Timothy M. Braje “Loading Classical Data into a Quantum Computer”, 2018 arXiv:1803.01958 [quant-ph]
  42. “Stabilization of quantum computations by symmetrization” In SIAM Journal on Computing 26.5 SIAM, 1997, pp. 1541–1557
  43. Michael A Nielsen “The entanglement fidelity and quantum error correction” In arXiv preprint quant-ph/9606012, 1996
  44. Jürgen Schmidhuber “Deep learning in neural networks: An overview” In Neural networks 61 Elsevier, 2015, pp. 85–117
  45. Ian T Jolliffe and Jorge Cadima “Principal component analysis: a review and recent developments” In Philosophical transactions of the royal society A: Mathematical, Physical and Engineering Sciences 374.2065 The Royal Society Publishing, 2016, pp. 20150202
  46. Laurens Van der Maaten and Geoffrey Hinton “Visualizing data using t-SNE.” In Journal of machine learning research 9.11, 2008
  47. Andrew P Bradley “The use of the area under the ROC curve in the evaluation of machine learning algorithms” In Pattern recognition 30.7 Elsevier, 1997, pp. 1145–1159
  48. “Gans trained by a two time-scale update rule converge to a local nash equilibrium” In Advances in neural information processing systems 30, 2017
  49. “Evaluating generative networks using Gaussian mixtures of image features” In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2023, pp. 279–288
  50. “Barren plateaus in quantum neural network training landscapes” In Nature communications 9.1 Nature Publishing Group UK London, 2018, pp. 4812
  51. “Dynamics of superconducting qubit relaxation times”, 2022 arXiv:2105.15201 [quant-ph]
Citations (4)

Summary

  • The paper introduces quantum diffusion models that extend classical diffusion techniques to directly generate quantum states using parameterized quantum circuits.
  • It presents a dual approach with full and latent models, applying iterative noise addition and learned reverse denoising to approximate target states.
  • The study evaluates implementation challenges on NISQ devices, details training objectives via quantum variational methods, and assesses performance using fidelity metrics.

Quantum Diffusion Models (QDMs), as proposed in "Quantum Diffusion Models" (2311.15444), represent an extension of classical generative diffusion models to the quantum domain. The core objective is the direct generation of quantum states, ρ\rho, leveraging parameterized quantum circuits (PQCs) in place of classical neural networks within the diffusion framework. This approach aims to harness the principles of diffusion—iterative noise addition followed by learned denoising—for tasks involving quantum state preparation and quantum generative modeling.

Model Architecture and Diffusion Process

The QDM framework adapts the forward and reverse processes characteristic of classical diffusion models.

Forward Process (Noise Addition): A target quantum state ρ0\rho_0 is gradually corrupted over a sequence of TT discrete time steps by applying quantum noise channels, Nt\mathcal{N}_t. This process generates a sequence of increasingly noisy states {ρ1,ρ2,...,ρT}\{\rho_1, \rho_2, ..., \rho_T\}, where ρt=Nt(ρt1)\rho_t = \mathcal{N}_t(\rho_{t-1}). The noise schedule, often defined by parameters βt\beta_t, controls the amount of noise added at each step. The sequence is designed such that the final state ρT\rho_T approximates a tractable distribution, typically the maximally mixed state ρTI2N\rho_T \approx \frac{I}{2^N} for an NN-qubit system. The choice of quantum channel Nt\mathcal{N}_t is crucial; possibilities include depolarizing channels, amplitude damping, or phase damping channels, applied locally or globally. The overall forward process can be described by the transition kernel q(ρtρt1)q(\rho_t | \rho_{t-1}).

Reverse Process (Denoising/Generation): The generative process starts from the maximally mixed state ρT\rho_T and iteratively applies a learned reverse quantum channel Mθt\mathcal{M}_{\theta_t} to denoise the state, aiming to recover the original state distribution: ρt1Mθt(ρt)\rho_{t-1} \approx \mathcal{M}_{\theta_t}(\rho_t). This reverse process is parameterized by PQCs, denoted U(θt)U(\theta_t). The PQC takes the noisy state ρt\rho_t as input (potentially encoded via ancilla qubits or specific measurement schemes) and applies a unitary transformation U(θt)U(\theta_t), possibly followed by partial trace or other non-unitary operations derived from measurements, to approximate the denoised state ρt1\rho_{t-1}. The learnable parameters θt\theta_t within the PQC U(θt)U(\theta_t) are optimized during training to accurately model the reverse transitions pθ(ρt1ρt)p_\theta(\rho_{t-1} | \rho_t).

The PQC U(θt)U(\theta_t) serves as the quantum analogue of the neural network in classical diffusion models. Its architecture (ansatz) typically consists of layers of single-qubit rotations and multi-qubit entangling gates, where the rotation angles are functions of the parameters θt\theta_t and potentially the time step tt.

Full vs. Latent Quantum Diffusion Models

The paper proposes two main variants:

1. Full Quantum Diffusion Model: In this version, both the forward and reverse processes operate directly on quantum states residing in the Hilbert space. * Implementation: The forward process involves applying sequences of quantum channels. The reverse process uses PQCs U(θt)U(\theta_t) at each step tt to approximate the inverse of the noise channel Nt\mathcal{N}_t. Training optimizes the parameters {θt}t=1T\{\theta_t\}_{t=1}^T to maximize the likelihood of generating states from the target distribution, often via a quantum variational lower bound (ELBO) or by minimizing a quantum analogue of the score-matching objective. The input to the PQC at step tt is the quantum state ρt\rho_t, and the output is intended to be ρt1\rho_{t-1}. * Challenges: Requires coherent manipulation and storage of quantum states throughout the diffusion process, making it demanding for NISQ hardware. Measurement and state tomography might be needed during training or sampling, adding overhead. Defining and implementing the parameterized reverse channel Mθt\mathcal{M}_{\theta_t} based on the unitary U(θt)U(\theta_t) requires careful consideration (e.g., using ancilla-based methods or quantum channel models).

2. Latent Quantum Diffusion Model: This hybrid approach combines a classical diffusion model operating in a latent space with a quantum generative circuit. * Implementation: A classical diffusion model generates latent vectors zRdz \in \mathbb{R}^d. These latent vectors are then mapped to target quantum states ψ(z)|\psi(z)\rangle or ρ(z)\rho(z) using a fixed or parameterized quantum circuit, GϕG_\phi. This generator GϕG_\phi could be, for instance, a PQC trained variationally or potentially a form of Quantum Generative Adversarial Network (QGAN) generator. The diffusion process itself (noise addition and denoising) occurs entirely classically on the latent variables zz. The forward process corrupts latent vectors z0z_0 to zTN(0,I)z_T \sim \mathcal{N}(0, I), and the reverse process learns to map ztz_t to zt1z_{t-1}. Sampling involves running the classical reverse diffusion to get z0z_0 and then preparing the state ρ(z0)\rho(z_0) using GϕG_\phi. * Advantages: Leverages mature classical diffusion techniques for the core diffusion dynamics, potentially simplifying training and implementation. The quantum component GϕG_\phi is only needed for the final state generation step. * Challenges: Requires designing an effective mapping from the latent space to the quantum state space. The expressivity and trainability of the quantum generator circuit GϕG_\phi are critical. The entanglement structure and specific properties of the target quantum states must be captured by this mapping.

Conditioned Quantum Diffusion Models

Conditioning allows generating quantum states that satisfy specific properties or belong to a certain class. Similar to classical conditional diffusion, information cc (e.g., desired energy level, specific entanglement pattern, classical label) can be incorporated into the QDM.

  • Implementation: The conditioning information cc can be embedded into the PQC parameters θt\theta_t or provided as an additional input to the PQC U(θt,c)U(\theta_t, c) at each step of the reverse process. For instance, parameters in the PQC could be made functions of cc, or cc could be encoded onto auxiliary qubits that interact with the main system qubits during the unitary evolution U(θt,c)U(\theta_t, c). In the latent QDM, conditioning is typically handled within the classical diffusion model operating on the latent space (z,c)(z, c). The final quantum state ρ(z0,c)\rho(z_0, c) would then be generated based on the sampled conditioned latent vector z0z_0.

Implementation Details and Training

Training Objective: The training typically aims to optimize the parameters θ={θt}t=1T\theta = \{\theta_t\}_{t=1}^T of the reverse process PQCs. This can be framed as maximizing a variational lower bound (ELBO) on the log-likelihood of the target quantum state distribution, adapted to the quantum setting. Alternatively, a loss function based on minimizing the difference between the PQC-driven reverse transition pθ(ρt1ρt)p_\theta(\rho_{t-1} | \rho_t) and the true posterior q(ρt1ρt,ρ0)q(\rho_{t-1} | \rho_t, \rho_0) can be formulated. This often simplifies to a form analogous to score matching, possibly involving minimizing the distance (e.g., trace distance or Hilbert-Schmidt distance) between the output of the PQC and the expected denoised state. L(θ)=t=1TEq(ρ0)Eq(ρtρ0)[D(q(ρt1ρt,ρ0)pθ(ρt1ρt))]L(\theta) = \sum_{t=1}^T E_{q(\rho_0)} E_{q(\rho_t | \rho_0)} [ D( q(\rho_{t-1} | \rho_t, \rho_0) || p_\theta(\rho_{t-1} | \rho_t) ) ], where DD is a suitable distance measure between quantum states or channels.

Gradient Computation: Gradients of the loss function with respect to the PQC parameters θt\theta_t are required for optimization. The parameter-shift rule or other quantum gradient estimation techniques (e.g., finite differences, linear response theory) can be employed. This typically involves executing the PQC multiple times with shifted parameters and measuring appropriate observables.

Sampling: Generating a new quantum state involves starting from the maximally mixed state ρT\rho_T and iteratively applying the trained reverse PQCs U(θ^t)U(\hat{\theta}_t) for t=T,T1,...,1t = T, T-1, ..., 1.

1
2
3
4
5
6
7
8
9
function sample_qdm(T, trained_pqcs U_theta):
  // Initialize state to maximally mixed state
  rho = I / (2**N)
  for t from T down to 1:
    // Apply the trained PQC for step t
    // This might involve encoding rho_t and applying U_theta_t
    // Or directly applying a learned channel M_theta_t
    rho = apply_reverse_step(rho, U_theta[t], t)
  return rho // The generated quantum state rho_0

NISQ Implementation: The paper reports an implementation of a simplified QDM on real NISQ hardware. This likely involved:

  • Small number of qubits (e.g., 2-4 qubits).
  • Shallow PQC ansätze to mitigate decoherence and gate errors.
  • A small number of diffusion steps TT.
  • Simplified noise models (e.g., local depolarizing noise).
  • Extensive measurement and error mitigation techniques. The evaluation would compare the experimentally generated states against classically simulated ideal states or target states using metrics like fidelity. Challenges include PQC trainability (barren plateaus), gate infidelity, readout errors, and limited qubit connectivity.

Evaluation Metrics

Performance evaluation combines quantitative and qualitative methods:

  • Quantitative:
    • Fidelity: F(ρgen,ρtarget)=(Tr[ρtargetρgenρtarget])2F(\rho_{gen}, \rho_{target}) = (\text{Tr}[\sqrt{\sqrt{\rho_{target}} \rho_{gen} \sqrt{\rho_{target}}}])^2. Measures the closeness between the generated state ρgen\rho_{gen} and the target state ρtarget\rho_{target}.
    • Trace Distance: T(ρgen,ρtarget)=12ρgenρtarget1T(\rho_{gen}, \rho_{target}) = \frac{1}{2} ||\rho_{gen} - \rho_{target}||_1. Provides another measure of distinguishability.
    • Observable Expectation Values: Comparing Oρgen\langle O \rangle_{\rho_{gen}} with Oρtarget\langle O \rangle_{\rho_{target}} for relevant observables OO (e.g., Hamiltonians, entanglement witnesses).
    • Entanglement Measures: Quantifying entanglement (e.g., concurrence, negativity) in generated states if the target states are entangled.
  • Qualitative: Assessing properties of the generated state ensemble, such as the distribution of measurement outcomes in a specific basis or visualizing state representations (e.g., Bloch sphere for single qubits, Q-functions).

Practical Implications and Applications

QDMs offer a potentially powerful framework for generative tasks in the quantum domain:

  • Quantum State Preparation: Generating specific ground states of Hamiltonians, preparing resource states for quantum computation (e.g., cluster states), or initializing quantum algorithms. Conditioned QDMs could prepare states with specific energy or entanglement properties.
  • Quantum Simulation: Learning distributions of states relevant to physical systems, potentially aiding in the paper of many-body physics or quantum chemistry.
  • Quantum Machine Learning: Serving as generative models within broader QML pipelines, for tasks like anomaly detection or data augmentation on quantum datasets.
  • Error Mitigation: Potentially learning to reverse the effects of noise channels, although this application requires further investigation.

Compared to other quantum generative models like QGANs or Quantum Boltzmann Machines, QDMs might offer more stable training dynamics, similar to their classical counterparts, although training PQCs remains challenging. The iterative refinement process might allow for generating complex states with high fidelity. However, the computational cost, particularly for the full QDM requiring coherent quantum evolution, can be significant in terms of circuit depth and coherence times. The latent QDM shifts some burden to classical computation but relies heavily on the effectiveness of the classical-to-quantum mapping.

Limitations and Future Directions

Current limitations primarily stem from NISQ hardware constraints: qubit count, coherence times, gate fidelities, and connectivity severely restrict the size (NN) and depth (TT and PQC depth) of implementable QDMs. Barren plateaus can hinder the training of deep PQCs. The theoretical understanding of QDM convergence, expressivity, and the optimal choice of quantum noise channels and PQC ansätze requires further development.

Future research directions include:

  • Developing more hardware-efficient PQC ansätze and training methods for QDMs.
  • Exploring alternative quantum noise models and diffusion schedules tailored for specific quantum systems.
  • Rigorous theoretical analysis of QDM properties and comparison with other quantum generative approaches.
  • Applying QDMs to specific problems in physics, chemistry, and materials science.
  • Investigating fault-tolerant implementations of QDMs.
  • Improving the classical-to-quantum mapping in latent QDMs.

Conclusion

Quantum Diffusion Models (2311.15444) introduce a novel approach to quantum state generation by adapting the classical diffusion paradigm. Utilizing Parameterized Quantum Circuits for the reverse denoising process, QDMs offer pathways (full quantum and latent) to iteratively construct target quantum states from noise. While practical implementation faces significant hurdles on current NISQ devices, the framework presents a promising direction for generative modeling in quantum computation and simulation, with potential applications ranging from state preparation to quantum machine learning. Further research into efficient implementations, theoretical properties, and practical applications will be crucial to realizing their full potential.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Youtube Logo Streamline Icon: https://streamlinehq.com

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube