Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 24 tok/s Pro
GPT-5 High 25 tok/s Pro
GPT-4o 113 tok/s Pro
Kimi K2 216 tok/s Pro
GPT OSS 120B 428 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Quantum Circuit Born Machine (QCBM)

Updated 19 October 2025
  • QCBM is a quantum generative model that encodes classical probability distributions into the measurement statistics of quantum states, offering a compact and intractable description for classical data.
  • The training methodology leverages gradient-based optimization with maximum mean discrepancy (MMD) loss and the parameter-shift rule, circumventing the need for explicit likelihoods.
  • Empirical evidence shows that deeper circuits and unbiased gradient estimators significantly improve sample fidelity on tasks like Bars-and-Stripes and Gaussian mixtures, highlighting potential quantum advantages.

A Quantum Circuit Born Machine (QCBM) is a quantum generative model that encodes a classical probability distribution into the measurement statistics of a parameterized quantum state. By engineering a quantum circuit comprising layers of single-qubit rotations and entangling gates, a QCBM represents the distribution in the amplitudes of the resulting pure state. Sampling is performed via projective measurement in the computational basis, yielding classical bit strings from which a generative model can be constructed. This approach offers the possibility of leveraging quantum computational resources to model and sample from distributions that may be intractable for classical generative models.

1. Representation of Probability Distributions in QCBMs

The fundamental principle underlying the QCBM is the mapping of a classical distribution to the squared amplitudes of a quantum pure state. Starting from the initial product state 0n|0\rangle^{\otimes n}, the parameterized circuit U(θ)U(\theta) evolves the state to %%%%2%%%%. The probability of measuring a bit string x{0,1}nx \in \{0,1\}^n is

pθ(x)=xψθ2.p_\theta(x) = |\langle x|\psi_\theta\rangle|^2.

The entire classical distribution is defined implicitly by these measurement probabilities. This encodes potentially highly-correlated and nontrivial classical statistics within a compact quantum description.

Computational complexity arguments assert a greater expressivity for quantum circuits vis-à-vis classical neural networks; it is #P-hard to sample from the output distribution of certain quantum circuits, and thus QCBMs can represent distributions inaccessible to polynomial-size classical generative models. Random circuit sampling and paradigms related to quantum supremacy support the conjecture of a fundamental separation in generative capacity.

2. Sampling and the Challenge of Training without Likelihoods

Sampling in a QCBM is performed simply by repeated projective measurements in the computational basis, yielding bit strings xx sampled according to pθ(x)p_\theta(x). The measurement process is efficient and naturally suited to quantum hardware: no elaborate postprocessing or classical sampling loops are needed.

However, a critical limitation in both theory and practice is the lack of tractable expression for pθ(x)p_\theta(x). As with classical implicit generative models (e.g. GANs), QCBMs cannot compute likelihoods directly except for trivial cases. This impedes the direct use of maximum likelihood estimation (MLE) and necessitates the use of alternative loss functions suitable for “likelihood-free” training. As a result, methods based on two-sample tests and kernel embeddings become essential.

3. Gradient-Based Learning Algorithms: MMD and the Parameter-Shift Rule

QCBM training proceeds via minimization of a loss function that quantifies the divergence between the model-generated and the target distribution. The squared maximum mean discrepancy (MMD) is employed as a distributional two-sample test in a reproducing kernel Hilbert space (RKHS), defined as

LMMD=Expθ[ϕ(x)]Eyπ[ϕ(y)]2,\mathcal{L}_{\text{MMD}} = \left\| E_{x\sim p_\theta}[\phi(x)] - E_{y\sim \pi}[\phi(y)] \right\|^2,

which can be expanded as

Ex,xpθ[K(x,x)]2Expθ,yπ[K(x,y)]+Ey,yπ[K(y,y)],E_{x,x'\sim p_\theta}[K(x,x')] - 2 E_{x\sim p_\theta, y\sim \pi}[K(x,y)] + E_{y,y'\sim \pi}[K(y,y')],

where KK is a positive definite kernel. Training is conducted by sampling bit strings from both the model and the data, evaluating the kernel, and computing the loss.

Crucially, the gradient of the MMD loss with respect to circuit parameters θ\theta is estimated using the parameter-shift rule: for a gate U(η)=exp(iηΣ/2)U(\eta) = \exp(-i\eta \Sigma/2) with Σ2=1\Sigma^2=1, the derivative of the probability is

pθ(x)θ=12[pθ+(x)pθ(x)],\frac{\partial p_\theta(x)}{\partial \theta} = \frac{1}{2} \left[ p_{\theta^+}(x) - p_{\theta^-}(x) \right],

with parameter shifts θ±=θ±π2\theta^\pm = \theta \pm \frac{\pi}{2}. This method yields an unbiased gradient estimator, compatible with optimization algorithms such as Adam or L-BFGS-B. Unlike finite-difference methods, this parameter-shift rule leverages the analytic structure of quantum gates and supports efficient hybrid training.

4. Empirical Results: Expressivity, Circuit Depth, and Optimization

Experimental evaluation was undertaken on the Bars-and-Stripes (BAS) dataset (a 3×3 binary image task with 14 valid patterns) and Gaussian mixture distributions. For BAS, a 9-qubit QCBM of depth d=10d=10 with data-informed Chow-Liu-tree-inspired connectivity was used. The empirical findings are as follows:

  • Gradient-based optimizers (Adam, L-BFGS-B) dramatically reduced the MMD loss—valid sample generation rates reached 99.9% using computed gradients and appropriate circuit architectures.
  • Circuit depth proved essential: deeper (higher dd) circuits captured more complex structure, and shallow circuits underfit the target distribution.
  • Optimizers exploiting unbiased gradients outperformed gradient-free approaches (CMA-ES), especially in the presence of statistical noise from finite sampling.

On Gaussian mixtures encoded with 10 qubits, the QCBM matched the smooth target histogram with high fidelity. Importantly, the paper found no evidence of vanishing or exploding gradients (unlike classical deep neural networks), even for deep quantum circuits, owing to the intrinsic properties of unitary quantum evolution.

5. Practical Feasibility and Quantum Advantage

The proposed learning algorithm operates as a quantum–classical hybrid, requiring only circuit preparation, measurement, and classical postprocessing. Moderate measurement numbers (NN per gradient estimate) suffice for robust learning, which aligns with the measurement overheads of near-term quantum hardware. The absence of vanishing gradients and the effectiveness of projective measurement-based sampling suggest practical deployability on current devices.

Potential quantum advantages derive from both expressivity—capability to encode classically intractable distributions—and sampling efficiency. This points toward QCBMs being candidates for quantum-enhanced generative modeling, combinatorial optimization, and learning in high-dimensional spaces.

Prospective avenues include alternative loss functions (such as kernel Sinkhorn divergences or adversarial variants) and more sophisticated circuit architectures optimized for specific quantum devices. The flexibility of the QCBM framework suggests its suitability for diverse tasks where quantum modeling of distributions, especially beyond classical feasibility, is desired.

6. Summary and Outlook

The QCBM paradigm offers a rigorous route to representing and learning classical probability distributions as quantum pure states, sampled via projective measurements. The lack of tractable likelihoods is overcome by gradient-based training using MMD loss and parameter-shift gradients, which are efficient and robust even with deep circuits. The model’s strong empirical performance on nontrivial generative tasks underscores the importance of circuit depth and optimizer selection. Importantly, the algorithm is both noise-resilient and well-poised for experimental demonstration on near-term quantum hardware. Extensions to combinatorial optimization, complex density modeling, and possible hybridization with adversarial training further enrich the research landscape for QCBMs (Liu et al., 2018).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Quantum Circuit Born Machine (QCBM).