Papers
Topics
Authors
Recent
Search
2000 character limit reached

Quantum Volume (QV): Benchmarking NISQ Devices

Updated 23 April 2026
  • Quantum Volume (QV) is a holistic metric that quantifies NISQ device capabilities by measuring the largest size of random, square circuits with heavy-output probabilities exceeding a 2/3 threshold.
  • The methodology involves executing random circuits with equal width and depth and aggregating heavy-output probabilities across many trials to evaluate device performance.
  • Optimizations in compiler techniques and error mitigation can significantly boost effective QV, making it a key performance indicator for both hardware and software advancements.

Quantum Volume (QV) is a system-level, single-number metric quantifying the balanced interplay of width, depth, fidelity, connectivity, and compilation quality for a noisy intermediate-scale quantum (NISQ) device. QV captures the largest size of a random, square quantum circuit (equal width and depth) that a device can execute such that the measured heavy-output probability (HOP) exceeds a fixed threshold, typically 2/3. Critically, QV is designed to be architecture-agnostic and holistic, folding hardware, compiler, and device variability into a single operational benchmark (Pelofske et al., 2022).

1. Formal Definition and Benchmarking Protocol

Formally, a device’s quantum volume is

QV=2d\mathrm{QV}=2^d

where dd is the largest integer for which the device can reliably run random mm-qubit, depth-mm circuits (with m=dm=d) and achieve mean heavy-output probability (HOP) above threshold, with high statistical confidence. The “heavy set” for each random circuit UU is defined as

HU={x:pU(x)>pmedian}H_U=\{x: p_U(x)>p_\text{median}\}

where pU(x)=xU02p_U(x)=|\langle x|U|0\rangle|^2 and pmedianp_\text{median} is the median of all outcome probabilities. The heavy-output probability for each random instance is the fraction of measurement outcomes in the heavy set. The QV pass criterion requires that over kk circuits (commonly dd0), the mean HOP dd1, with the lower dd2 bound and the dd3 confidence lower bound also exceeding dd4 (Pelofske et al., 2022, Wack et al., 2021).

Circuit construction proceeds layer-wise: every layer consists of a random permutation of all dd5 qubits and a set of disjoint two-qubit SU(4) random unitaries. If dd6 is odd, one qubit is idle each layer. The device "passes" for width dd7 if the statistical pass criteria above are satisfied (Cross et al., 2018, Mori et al., 2022).

2. Holistic System Characterization

Quantum Volume is expressly designed to be a holistic system benchmark (Amico et al., 2023, Cross et al., 2018, Jurcevic et al., 2020). It probes the interplay and limitations arising from:

  • Gate fidelities: Imperfect single- and two-qubit gate errors accumulate with increasing depth, reducing HOP at larger circuit sizes.
  • Qubit connectivity: Non-all-to-all connectivity increases SWAP overhead during random permutations, compounding error and reducing effective maximum dd8.
  • Coherence time: Maximum executable depth at a given width is ultimately coherence limited.
  • Compiler optimization: Improved gate decompositions, noise-adaptive routing, and advanced transpiler routines can compress circuit depth and decrease cumulative error.
  • Variability: QV is sensitive to gate calibration drifts, cross-talk, and device-to-device as well as subset-to-subset performance variations (Pelofske et al., 2022).

This comprehensive scope is achieved by essentially requiring the device to successfully execute deep, wide, entangling, and randomly structured circuits—which collectively simulate realistic complex workloads (in a worst-case sense)—at scale.

3. Statistical and Practical Considerations

The QV test is operationalized as follows:

  • For each width dd9, generate mm0 random depth-mm1 circuits.
  • For each circuit, obtain HOP by comparing measured bitstrings to the classically precomputed heavy set.
  • Aggregate results over all circuits and test if mean HOP and confidence intervals clear the mm2 threshold (Baldwin et al., 2021).

The ideal mean HOP is mm3 for large mm4 (Porter-Thomas limit). In the fully depolarized limit, HOP falls to 1/2. The threshold mm5 is chosen to robustly exclude trivial classical sampling (Baldwin et al., 2021). For physically feasible QV tests (up to mm6 classically), the test is computationally bounded by classical simulation for heavy set identification; beyond this, scalable mirror-circuit or parity-constrained variants are required (Bistroń et al., 4 Feb 2025, Amico et al., 2023).

Typical QV experiments on major platforms include IBM Q, IonQ, Rigetti, OQC, Quantinuum, with measured values (2022) ranging from mm7 (OQC Lucy) to mm8 (Quantinuum H1-2). Device-to-device and subset-to-subset variability is significant—only specific qubit subsets satisfy the QV pass for a given mm9, and the same subset may drift in/out of pass region over time (Pelofske et al., 2022).

4. Compilation, Transpilation, and Error Mitigation Impact

Compilation and classical pre-processing are major QV determinants. Enhanced transpilation (qubit subset enumeration, noise-adaptive layout, advanced routing, and pulse-level schedule optimization) has been shown to substantially boost achievable QV compared to black-box, default transpilation (Jurcevic et al., 2020, Pelofske et al., 2022). Custom passmanagers, such as IBM’s QV passmanager (incorporating CPLEX routing and pulse optimization), yield fewer compilation failures and higher maximal mm0, at the cost of significant classical resources (upwards of 100,000 CPU-hours for mm1).

Error mitigation techniques (notably zero-noise extrapolation and dynamical decoupling) have demonstrably increased effective QV by one or more increments (e.g., mm2 from mm3 to mm4 on IBM devices), as measured either with the canonical or mirror quantum volume protocol (LaRose et al., 2022, Pelofske et al., 2023). The “effective quantum volume” must be reported together with applied mitigation and shot overhead to maintain a fair metric (Amico et al., 2023, Pelofske et al., 2023).

5. Variants, Extensions, and Limitations

Original QV specifies “square circuits” (width = depth). Miller et al. extend the concept with Quantum Volumetric Classes, QV-mm5, targeting random circuits of width mm6 and depth mm7. These variants (QV-2, QV-3, etc.) better capture the scaling of circuit depths for workloads with mm8, aligning volumetric benchmarking with algorithmic structure (Miller et al., 2022, Rodenburg, 31 Jan 2025). In practical benchmarking (91% of surveyed quantum algorithms), square (QV-1), quadratic (QV-2), and cubic (QV-3) circuit families suffice.

QV is resource intensive: typical experiments require mm9 random circuits and m=dm=d0 shots/circuit. Scalability to m=dm=d1 is classically blocked by exponential cost of the heavy set identification. Parity-constrained QV variants eliminate the O(m=dm=d2) simulation overhead by structuring the random circuits to have known heavy output subspaces, enabling benchmarking of devices with size far beyond classical simulation (Bistroń et al., 4 Feb 2025). Mirror Quantum Volume (MQV) and related techniques use inversion-based circuits to avoid this bottleneck, but may mask some error types (Amico et al., 2023).

QV has several caveats: it does not directly reflect algorithmic performance for non-scrambling, highly structured workloads; it ignores circuit speed (addressed by metrics such as CLOPS); its single-number nature gives only coarse granularity; and its accuracy depends on transparent reporting of compiler/mapping details (Pelofske et al., 2022, Wack et al., 2021).

6. Theoretical Foundations and Modeling

Quantum Volume reflects a rigorous theoretical interplay between native error rates, connectivity, and gate durations. The effective error per circuit layer governs the achievable m=dm=d3; for simple depolarizing models, the cross-over point at which mean HOP drops below m=dm=d4 is set by cumulative error per layer exceeding approximately m=dm=d5 (Jaeger et al., 2024). For NISQ architectures, QV can be estimated analytically from gate error rates, connectivity exponent m=dm=d6, and circuit depth scaling (Rodenburg, 31 Jan 2025). Extensions to fault-tolerant architectures incorporate error-corrected logical gates and magic-state distillation overheads to predict QV-k in the logical regime.

Platform-specific QV definitions exist for photonic and measurement-based architectures. For photonic MBQC with Gottesman-Kitaev-Preskill encoding, QV is analytically derived as a function of GKP squeezing and photon transmission efficiency via error-mapping to effective Pauli error probability, then standard QV pass criteria (Zhang et al., 2022).

Topologically, Quantum Volume appears in the context of two-dimensional insulators as the Brillouin zone integral of the square root of the quantum-metric determinant; in this setting, QV variation tracks topological phase transitions and bounds the count of symmetry-protected boundary states (Chiu, 29 Jan 2025).

7. Current Best Practices and Outlook

Best practice prescriptions for maximizing QV include:

  • Enumerate connected qubit subsets and select for highest HOP.
  • Employ aggressive compiler optimizations: noise-adaptive layouts, advanced routing, pulse-level schedule.
  • Exploit dynamic decoupling and readout error mitigation wherever permissible.
  • Document and report error mitigation overheads and compilation-level optimizations for transparency (Pelofske et al., 2022, Amico et al., 2023).

Principal open challenges include: standardizing compiler stacks for reproducibility, reducing classical overhead for heavy set computations (especially for m=dm=d7), integrating QV full-stack analysis with error-corrected logical layers, benchmarking non-square (rectangular) algorithmic workloads, and systematically correlating QV to application performance metrics (Pelofske et al., 2022).

Quantum Volume remains the prevailing system-level metric for quantifying NISQ device capabilities. Its robustness, platform-independence, and sensitivity to both hardware and software-stack optimizations have made it foundational for benchmarking and progress tracking in quantum computing (Pelofske et al., 2022, Cross et al., 2018, Wack et al., 2021).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Quantum Volume (QV).