Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

High-Fidelity Magic-State Preparation with a Biased-Noise Architecture (2109.02677v1)

Published 6 Sep 2021 in quant-ph

Abstract: Magic state distillation is a resource intensive subroutine that consumes noisy input states to produce high-fidelity resource states that are used to perform logical operations in practical quantum-computing architectures. The resource cost of magic state distillation can be reduced by improving the fidelity of the raw input states. To this end, we propose an initialization protocol that offers a quadratic improvement in the error rate of the input magic states in architectures with biased noise. This is achieved by preparing an error-detecting code which detects the dominant errors that occur during state preparation. We obtain this advantage by exploiting the native gate operations of an underlying qubit architecture that experiences biases in its noise profile. We perform simulations to analyze the performance of our protocol with the XZZX surface code. Even at modest physical parameters with a two-qubit gate error rate of $0.7\%$ and total probability of dominant errors in the gate $O(103)$ larger compared to that of non-dominant errors, we find that our preparation scheme delivers magic states with logical error rate $O(10{-8})$ after a single round of the standard 15-to-1 distillation protocol; two orders of magnitude lower than using conventional state preparation. Our approach therefore promises considerable savings in overheads with near-term technology.

Citations (11)

Summary

We haven't generated a summary for this paper yet.