Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Investigating the limits of randomized benchmarking protocols (1308.2928v1)

Published 13 Aug 2013 in quant-ph

Abstract: In this paper, we analyze the performance of randomized benchmarking protocols on gate sets under a variety of realistic error models that include systematic rotations, amplitude damping, leakage to higher levels, and 1/f noise. We find that, in almost all cases, benchmarking provides better than a factor-of-two estimate of average error rate, suggesting that randomized benchmarking protocols are a valuable tool for verification and validation of quantum operations. In addition, we derive new models for fidelity decay curves under certain types of non-Markovian noise models such as 1/f and leakage errors. We also show that, provided the standard error of the fidelity measurements is small, only a small number of trials are required for high confidence estimation of gate errors.

Citations (121)

Summary

We haven't generated a summary for this paper yet.