Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 102 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 30 tok/s
GPT-5 High 27 tok/s Pro
GPT-4o 110 tok/s
GPT OSS 120B 475 tok/s Pro
Kimi K2 203 tok/s Pro
2000 character limit reached

$k$-extendibility of high-dimensional bipartite quantum states (1504.06459v2)

Published 24 Apr 2015 in quant-ph, math.FA, and math.PR

Abstract: The idea of detecting the entanglement of a given bipartite state by searching for symmetric extensions of this state was first proposed by Doherty, Parrilo and Spedialeri. The complete family of separability tests it generates, often referred to as the hierarchy of $k$-extendibility tests, has already proved to be most promising. The goal of this paper is to try and quantify the efficiency of this separability criterion in typical scenarios. For that, we essentially take two approaches. First, we compute the average width of the set of $k$-extendible states, in order to see how it scales with the one of separable states. And second, we characterize when random-induced states are, depending on the ancilla dimension, with high probability violating or not the $k$-extendibility test, and compare the obtained result with the corresponding one for entanglement vs separability. The main results can be precisely phrased as follows: on $\mathbf{C}d\otimes\mathbf{C}d$, when $d$ grows, the average width of the set of $k$-extendible states is equivalent to $(2/\sqrt{k})/d$, while random states obtained as partial traces over an environment $\mathbf{C}s$ of uniformly distributed pure states are violating the $k$-extendibility test with probability going to $1$ if $s<((k-1)2/4k)d2$. Both statements converge to the conclusion that, if $k$ is fixed, $k$-extendibility is asymptotically a weak approximation of separability, even though any of the other well-studied separability relaxations is outperformed by $k$-extendibility as soon as $k$ is above a certain (dimension independent) value.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Authors (1)

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube