When does subtracting a rank-one approximation decrease tensor rank?
Abstract: Subtracting a critical rank-one approximation from a matrix always results in a matrix with a lower rank. This is not true for tensors in general. Motivated by this, we ask the question: what is the closure of the set of those tensors for which subtracting some of its critical rank-one approximation from it and repeating the process we will eventually get to zero? In this article, we show how to construct this variety of tensors and we show how this is connected to the bottleneck points of the variety of rank-one tensors (and in general to the singular locus of the hyperdeterminant), and how this variety can be equal to and in some cases be more than (weakly) orthogonally decomposable tensors.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.