Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
153 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Alternating Least Squares Tensor Completion in The TT-Format (1509.00311v1)

Published 1 Sep 2015 in math.NA

Abstract: We consider the problem of fitting a low rank tensor $A\in\mathbb{R}{{\mathcal I}}$, ${\mathcal I} = {1,\ldots,n}{d}$, to a given set of data points ${M_i\in\mathbb{R}\mid i\in P}$, $P\subset{\mathcal I}$. The low rank format under consideration is the hierarchical or TT or MPS format. It is characterized by rank bounds $r$ on certain matricizations of the tensor. The number of degrees of freedom is in ${\cal O}(r2dn)$. For a fixed rank and mode size $n$ we observe that it is possible to reconstruct random (but rank structured) tensors as well as certain discretized multivariate (but rank structured) functions from a number of samples that is in ${\cal O}(\log N)$ for a tensor having $N=nd$ entries. We compare an alternating least squares fit (ALS) to an overrelaxation scheme inspired by the LMaFit method for matrix completion. Both approaches aim at finding a tensor $A$ that fulfils the first order optimality conditions by a nonlinear Gauss-Seidel type solver that consists of an alternating fit cycling through the directions $\mu=1,\ldots,d$. The least squares fit is of complexity ${\cal O}(r4d#P)$ per step, whereas each step of ADF is in ${\cal O}(r2d#P)$, albeit with a slightly higher number of necessary steps. In the numerical experiments we observe robustness of the completion algorithm with respect to noise and good reconstruction capability. Our tests provide evidence that the algorithm is suitable in higher dimension ($>$10) as well as for moderate ranks. Keywords: MPS, Tensor Completion, Tensor Train, TT, Hierarchical Tucker, HT, ALS.

Summary

We haven't generated a summary for this paper yet.