Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient iterative thresholding algorithms with functional feedbacks and convergence analysis (2005.06693v1)

Published 14 May 2020 in cs.IT and math.IT

Abstract: An accelerated class of adaptive scheme of iterative thresholding algorithms is studied analytically and empirically. They are based on the feedback mechanism of the null space tuning techniques (NST+HT+FB). The main contribution of this article is the accelerated convergence analysis and proofs with a variable/adaptive index selection and different feedback principles at each iteration. These convergence analysis require no longer a priori sparsity information $s$ of a signal. %key theory in this paper is the concept that the number of indices selected at each iteration should be considered in order to speed up the convergence. It is shown that uniform recovery of all $s$-sparse signals from given linear measurements can be achieved under reasonable (preconditioned) restricted isometry conditions. Accelerated convergence rate and improved convergence conditions are obtained by selecting an appropriate size of the index support per iteration. The theoretical findings are sufficiently demonstrated and confirmed by extensive numerical experiments. It is also observed that the proposed algorithms have a clearly advantageous balance of efficiency, adaptivity and accuracy compared with all other state-of-the-art greedy iterative algorithms.

Summary

We haven't generated a summary for this paper yet.