A preconditioned deepest descent algorithm for a class of optimization problems involving the $p(x)$-Laplacian operator
Abstract: In this paper we are concerned with a class of optimization problems involving the $p(x)$-Laplacian operator, which arise in imaging and signal analysis. We study the well-posedness of this kind of problems in an amalgam space considering that the variable exponent $p(x)$ is a log-H\"older continuous function. Further, we propose a preconditioned descent algorithm for the numerical solution of the problem, considering a "frozen exponent" approach in a finite dimension space. Finally, we carry on several numerical experiments to show the advantages of our method. Specifically, we study two detailed example whose motivation lies in a possible extension of the proposed technique to image processing.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.