Coderivative-Based Newton Methods in Structured Nonconvex and Nonsmooth Optimization (2403.04262v2)
Abstract: This paper proposes and develops new Newton-type methods to solve structured nonconvex and nonsmooth optimization problems with justifying their fast local and global convergence by means of advanced tools of variational analysis and generalized differentiation. The objective functions belong to a broad class of prox-regular functions with specification to constrained optimization of nonconvex structured sums. We also develop a novel line search method, which is an extension of the proximal gradient algorithm while allowing us to globalize the proposed coderivative-based Newton methods by incorporating the machinery of forward-backward envelopes. Applications and numerical experiments, which are provided for nonconvex least squares regression models, Student's t-regression, and image restoration problems, demonstrate the efficiency of the proposed algorithms