Results 1 to 10 of about 491 (55)

Convergence of Linear Bregman ADMM for Nonconvex and Nonsmooth Problems with Nonseparable Structure

open access: yesComplexity, 2020
The alternating direction method of multipliers (ADMM) is an effective method for solving two-block separable convex problems and its convergence is well understood.
Miantao Chao, Zhao Deng, Jinbao Jian
doaj   +2 more sources

A Modified Approach to Distributed Bregman ADMM for a Class of Nonconvex Consensus Problems

open access: yesJournal of Mathematics
This article presents a refined iteration of the distributed Bregman alternating direction method of multipliers (ADMM) tailored to tackle nonconvex consensus issues, especially those with multiple blocks.
Zhonghui Xue, Qianfeng Ma, Yazheng Dang
doaj   +2 more sources

Bregman iterative regularization using model functions for nonconvex nonsmooth optimization

open access: yesFrontiers in Applied Mathematics and Statistics, 2022
In this paper, we propose a new algorithm called ModelBI by blending the Bregman iterative regularization method and the model function technique for solving a class of nonconvex nonsmooth optimization problems.
Haoxing Yang   +3 more
doaj   +1 more source

Convergence Analysis of Multiblock Inertial ADMM for Nonconvex Consensus Problem

open access: yesJournal of Mathematics, Volume 2023, Issue 1, 2023., 2023
The alternating direction method of multipliers (ADMM) is one of the most powerful and successful methods for solving various nonconvex consensus problem. The convergence of the conventional ADMM (i.e., 2‐block) for convex objective functions has been stated for a long time.
Yang Liu, Yazheng Dang, Qiang Wu
wiley   +1 more source

Non‐convex nonlocal adaptive tight frame image deblurring

open access: yesIET Image Processing, Volume 16, Issue 7, Page 1908-1923, 29 May 2022., 2022
Abstract The challenge of the image restoration is to recover more detailed information from the degraded images. Based on the observations that wavelet frames have efficient representation ability to image details and the nonconvex regularization in the model may admit unbiased solutions, in this paper, in order to recover more details, a wavelet ...
Zhengwei Shen
wiley   +1 more source

A convex function satisfying the Łojasiewicz inequality but failing the gradient conjecture both at zero and infinity

open access: yesBulletin of the London Mathematical Society, Volume 54, Issue 2, Page 590-608, April 2022., 2022
Abstract We construct an example of a smooth convex function on the plane with a strict minimum at zero, which is real analytic except at zero, for which Thom's gradient conjecture fails both at zero and infinity. More precisely, the gradient orbits of the function spiral around zero and at infinity.
Aris Daniilidis   +2 more
wiley   +1 more source

Factorization of completely positive matrices using iterative projected gradient steps

open access: yesNumerical Linear Algebra with Applications, Volume 28, Issue 6, December 2021., 2021
Abstract We aim to factorize a completely positive matrix by using an optimization approach which consists in the minimization of a nonconvex smooth function over a convex and compact set. To solve this problem we propose a projected gradient algorithm with parameters that take into account the effects of relaxation and inertia.
Radu Ioan Boţ, Dang‐Khoa Nguyen
wiley   +1 more source

A LogTVSCAD Nonconvex Regularization Model for Image Deblurring in the Presence of Impulse Noise

open access: yesDiscrete Dynamics in Nature and Society, Volume 2021, Issue 1, 2021., 2021
This paper proposes a nonconvex model (called LogTVSCAD) for deblurring images with impulsive noises, using the log‐function penalty as the regularizer and adopting the smoothly clipped absolute deviation (SCAD) function as the data‐fitting term. The proposed nonconvex model can effectively overcome the poor performance of the classical TVL1 model for ...
Zhijun Luo   +3 more
wiley   +1 more source

Proximal Linearized Iteratively Reweighted Algorithms for Nonconvex and Nonsmooth Optimization Problem

open access: yesAxioms, 2022
The nonconvex and nonsmooth optimization problem has been attracting increasing attention in recent years in image processing and machine learning research. The algorithm-based reweighted step has been widely used in many applications.
Juyeb Yeo, Myeongmin Kang
doaj   +1 more source

Convergence Rate Analysis of the Proximal Difference of the Convex Algorithm

open access: yesMathematical Problems in Engineering, Volume 2021, Issue 1, 2021., 2021
In this paper, we study the convergence rate of the proximal difference of the convex algorithm for the problem with a strong convex function and two convex functions. By making full use of the special structure of the difference of convex decomposition, we prove that the convergence rate of the proximal difference of the convex algorithm is linear ...
Xueyong Wang   +4 more
wiley   +1 more source

Home - About - Disclaimer - Privacy