Article ID Journal Published Year Pages File Type
4637816 Journal of Computational and Applied Mathematics 2017 17 Pages PDF
Abstract

The alternating direction method with multipliers (ADMM) has been one of most powerful and successful methods for solving a two-block linearly constrained convex minimization model whose objective function is the sum of two functions without coupled variables. It is known that the numerical efficiency is inherited for a large number of applications, but the convergence is not guaranteed if the ADMM is directly extended to a multiple-block convex minimization model whose objective function has more than two functions. This viewpoint was in fact the motivation for developing efficient algorithms that cannot only preserve the numerical advantages of the direct extension of ADMM but also guarantee convergence. One way is to correct the output of the direct extension of ADMM slightly via a simple correction step, and the other is to employ a simple proximal to solve inexactly each subproblem in the direct extension of ADMM. In this paper, in order to solve the multi-block separable convex minimization model efficiently, we present a method which is a combination of the above two ways, that is, we first solve each subproblem with a simple proximal, then we correct the output via a simple correction step. Theoretically, we derive global convergence results for this method and establish a worst-case O(1/k)O(1/k) iteration complexity. Numerically, the efficiency of this method can be showed by testing the problem of recovering low-rank and sparse components of matrices from incomplete and noisy observation.

Related Topics
Physical Sciences and Engineering Mathematics Applied Mathematics
Authors
, ,