A Unified Convergence Analysis of Block Successive Minimization Methods for Nonsmooth OptimizationSIAM Journal on Optimization (2013)
The block coordinate descent (BCD) method is widely used for minimizing a continuous function f of several block variables. At each iteration of this method, a single block of variables is optimized, while the remaining variables are held fixed. To ensure the convergence of the BCD method, the subproblem of each block variable needs to be solved to its unique global optimal. Unfortunately, this requirement is often too restrictive for many practical scenarios. In this paper, we study an alternative inexact BCD approach which updates the variable blocks by successively minimizing a sequence of approximations of f which are either locally tight upper bounds of f or strictly convex local approximations of f. The main contributions of this work include the characterizations of the convergence conditions for a fairly wide class of such methods, especially for the cases where the objective functions are either nondifferentiable or nonconvex. Our results unify and extend the existing convergence results for many classical algorithms such as the BCD method, the difference of convex functions (DC) method, the expectation maximization (EM) algorithm, as well as the block forward-backward splitting algorithm, all of which are popular for large scale optimization problems involving big data.
- block coordinate descent,
- block successive upper-bound minimization,
- successive convex approximation,
- successive inner approximation
Citation InformationMeisam Razaviyayn, Mingyi Hong and Zhi-Quan Luo. "A Unified Convergence Analysis of Block Successive Minimization Methods for Nonsmooth Optimization" SIAM Journal on Optimization Vol. 23 Iss. 2 (2013)
Available at: http://works.bepress.com/mingyi_hong/1/