4.6 Article

ON THE NONASYMPTOTIC CONVERGENCE OF CYCLIC COORDINATE DESCENT METHODS

Journal

SIAM JOURNAL ON OPTIMIZATION
Volume 23, Issue 1, Pages 576-601

Publisher

SIAM PUBLICATIONS
DOI: 10.1137/110840054

Keywords

convex optimization; cyclic coordinate descent; convergence rates; sparsity

Ask authors/readers for more resources

Cyclic coordinate descent is a classic optimization method that has witnessed a resurgence of interest in signal processing, statistics, and machine learning. Reasons for this renewed interest include the simplicity, speed, and stability of the method, as well as its competitive performance on l(1) regularized smooth optimization problems. Surprisingly, very little is known about its nonasymptotic convergence behavior on these problems. Most existing results either just prove convergence or provide asymptotic rates. We fill this gap in the literature by proving O(1/k) convergence rates (where k is the iteration count) for two variants of cyclic coordinate descent under an isotonicity assumption. Our analysis proceeds by comparing the objective values attained by the two variants with each other, as well as with the gradient descent algorithm. We show that the iterates generated by the cyclic coordinate descent methods remain better than those of gradient descent uniformly over time.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available