4.5 Article

Calculus of the Exponent of Kurdyka-Aojasiewicz Inequality and Its Applications to Linear Convergence of First-Order Methods

Journal

FOUNDATIONS OF COMPUTATIONAL MATHEMATICS
Volume 18, Issue 5, Pages 1199-1232

Publisher

SPRINGER
DOI: 10.1007/s10208-017-9366-8

Keywords

First-order methods; Convergence rate; Kurdyka-Lojasiewicz inequality; Linear convergence; Luo-Tseng error bound; Sparse optimization

Funding

  1. Australian Research Council [FT130100038]
  2. Hong Kong Research Grants Council [PolyU253008/15p]
  3. Australian Research Council [FT130100038] Funding Source: Australian Research Council

Ask authors/readers for more resources

In this paper, we study the Kurdyka-Aojasiewicz (KL) exponent, an important quantity for analyzing the convergence rate of first-order methods. Specifically, we develop various calculus rules to deduce the KL exponent of new (possibly nonconvex and nonsmooth) functions formed from functions with known KL exponents. In addition, we show that the well-studied Luo-Tseng error bound together with a mild assumption on the separation of stationary values implies that the KL exponent is . The Luo-Tseng error bound is known to hold for a large class of concrete structured optimization problems, and thus we deduce the KL exponent of a large class of functions whose exponents were previously unknown. Building upon this and the calculus rules, we are then able to show that for many convex or nonconvex optimization models for applications such as sparse recovery, their objective function's KL exponent is . This includes the least squares problem with smoothly clipped absolute deviation regularization or minimax concave penalty regularization and the logistic regression problem with regularization. Since many existing local convergence rate analysis for first-order methods in the nonconvex scenario relies on the KL exponent, our results enable us to obtain explicit convergence rate for various first-order methods when they are applied to a large variety of practical optimization models. Finally, we further illustrate how our results can be applied to establishing local linear convergence of the proximal gradient algorithm and the inertial proximal algorithm with constant step sizes for some specific models that arise in sparse recovery.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available