4.6 Article

CONVERGENCE RATE ANALYSIS OF A SEQUENTIAL CONVEX PROGRAMMING METHOD WITH LINE SEARCH FOR A CLASS OF CONSTRAINED DIFFERENCE-OF-CONVEX OPTIMIZATION PROBLEMS

Journal

SIAM JOURNAL ON OPTIMIZATION
Volume 31, Issue 3, Pages 2024-2054

Publisher

SIAM PUBLICATIONS
DOI: 10.1137/20M1314057

Keywords

constrained optimization; difference-of-convex optimization; Kurdyka-Lojasiewicz property; linear convergence

Funding

  1. Hong Kong Research Grants Council [PolyU153005/17p]

Ask authors/readers for more resources

This paper studies the sequential convex programming method with monotone line search (SCPls) for a class of difference-of-convex optimization problems with smooth inequality constraints. The convergence rate of the sequence generated by SCPls in nonconvex and convex settings under suitable Kurdyka-Lojasiewicz (KL) assumptions is analyzed. The paper also discusses how to deduce the KL exponent of the extended objective function from its Lagrangian in convex settings.
In this paper, we study the sequential convex programming method with monotone line search (SCPls) in [Z. Lu, Sequential Convex Programming Methods for a Class of Structured Nonlinear Programming, https://arxiv.org/abs/1210.3039, 2012] for a class of difference-of-convex optimization problems with multiple smooth inequality constraints. The SCPls is a representative variant of moving-ball-approximation-type algorithms [A. Auslender, R. Shefi, and M. Teboulle, SIAM J. Optim., 20 (2010), pp. 3232-3259; J. Bolte, Z. Chen, and E. Pauwels, Math. Program., 182 (2020) pp. 1-36; J. Bolte and E. Pauwels, Math. Oper. Res., 41 (2016), pp. 442-465; R. Shefi and M. Teboulle, Math. Program., 159 (2016), pp. 137-164] for constrained optimization problems. We analyze the convergence rate of the sequence generated by SCPls in both nonconvex and convex settings by imposing suitable Kurdyka-Lojasiewicz (KL) assumptions. Specifically, in the nonconvex settings, we assume that a special potential function related to the objective and the constraints is a KL function, while in the convex settings we impose KL assumptions directly on the extended objective function (i.e., sum of the objective and the indicator function of the constraint set). A relationship between these two different KL assumptions is established in the convex settings under additional differentiability assumptions. We also discuss how to deduce the KL exponent of the extended objective function from its Lagrangian in the convex settings, under additional assumptions on the constraint functions. Thanks to this result, the extended objectives of some constrained optimization models such as minimizing l(1) subject to logistic/Poisson loss are found to be KL functions with exponent 1/2 under mild assumptions. To illustrate how our results can be applied, we consider SCPls for minimizing l(1-2) [P. Yin, Y. Lou, Q. He, and J. Xin, SIAM J. Sci. Comput., 37 (2015), pp. A536-A563] subject to residual error measured by l(2) norm/Lorentzian norm [R. E. Carrillo, A. B. Ramirez, G. R. Arce, K. E. Barner, and B. M. Sadler, EURASIP J. Adv. Signal Process. (2016), 108]. We first discuss how the various conditions required in our analysis can be verified, and then perform numerical experiments to illustrate the convergence behaviors of SCPls.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available