4.8 Article

Faster First-Order Methods for Stochastic Non-Convex Optimization on Riemannian Manifolds

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2019.2933841

关键词

Optimization; Complexity theory; Manifolds; Convergence; Signal processing algorithms; Stochastic processes; Minimization; Riemannian optimization; stochastic variance-reduced algorithm; non-convex optimization; online learning

资金

  1. NUS [R-263-000-C67-646]
  2. ECRA [R-263-000-C87-133]
  3. MOE Tier-II [R-263-000-D17-112]
  4. Natural Science Foundation of China (NSFC) [61876090]

向作者/读者索取更多资源

The paper introduces an efficient Riemannian Stochastic Path Integrated Differential Estimator algorithm for solving finite and online Riemannian non-convex minimization problems with lower computational cost compared to prior algorithms. The algorithm's recursive gradient estimation mechanism allows it to achieve epsilon-approximate stationary points within a certain number of stochastic gradient evaluations, outperforming other Riemannian optimization methods.
First-order non-convex Riemannian optimization algorithms have gained recent popularity in structured machine learning problems including principal component analysis and low-rank matrix completion. The current paper presents an efficient Riemannian Stochastic Path Integrated Differential EstimatoR (R-SPIDER) algorithm to solve the finite-sum and online Riemannian non-convex minimization problems. At the core of R-SPIDER is a recursive semi-stochastic gradient estimator that can accurately estimate Riemannian gradient under not only exponential mapping and parallel transport, but also general retraction and vector transport operations. Compared with prior Riemannian algorithms, such a recursive gradient estimation mechanism endows R-SPIDER with lower computational cost in first-order oracle complexity. Specifically, for finite-sum problems with n components, R-SPIDER is proved to converge to an epsilon-approximate stationary point within O(min(n + root n/epsilon(2),1/epsilon(3))) stochastic gradient evaluations, beating the best-known complexity O(n+1/epsilon(4)); for online optimization, R-SPIDER is shown to converge with O(1/epsilon(3)) complexity which is, to the best of our knowledge, the first non-asymptotic result for online Riemannian optimization. For the special case of gradient dominated functions, we further develop a variant of R-SPIDER with improved linear rate of convergence. Extensive experimental results demonstrate the advantage of the proposed algorithms over the state-of-the-art Riemannian non-convex optimization methods.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据