4.6 Article

Lower bounds for non-convex stochastic optimization

期刊

MATHEMATICAL PROGRAMMING
卷 199, 期 1-2, 页码 165-214

出版社

SPRINGER HEIDELBERG
DOI: 10.1007/s10107-022-01822-7

关键词

-

向作者/读者索取更多资源

In this work, we establish lower bounds on the complexity of finding epsilon-stationary points using stochastic first-order methods. We prove that any algorithm accessing smooth, potentially non-convex functions through queries to an unbiased stochastic gradient oracle with bounded variance requires at least epsilon(-4) queries to find an epsilon-stationary point in the worst case. The tightness of the lower bound establishes the minimax optimality of stochastic gradient descent in this model. In a stricter model where the noisy gradient estimates satisfy a mean-squared smoothness property, we further prove a lower bound of epsilon(-3) queries, establishing the optimality of recently proposed variance reduction techniques.
We lower bound the complexity of finding epsilon-stationary points (with gradient norm at most epsilon) using stochastic first-order methods. In a well-studied model where algorithms access smooth, potentially non-convex functions through queries to an unbiased stochastic gradient oracle with bounded variance, we prove that (in the worst case) any algorithm requires at least epsilon(-4) queries to find an epsilon-stationary point. The lower bound is tight, and establishes that stochastic gradient descent is minimax optimal in this model. In a more restrictive model where the noisy gradient estimates satisfy a meansquared smoothness property, we prove a lower bound of epsilon(-3) queries, establishing the optimality of recently proposed variance reduction techniques.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据