4.6 Article

STOCHASTIC FIRST- AND ZEROTH-ORDER METHODS FOR NONCONVEX STOCHASTIC PROGRAMMING

Journal

SIAM JOURNAL ON OPTIMIZATION
Volume 23, Issue 4, Pages 2341-2368

Publisher

SIAM PUBLICATIONS
DOI: 10.1137/120880811

Keywords

stochastic approximation; nonconvex optimization; stochastic programming; simulation-based optimization

Funding

  1. NSF [CMMI-1000347, DMS-1319050]
  2. ONR [N00014-13-1-0036]
  3. NSF CAREER Award [CMMI-1254446]

Ask authors/readers for more resources

In this paper, we introduce a new stochastic approximation type algorithm, namely, the randomized stochastic gradient (RSG) method, for solving an important class of nonlinear (possibly nonconvex) stochastic programming problems. We establish the complexity of this method for computing an approximate stationary point of a nonlinear programming problem. We also show that this method possesses a nearly optimal rate of convergence if the problem is convex. We discuss a variant of the algorithm which consists of applying a postoptimization phase to evaluate a short list of solutions generated by several independent runs of the RSG method, and we show that such modification allows us to improve significantly the large-deviation properties of the algorithm. These methods are then specialized for solving a class of simulation-based optimization problems in which only stochastic zeroth-order information is available.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available