4.0 Article

Stable Bayesian optimization

Journal

Publisher

SPRINGERNATURE
DOI: 10.1007/s41060-018-0119-9

Keywords

Bayesian optimization; Gaussian process; Stable Bayesian optimization; Acquisition function

Funding

  1. Australian Government through the Australian Research Council (ARC)
  2. Telstra-Deakin Centre of Excellence in Big Data and Machine Learning
  3. ARC Australian Laureate Fellowship [FL170100006]
  4. Australian Research Council [FL170100006] Funding Source: Australian Research Council

Ask authors/readers for more resources

Tuning hyperparameters of machine learning models is important for their performance. Bayesian optimization has recently emerged as a de-facto method for this task. The hyperparameter tuning is usually performed by looking at model performance on a validation set. Bayesian optimization is used to find the hyperparameter set corresponding to the best model performance. However, in many cases, the function representing the model performance on the validation set contains several spurious sharp peaks due to limited datapoints. The Bayesian optimization, in such cases, has a tendency to converge to sharp peaks instead of other more stable peaks. When a model trained using these hyperparameters is deployed in the real world, its performance suffers dramatically. We address this problem through a novel stable Bayesian optimization framework. We construct two new acquisition functions that help Bayesian optimization to avoid the convergence to the sharp peaks. We conduct a theoretical analysis and guarantee that Bayesian optimization using the proposed acquisition functions prefers stable peaks over unstable ones. Experiments with synthetic function optimization and hyperparameter tuning for support vector machines show the effectiveness of our proposed framework.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.0
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available