4.7 Article

Multi-objective optimization of neural network with stochastic directed search

期刊

EXPERT SYSTEMS WITH APPLICATIONS
卷 237, 期 -, 页码 -

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.eswa.2023.121535

关键词

Multi-objective optimization; Neural networks; Continuation method; Pareto optimization; Directed search

向作者/读者索取更多资源

This paper proposes a novel approach, the stochastic directed search, for optimizing deep neural networks. The method allows efficient fine-tuning of neural networks and performs well in solving high-dimensional multi-objective problems. The effectiveness of the algorithm is demonstrated through experiments.
This work proposes a novel approach, the stochastic directed search, for optimizing deep neural networks framed as high-dimensional multi-objective problems which typically cannot be addressed with current methods. The proposed algorithm is based on the gradient-based predictor-corrector directed search method and it allows the efficient fine-tuning of a neural network (NN) without retraining the entire model. The contributions include the computation of the Jacobian using batches of data to account for GPU memory limitations, special routines for better approximating boundaries, and an early stopping criterion that limits the search space. Additionally, the proposed method scales well with the problem dimension given that it employs the Jacobian to steer the search in the objective space. The effectiveness of the algorithm is exemplified by fine-tuning a forecasting NN model, responsible for producing multi-horizon quantile forecasts of the S&P 500 Futures financial time series. The stochastic directed search optimizes the neural network in only 1.5% of the training time, i.e. 29.56 s, and the Pareto front obtained shows that variations of around 30% can be obtained for the objective with a degradation of only 5% in the loss function. When solving large problem instances with dimensions up to 240,000, the results show that it outperforms NSGA-II and NSGA-III by reducing function evaluations by a factor of 100, while increasing by more than 5% the hypervolume of Pareto fronts. Notably, this work showcases how a large NN model can be fine-tuned using a multi-objective framework.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据