4.7 Article

Multi-objective optimization of neural network with stochastic directed search

Journal

EXPERT SYSTEMS WITH APPLICATIONS
Volume 237, Issue -, Pages -

Publisher

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.eswa.2023.121535

Keywords

Multi-objective optimization; Neural networks; Continuation method; Pareto optimization; Directed search

Ask authors/readers for more resources

This paper proposes a novel approach, the stochastic directed search, for optimizing deep neural networks. The method allows efficient fine-tuning of neural networks and performs well in solving high-dimensional multi-objective problems. The effectiveness of the algorithm is demonstrated through experiments.
This work proposes a novel approach, the stochastic directed search, for optimizing deep neural networks framed as high-dimensional multi-objective problems which typically cannot be addressed with current methods. The proposed algorithm is based on the gradient-based predictor-corrector directed search method and it allows the efficient fine-tuning of a neural network (NN) without retraining the entire model. The contributions include the computation of the Jacobian using batches of data to account for GPU memory limitations, special routines for better approximating boundaries, and an early stopping criterion that limits the search space. Additionally, the proposed method scales well with the problem dimension given that it employs the Jacobian to steer the search in the objective space. The effectiveness of the algorithm is exemplified by fine-tuning a forecasting NN model, responsible for producing multi-horizon quantile forecasts of the S&P 500 Futures financial time series. The stochastic directed search optimizes the neural network in only 1.5% of the training time, i.e. 29.56 s, and the Pareto front obtained shows that variations of around 30% can be obtained for the objective with a degradation of only 5% in the loss function. When solving large problem instances with dimensions up to 240,000, the results show that it outperforms NSGA-II and NSGA-III by reducing function evaluations by a factor of 100, while increasing by more than 5% the hypervolume of Pareto fronts. Notably, this work showcases how a large NN model can be fine-tuned using a multi-objective framework.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available