4.7 Article

Accelerating Extreme Search of Multidimensional Functions Based on Natural Gradient Descent with Dirichlet Distributions

Journal

MATHEMATICS
Volume 10, Issue 19, Pages -

Publisher

MDPI
DOI: 10.3390/math10193556

Keywords

natural gradient descent; optimization; K-L divergence; Dirichlet distribution; generalized Dirichlet distribution

Categories

Funding

  1. Russian Science Foundation [22-71-00009, 21-71-00017]
  2. North Caucasus Center for Mathematical Research
  3. Ministry of Science and Higher Education of the Russian Federation

Ask authors/readers for more resources

This paper introduces a method for extreme searching of multidimensional functions based on the natural gradient descent algorithm, which can help improve the accuracy of neural networks and reduce training time. By describing the multidimensional surface with probability distributions, the natural gradient can reduce the accuracy changes of gradient and step size. Experiments have shown that this approach can effectively converge in various types of neural networks.
The high accuracy attainment, using less complex architectures of neural networks, remains one of the most important problems in machine learning. In many studies, increasing the quality of recognition and prediction is obtained by extending neural networks with usual or special neurons, which significantly increases the time of training. However, engaging an optimization algorithm, which gives us a value of the loss function in the neighborhood of global minimum, can reduce the number of layers and epochs. In this work, we explore the extreme searching of multidimensional functions by proposed natural gradient descent based on Dirichlet and generalized Dirichlet distributions. The natural gradient is based on describing a multidimensional surface with probability distributions, which allows us to reduce the change in the accuracy of gradient and step size. The proposed algorithm is equipped with step-size adaptation, which allows it to obtain higher accuracy, taking a small number of iterations in the process of minimization, compared with the usual gradient descent and adaptive moment estimate. We provide experiments on test functions in four- and three-dimensional spaces, where natural gradient descent proves its ability to converge in the neighborhood of global minimum. Such an approach can find its application in minimizing the loss function in various types of neural networks, such as convolution, recurrent, spiking and quantum networks.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available