3.8 Proceedings Paper

On the Use of Low-discrepancy Sequences in the Training of Neural Networks

期刊

LARGE-SCALE SCIENTIFIC COMPUTING (LSSC 2021)
卷 13127, 期 -, 页码 421-430

出版社

SPRINGER INTERNATIONAL PUBLISHING AG
DOI: 10.1007/978-3-030-97549-4_48

关键词

Quasi-Monte Carlo algorithms; Neural networks; Stochastic gradient descent

资金

  1. Bulgarian Ministry of Education and Science [D01-205/23.11.2018]
  2. Ministry of Education and Science (MES) of Bulgaria [DO1-322/18.12.2019, D01-363/17.12.2020]
  3. National Center for High-performance and Distributed Computing (NCHDC), part of National Roadmap of RIs [D01-387/18.12.2020]
  4. Science and Education for Smart Growth Operational Program [BG05M2OP001-1.001-0003]
  5. European Union through the European structural and investment funds

向作者/读者索取更多资源

Quasi-Monte Carlo methods utilize specially designed deterministic sequences to achieve higher convergence rates compared to random numbers. The usefulness of low-discrepancy sequences, such as Sobol and Halton sequences, has been established in various fields including Mathematical Finance, optimization, and machine learning. This study explores different approaches to efficiently employ low-discrepancy sequences in the training process of neural networks, highlighting their advantage in benchmark problems and discussing practical issues in real-life applications.
The quasi-Monte Carlo methods use specially designed deterministic sequences with improved uniformity properties compared with random numbers, in order to achieve higher rates of convergence. Usually certain measures like the discrepancy are used in order to quantify these uniformity properties. The usefulness of certain families of sequences with low discrepancy, like the Sobol and Halton sequences, has been established in problems with high practical value as in Mathematical Finance. Multiple studies have been done about applying these sequences also in the domains of optimisation and machine learning. Currently many types of neural networks are used extensively to achieve break-through results in Machine Learning and Artificial Intelligence. The process of training these networks requires substantial computational resources, usually provided by using powerful GPUs or specially designed hardware. In this work we study different approaches to employ efficiently low-discrepancy sequences at various places in the training process where their uniformity properties can speed-up or improve the training process. We demonstrate the advantage of using Sobol low-discrepancy sequences in benchmark problems and we discuss various practical issues that arise in order to achieve acceptable performance in real-life problems.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据