4.5 Article

Optimal deep neural networks by maximization of the approximation power

期刊

COMPUTERS & OPERATIONS RESEARCH
卷 156, 期 -, 页码 -

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.cor.2023.106264

关键词

Machine learning; Artificial intelligence; Data science; Forecasting; Feedforward neural networks

向作者/读者索取更多资源

We propose an optimal architecture for deep neural networks of given size, which optimizes the approximation of linear regions by maximizing the lower bound. The accuracy of the approximation is influenced by the structure of the neural network, including the number, dependence, and hierarchy of nodes within and across layers. Experimental results show that our optimized architecture outperforms cross-validation methods and gridsearch for linear and nonlinear prediction models, as demonstrated on the Boston Housing dataset.
We propose an optimal architecture for deep neural networks of given size. The optimal architecture obtains from maximizing the lower bound of the maximum number of linear regions approximated by a deep neural network with a ReLu activation function. The accuracy of the approximation function relies on the neural network structure characterized by the number, dependence and hierarchy between the nodes within and across layers. We show how the accuracy of the approximation improves as we optimally choose the width and depth of the network. A Monte-Carlo simulation exercise illustrates the outperformance of the optimized architecture against cross-validation methods and gridsearch for linear and nonlinear prediction models. The application of this methodology to the Boston Housing dataset confirms empirically the outperformance of our method against state-of the-art machine learning models.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据