4.6 Article

Fluctuations, bias, variance and ensemble of learners: exact asymptotics for convex losses in high-dimension

出版社

IOP Publishing Ltd
DOI: 10.1088/1742-5468/ad0221

关键词

learning theory; machine learning

向作者/读者索取更多资源

This manuscript develops a quantitative and rigorous theory for studying the fluctuations in an ensemble of generalized linear models trained on high-dimensional correlated features. The results can be applied to various classification and regression tasks and help understand the impact of ensembling on test error as well as the roots of the "double-descent" phenomenon.
From the sampling of data to the initialisation of parameters, randomness is ubiquitous in modern Machine Learning practice. Understanding the statistical fluctuations engendered by the different sources of randomness in prediction is therefore key to understanding robust generalisation. In this manuscript we develop a quantitative and rigorous theory for the study of fluctuations in an ensemble of generalised linear models trained on different, but correlated, features in high-dimensions. In particular, we provide a complete description of the asymptotic joint distribution of the empirical risk minimiser for generic convex loss and regularisation in the high-dimensional limit. Our result encompasses a rich set of classification and regression tasks, such as the lazy regime of overparametrised neural networks, or equivalently the random features approximation of kernels. While allowing to study directly the mitigating effect of ensembling (or bagging) on the bias-variance decomposition of the test error, our analysis also helps disentangle the contribution of statistical fluctuations, and the singular role played by the interpolation threshold that are at the roots of the 'double-descent' phenomenon.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据