4.8 Review

Fitting elephants in modern machine learning by statistically consistent interpolation

期刊

NATURE MACHINE INTELLIGENCE
卷 3, 期 5, 页码 378-386

出版社

NATURE PORTFOLIO
DOI: 10.1038/s42256-021-00345-8

关键词

-

资金

  1. Crick-Clay Professorship (CSHL)
  2. H. N. Mahabala Chair Professorship (IIT Madras)

向作者/读者索取更多资源

This article discusses the good generalization ability of modern machine learning methods when interpolating noisy data, introduces the phenomenon of statistically consistent interpolation (SCI), explains the reasons for successful data interpolation, and compares the different approaches of modern machine learning, traditional physical theory, and biological brains in modeling natural phenomena.
Modern machine learning approaches, such as deep neural networks, generalize well despite interpolating noisy data, in contrast with textbook wisdom. Mitra describes the phenomenon of statistically consistent interpolation (SCI) to clarify why data interpolation succeeds, and discusses how SCI elucidates the differing approaches to modelling natural phenomena represented in modern machine learning, traditional physical theory and biological brains. Textbook wisdom advocates for smooth function fits and implies that interpolation of noisy data should lead to poor generalization. A related heuristic is that fitting parameters should be fewer than measurements (Occam's razor). Surprisingly, contemporary machine learning approaches, such as deep nets, generalize well, despite interpolating noisy data. This may be understood via statistically consistent interpolation (SCI), that is, data interpolation techniques that generalize optimally for big data. Here, we elucidate SCI using the weighted interpolating nearest neighbours algorithm, which adds singular weight functions to k nearest neighbours. This shows that data interpolation can be a valid machine learning strategy for big data. SCI clarifies the relation between two ways of modelling natural phenomena: the rationalist approach (strong priors) of theoretical physics with few parameters, and the empiricist (weak priors) approach of modern machine learning with more parameters than data. SCI shows that the purely empirical approach can successfully predict. However, data interpolation does not provide theoretical insights, and the training data requirements may be prohibitive. Complex animal brains are between these extremes, with many parameters, but modest training data, and with prior structure encoded in species-specific mesoscale circuitry. Thus, modern machine learning provides a distinct epistemological approach that is different both from physical theories and animal brains.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据