4.7 Article

Classification vs regression in overparameterized regimes: Does the loss function matter?

期刊

出版社

MICROTOME PUBL

关键词

classification; regression; overparameterized; support vector machines; survival; contamination

资金

  1. NSF [CCF-1740833, AST-144078, ECCS-1343398]
  2. Sloan Research Fellowship
  3. Google

向作者/读者索取更多资源

This study compares classification and regression tasks in an overparameterized linear model with Gaussian features. The findings show that under sufficient overparameterization, all training points are support vectors, and the solutions obtained by least-squares minimum-norm interpolation are identical to those produced by the hard margin support vector machine. Additionally, the study reveals the existence of regimes where interpolating solutions generalize well when evaluated by the 0-1 test loss function, but do not generalize when evaluated by the square loss function.
We compare classification and regression tasks in an overparameterized linear model with Gaussian features. On the one hand, we show that with sufficient overparameterization all training points are support vectors: solutions obtained by least-squares minimum-norm interpolation, typically used for regression, are identical to those produced by the hard margin support vector machine (SVM) that minimizes the hinge loss, typically used for training classifiers. On the other hand, we show that there exist regimes where these interpolating solutions generalize well when evaluated by the 0-1 test loss function, but do not generalize if evaluated by the square loss function, i.e. they approach the null risk. Our results demonstrate the very different roles and properties of loss functions used at the training phase (optimization) and the testing phase (generalization).

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据