4.7 Article

The generalization performance of ERM algorithm with strongly mixing observations

期刊

MACHINE LEARNING
卷 75, 期 3, 页码 275-295

出版社

SPRINGER
DOI: 10.1007/s10994-009-5104-z

关键词

Generalization performance; ERM principle; Relative uniform convergence; Exponentially strongly mixing

资金

  1. National 973 project [2007CB311002]
  2. NSFC key project [70501030, 10771053]
  3. Foundation of Hubei Educational Committee [Q200710001]

向作者/读者索取更多资源

The generalization performance is the main concern of machine learning theoretical research. The previous main bounds describing the generalization ability of the Empirical Risk Minimization (ERM) algorithm are based on independent and identically distributed (i.i.d.) samples. In order to study the generalization performance of the ERM algorithm with dependent observations, we first establish the exponential bound on the rate of relative uniform convergence of the ERM algorithm with exponentially strongly mixing observations, and then we obtain the generalization bounds and prove that the ERM algorithm with exponentially strongly mixing observations is consistent. The main results obtained in this paper not only extend the previously known results for i.i.d. observations to the case of exponentially strongly mixing observations, but also improve the previous results for strongly mixing samples. Because the ERM algorithm is usually very time-consuming and overfitting may happen when the complexity of the hypothesis space is high, as an application of our main results we also explore a new strategy to implement the ERM algorithm in high complexity hypothesis space.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据