4.8 Article

Scaling Up Generalized Kernel Methods

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2021.3059702

关键词

Kernel; Training; Stochastic processes; Convergence; Computational modeling; Scalability; Optimization; Kernel method; asynchronous parallel computation; stochastic gradient descent; coordinate descent; random feature

资金

  1. National Natural Science Foundation of China [62076138]
  2. Natural Science Foundation [BK20161534]
  3. six talent peaks project [XYDXX-042]
  4. 333 Project in Jiangsu Province [BRA2017455]

向作者/读者索取更多资源

This paper introduces a general sparse kernel learning formulation based on random feature approximation to tackle the challenge of big data era. A new large-scale sparse kernel learning algorithm (AsyDSSKL) is proposed using techniques of asynchronous parallel computation and doubly stochastic optimization. Experimental results demonstrate that AsyDSSKL has significant superiority in computational efficiency over existing kernel methods.
Kernel methods have achieved tremendous success in the past two decades. In the current big data era, data collection has grown tremendously. However, existing kernel methods are not scalable enough both at the training and predicting steps. To address this challenge, in this paper, we first introduce a general sparse kernel learning formulation based on the random feature approximation, where the loss functions are possibly non-convex. In order to reduce the scale of random features required in experiment, we also use that formulation based on the orthogonal random feature approximation. Then we propose a new asynchronous parallel doubly stochastic algorithm for large scale sparse kernel learning (AsyDSSKL). To the best our knowledge, AsyDSSKL is the first algorithm with the techniques of asynchronous parallel computation and doubly stochastic optimization. We also provide a comprehensive convergence guarantee to AsyDSSKL. Importantly, the experimental results on various large-scale real-world datasets show that, our AsyDSSKL method has the significant superiority on the computational efficiency at the training and predicting steps over the existing kernel methods.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据