4.8 Article

Fast and Accurate Least-Mean-Squares Solvers for High Dimensional Data

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2021.3139612

关键词

Regression; Least Mean Squares Solvers; Coresets; Sketches; Caratheodory's Theorem; Big Data

向作者/读者索取更多资源

This article introduces a new algorithm that can compute subsets with positive weights in a short amount of time, which is crucial for solving machine learning problems and matrix factorizations. The algorithm combines data summarization techniques like sketches and coresets, providing performance boost for existing LMS solvers.
Least-mean-squares (LMS) solvers such as Linear / Ridge-Regression and SVD not only solve fundamental machine learning problems, but are also the building blocks in a variety of other methods, such as matrix factorizations. We suggest an algorithm that gets a finite set of n d-dimensional real vectors and returns a subset of d + 1 vectors with positive weights whose weighted sum is exactly the same. The constructive proof in Caratheodory's Theorem computes such a subset in O(n(2)d(2)) time and thus not used in practice. Our algorithm computes this subset in O(nd + d(4)log n) time, using O(log n) calls to Caratheodory's construction on small but smart subsets. This is based on a novel paradigm of fusion between different data summarization techniques, known as sketches and coresets. For large values of d, we suggest a faster construction that takes O(nd) time and returns a weighted subset of O(d) sparsified input points. Here, a sparsified point means that some of its entries were set to zero. As an application, we show how to boost the performance of existing LMS solvers, such as those in scikit-learn library, up to x100. Generalization for streaming and distributed data is trivial. Extensive experimental results and open source code are provided.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据