4.6 Article

AN ALGORITHM FOR THE PRINCIPAL COMPONENT ANALYSIS OF LARGE DATA SETS

期刊

SIAM JOURNAL ON SCIENTIFIC COMPUTING
卷 33, 期 5, 页码 2580-2594

出版社

SIAM PUBLICATIONS
DOI: 10.1137/100804139

关键词

algorithm; principal component analysis; PCA; singular value decomposition; SVD; low rank

资金

  1. NSF [DMS0748488, DMS0610097]
  2. Israel Science Foundation [485/10]
  3. Alfred P. Sloan Research Fellowship
  4. mathematics departments of UCLA and Yale
  5. DOD Counterdrug Technology Development Program Office
  6. Division Of Mathematical Sciences [0748488, 0941476] Funding Source: National Science Foundation

向作者/读者索取更多资源

Recently popularized randomized methods for principal component analysis (PCA) efficiently and reliably produce nearly optimal accuracy-even on parallel processors-unlike the classical (deterministic) alternatives. We adapt one of these randomized methods for use with data sets that are too large to be stored in random-access memory (RAM). (The traditional terminology is that our procedure works efficiently out-of-core.) We illustrate the performance of the algorithm via several numerical examples. For example, we report on the PCA of a data set stored on disk that is so large that less than a hundredth of it can fit in our computer's RAM.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据