4.3 Article

Analysis and performance estimation of the Conjugate Gradient method on multiple GPUs

期刊

PARALLEL COMPUTING
卷 38, 期 10-11, 页码 552-575

出版社

ELSEVIER
DOI: 10.1016/j.parco.2012.07.002

关键词

Conjugate Gradient method; Sparse-Matrix Vector multiplication; Block Compressed Sparse Row format; Performance analysis; Performance estimation; Multiple GPUs

向作者/读者索取更多资源

The Conjugate Gradient (CG) method is a widely-used iterative method for solving linear systems described by a (sparse) matrix. The method requires a large amount of Sparse-Matrix Vector (SpMV) multiplications, vector reductions and other vector operations to be performed. We present a number of mappings for the SpMV operation on modern programmable CPUs using the Block Compressed Sparse Row (BCSR) format. Further, we show that reordering matrix blocks substantially improves the performance of the SpMV operation, especially when small blocks are used, so that our method outperforms existing state-of-the-art approaches, in most cases. Finally, a thorough analysis of the performance of both SpMV and CG methods is performed, which allows us to model and estimate the expected maximum performance for a given (unseen) problem. (C) 2012 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.3
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据