3.8 Proceedings Paper

Adaptive Sparse Matrix-Matrix Multiplication on the GPU

出版社

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3293883.3295701

关键词

SpGEMM; GPU; Sparse Matrix; Adaptive; ESC; bit-stable

资金

  1. German Research Foundation (DFG) [STE 2565/1-1]
  2. Austrian Science Fund (FWF) [I 3007]
  3. NVIDIA Corporation
  4. Austrian Science Fund (FWF) [I3007] Funding Source: Austrian Science Fund (FWF)

向作者/读者索取更多资源

In the ongoing efforts targeting the vectorization of linear algebra primitives, sparse matrix-matrix multiplication (SpGEMM) has received considerably less attention than sparse Matrix-Vector multiplication (SpMV). While both are equally important, this disparity can be attributed mainly to the additional formidable challenges raised by SpGEMM. In this paper, we present a dynamic approach for addressing SpGEMM on the GPU. Our approach works directly on the standard compressed sparse rows (CSR) data format. In comparison to previous SpGEMM implementations, our approach guarantees a homogeneous, load-balanced access pattern to the first input matrix and improves memory access to the second input matrix. It adaptively re-purposes GPU threads during execution and maximizes the time efficient on-chip scratchpad memory can be used. Adhering to a completely deterministic scheduling pattern guarantees bitstable results during repetitive execution, a property missing from other approaches. Evaluation on an extensive sparse matrix benchmark suggests our approach being the fastest SpGEMM implementation for highly sparse matrices (80% of the set). When bit-stable results are sought, our approach is the fastest across the entire test set.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据