4.7 Article

Multi-GPU implementation of a time-explicit finite volume solver using CUDA and a CUDA-Aware version of OpenMPI with application to shallow water flows

期刊

COMPUTER PHYSICS COMMUNICATIONS
卷 271, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.cpc.2021.108190

关键词

Flood simulations; Shallow-water equations; Multi-GPU; CUDA; MPI; HPC

资金

  1. National Sciences and Engineering Research Council of Canada
  2. Hydro-Quebec [RDCPJ 491880-15, RGPIN 132923-11]

向作者/读者索取更多资源

This study presents the development of a multi-GPU version of a time-explicit finite volume solver for the Shallow-Water Equations on a multi-GPU architecture, utilizing MPI, CUDA-Fortran, and the METIS library. By using multiple GPUs to accelerate message passing and conducting efficiency studies, it was found that efficiencies of over 80% can be achieved.
This paper shows the development of a multi-GPU version of a time-explicit finite volume solver for the Shallow-Water Equations (SWE) on a multi-GPU architecture. MPI is combined with CUDA-Fortran in order to use as many GPUs as needed and the METIS library is leveraged to perform a domain decomposition on the 2D unstructured triangular meshes of interest. A CUDA-Aware version of OpenMPI is adopted to speed up the messages between the MPI processes. A study of both speed-up and efficiency is conducted; first, for a classic dam-break flow in a canal, and then for two real domains with complex bathymetries. In both cases, meshes with up to 12 million cells are used. Using 24 to 28 GPUs on these meshes leads to an efficiency of 80% and more. Finally, the multi-GPU version is compared to the pure MPI multi-CPU version, and it is concluded that in this particular case, about 100 CPU cores would be needed to achieve the same performance as one GPU. The developed methodology is applicable for general time-explicit Riemann solvers for conservation laws. (C) 2021 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据