4.7 Article

When and why PINNs fail to train: A neural tangent kernel perspective

期刊

JOURNAL OF COMPUTATIONAL PHYSICS
卷 449, 期 -, 页码 -

出版社

ACADEMIC PRESS INC ELSEVIER SCIENCE
DOI: 10.1016/j.jcp.2021.110768

关键词

Physics-informed neural networks; Spectral bias; Multi-task learning; Gradient descent; Scientific machine learning

资金

  1. DOE [DE-SC0019116]
  2. AFOSR [FA9550-20-1-0060]
  3. DOE-ARPA [DE-AR0001201]
  4. U.S. Department of Energy (DOE) [DE-SC0019116] Funding Source: U.S. Department of Energy (DOE)

向作者/读者索取更多资源

This work investigates the Neural Tangent Kernel (NTK) of Physics-informed neural networks (PINNs) and demonstrates that it can converge to a deterministic kernel that remains constant during training under appropriate conditions. A novel gradient descent algorithm is proposed to adaptively calibrate the convergence rate of total training error using the eigenvalues of NTK. A series of numerical experiments are conducted to validate the theory and practical effectiveness of the proposed algorithms.
Physics-informed neural networks (PINNs) have lately received great attention thanks to their flexibility in tackling a wide range of forward and inverse problems involving partial differential equations. However, despite their noticeable empirical success, little is known about how such constrained neural networks behave during their training via gradient descent. More importantly, even less is known about why such models sometimes fail to train at all. In this work, we aim to investigate these questions through the lens of the Neural Tangent Kernel (NTK); a kernel that captures the behavior of fully-connected neural networks in the infinite width limit during training via gradient descent. Specifically, we derive the NTK of PINNs and prove that, under appropriate conditions, it converges to a deterministic kernel that stays constant during training in the infinite-width limit. This allows us to analyze the training dynamics of PINNs through the lens of their limiting NTK and find a remarkable discrepancy in the convergence rate of the different loss components contributing to the total training error. To address this fundamental pathology, we propose a novel gradient descent algorithm that utilizes the eigenvalues of the NTK to adaptively calibrate the convergence rate of the total training error. Finally, we perform a series of numerical experiments to verify the correctness of our theory and the practical effectiveness of the proposed algorithms. The data and code accompanying this manuscript are publicly available at https://github.com/PredictiveIntelligenceLab/PINNsNTK. (C) 2021 Elsevier Inc. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据