4.7 Article

Reliable extrapolation of deep neural operators informed by physics or sparse observations

出版社

ELSEVIER SCIENCE SA
DOI: 10.1016/j.cma.2023.116064

关键词

Neural operators; DeepONet; Extrapolation complexity; Fine-tuning; Multifidelity learning; Out-of-distribution inference

向作者/读者索取更多资源

Deep neural operators can learn nonlinear mappings between infinite-dimensional function spaces. They provide a new simulation paradigm in science and engineering as surrogate solvers of partial differential equations. However, pure data-driven neural operators and deep learning models are usually limited to interpolation scenarios, and extrapolation may lead to large errors and failure. In this study, we systematically investigate the extrapolation behavior of DeepONets and propose a bias-variance trade-off strategy for extrapolation. We also develop a complete workflow and propose reliable learning methods that guarantee safe prediction under extrapolation conditions.
Deep neural operators can learn nonlinear mappings between infinite-dimensional function spaces via deep neural networks. As promising surrogate solvers of partial differential equations (PDEs) for real-time prediction, deep neural operators such as deep operator networks (DeepONets) provide a new simulation paradigm in science and engineering. Pure data-driven neural operators and deep learning models, in general, are usually limited to interpolation scenarios, where new predictions utilize inputs within the support of the training set. However, in the inference stage of real-world applications, the input may lie outside the support, i.e., extrapolation is required, which may result to large errors and unavoidable failure of deep learning models. Here, we address this challenge of extrapolation for deep neural operators. First, we systematically investigate the extrapolation behavior of DeepONets by quantifying the extrapolation complexity, via the 2-Wasserstein distance between two function spaces and propose a new strategy of bias-variance trade-off for extrapolation with respect to model capacity. Subsequently, we develop a complete workflow, including extrapolation determination, and we propose five reliable learning methods that guarantee a safe prediction under extrapolation by requiring additional information-the governing PDEs of the system or sparse new observations. The proposed methods are based on either fine-tuning a pre-trained DeepONet or multifidelity learning. We demonstrate the effectiveness of the proposed framework for various types of parametric PDEs. Our systematic comparisons provide practical guidelines for selecting a proper extrapolation method depending on the available information, desired accuracy, and required inference speed.(c) 2023 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据