期刊
JOURNAL OF MACHINE LEARNING RESEARCH
卷 24, 期 -, 页码 -出版社
MICROTOME PUBL
关键词
Deep learning; dimensionality reduction; partial differential equations
High-dimensional spatio-temporal dynamics can be encoded in a low-dimensional subspace. To address practical engineering challenges, we propose a general framework called Neural Implicit Flow (NIF), which enables mesh-agnostic, low-rank representation of large-scale parametric spatio-temporal data.
High-dimensional spatio-temp oral dynamics can often be encoded in a low-dimensional sub-space. Engineering applications for modeling, characterization, design, and control of such large-scale systems often rely on dimensionality reduction to make solutions computationally tractable in real time. Common existing paradigms for dimensionality reduction include linear methods, such as the singular value decomposition (SVD), and nonlinear methods, such as variants of convolutional autoencoders (CAE). However, these encoding techniques lack the ability to efficiently represent the complexity associated with spatio-temp oral data, which often requires variable geometry, non-uniform grid resolution, adaptive meshing, and/or parametric dependencies. To resolve these practical engineering challenges, we propose a general framework called Neural Implicit Flow (NIF) that enables a mesh-agnostic, low-rank representation of large-scale, parametric, spatial-temp oral data. NIF consists of two modified multilayer perceptrons (MLPs): (i) ShapeNet, which isolates and represents the spatial complexity, and (ii) ParameterNet, which accounts for any other input complexity, including parametric dependencies, time, and sensor measurements. We demonstrate the utility of NIF for parametric surrogate modeling, enabling the interpretable represen-tation and compression of complex spatio-temp oral dynamics, efficient many-spatial-query tasks, and improved generalization performance for sparse reconstruction.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据