Reduced Basis Approximations of Parameterized Dynamical Partial Differential Equations via Neural Networks
From MaRDI portal
Publication:6380889
arXiv2110.10775MaRDI QIDQ6380889FDOQ6380889
Authors: Peter Sentz, Kristian Beckwith, Eric C. Cyr, Luke Olson, Ravi G. Patel
Publication date: 20 October 2021
Abstract: Projection-based reduced order models are effective at approximating parameter-dependent differential equations that are parametrically separable. When parametric separability is not satisfied, which occurs in both linear and nonlinear problems, projection-based methods fail to adequately reduce the computational complexity. Devising alternative reduced order models is crucial for obtaining efficient and accurate approximations to expensive high-fidelity models. In this work, we develop a time-stepping procedure for dynamical parameter-dependent problems, in which a neural-network is trained to propagate the coefficients of a reduced basis expansion. This results in an online stage with a computational cost independent of the size of the underlying problem. We demonstrate our method on several parabolic partial differential equations, including a problem that is not parametrically separable.
This page was built for publication: Reduced Basis Approximations of Parameterized Dynamical Partial Differential Equations via Neural Networks
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6380889)