Adaptive greedy algorithms based on parameter-domain decomposition and reconstruction for the reduced basis method

From MaRDI portal
Publication:6328630

arXiv1911.02214MaRDI QIDQ6328630FDOQ6328630


Authors: Jiahua Jiang, Yanlai Chen Edit this on Wikidata


Publication date: 6 November 2019

Abstract: The reduced basis method (RBM) empowers repeated and rapid evaluation of parametrized partial differential equations through an offline-online decomposition, a.k.a. a learning-execution process. A key feature of the method is a greedy algorithm repeatedly scanning the training set, a fine discretization of the parameter domain, to identify the next dimension of the parameter-induced solution manifold along which we expand the surrogate solution space. Although successfully applied to problems with fairly high parametric dimensions, the challenge is that this scanning cost dominates the offline cost due to it being proportional to the cardinality of the training set which is exponential with respect to the parameter dimension. In this work, we review three recent attempts in effectively delaying this curse of dimensionality, and propose two new hybrid strategies through successive refinement and multilevel maximization of the error estimate over the training set. All five offline-enhanced methods and the original greedy algorithm are tested and compared on {two types of problems: the thermal block problem and the geometrically parameterized Helmholtz problem.













This page was built for publication: Adaptive greedy algorithms based on parameter-domain decomposition and reconstruction for the reduced basis method

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6328630)