Convergence rate analysis for deep Ritz method

From MaRDI portal
Publication:5077692

DOI10.4208/CICP.OA-2021-0195zbMATH Open1491.65117arXiv2103.13330OpenAlexW4220826888MaRDI QIDQ5077692FDOQ5077692


Authors: Chenguang Duan, Y. M. Lai, Dingwei Li, Xiliang Lu, Yu Ling Jiao, Jerry Zhijian Yang Edit this on Wikidata


Publication date: 19 May 2022

Published in: Communications in Computational Physics (Search for Journal in Brave)

Abstract: Using deep neural networks to solve PDEs has attracted a lot of attentions recently. However, why the deep learning method works is falling far behind its empirical success. In this paper, we provide a rigorous numerical analysis on deep Ritz method (DRM) cite{wan11} for second order elliptic equations with Neumann boundary conditions. We establish the first nonasymptotic convergence rate in H1 norm for DRM using deep networks with mathrmReLU2 activation functions. In addition to providing a theoretical justification of DRM, our study also shed light on how to set the hyper-parameter of depth and width to achieve the desired convergence rate in terms of number of training samples. Technically, we derive bounds on the approximation error of deep mathrmReLU2 network in H1 norm and on the Rademacher complexity of the non-Lipschitz composition of gradient norm and mathrmReLU2 network, both of which are of independent interest.


Full work available at URL: https://arxiv.org/abs/2103.13330




Recommendations




Cites Work


Cited In (30)

Uses Software





This page was built for publication: Convergence rate analysis for deep Ritz method

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5077692)