Higher-order quasi-Monte Carlo training of deep neural networks

From MaRDI portal
Publication:5015302

DOI10.1137/20M1369373zbMATH Open1492.65043arXiv2009.02713OpenAlexW3216945074MaRDI QIDQ5015302FDOQ5015302


Authors: Marcello Longo, Siddhartha Mishra, T. Konstantin Rusch, Christoph Schwab Edit this on Wikidata


Publication date: 7 December 2021

Published in: SIAM Journal on Scientific Computing (Search for Journal in Brave)

Abstract: We present a novel algorithmic approach and an error analysis leveraging Quasi-Monte Carlo points for training deep neural network (DNN) surrogates of Data-to-Observable (DtO) maps in engineering design. Our analysis reveals higher-order consistent, deterministic choices of training points in the input data space for deep and shallow Neural Networks with holomorphic activation functions such as tanh. These novel training points are proved to facilitate higher-order decay (in terms of the number of training samples) of the underlying generalization error, with consistency error bounds that are free from the curse of dimensionality in the input data space, provided that DNN weights in hidden layers satisfy certain summability conditions. We present numerical experiments for DtO maps from elliptic and parabolic PDEs with uncertain inputs that confirm the theoretical analysis.


Full work available at URL: https://arxiv.org/abs/2009.02713




Recommendations




Cites Work


Cited In (10)

Uses Software





This page was built for publication: Higher-order quasi-Monte Carlo training of deep neural networks

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5015302)