Higher-order quasi-Monte Carlo training of deep neural networks

From MaRDI portal
Publication:5015302




Abstract: We present a novel algorithmic approach and an error analysis leveraging Quasi-Monte Carlo points for training deep neural network (DNN) surrogates of Data-to-Observable (DtO) maps in engineering design. Our analysis reveals higher-order consistent, deterministic choices of training points in the input data space for deep and shallow Neural Networks with holomorphic activation functions such as tanh. These novel training points are proved to facilitate higher-order decay (in terms of the number of training samples) of the underlying generalization error, with consistency error bounds that are free from the curse of dimensionality in the input data space, provided that DNN weights in hidden layers satisfy certain summability conditions. We present numerical experiments for DtO maps from elliptic and parabolic PDEs with uncertain inputs that confirm the theoretical analysis.



Cites work



Describes a project that uses

Uses Software





This page was built for publication: Higher-order quasi-Monte Carlo training of deep neural networks

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5015302)