Analysis of the rate of convergence of fully connected deep neural network regression estimates with smooth activation function (Q2222227)
From MaRDI portal
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Analysis of the rate of convergence of fully connected deep neural network regression estimates with smooth activation function |
scientific article |
Statements
Analysis of the rate of convergence of fully connected deep neural network regression estimates with smooth activation function (English)
0 references
26 January 2021
0 references
There is an investigation on regression estimators based on deep neural networks (DNN). In a previous article, [the author and \textit{M. Kohler}, ``On the rate of convergence of fully connected deep neural network regression estimates'', Preprint, \url{arXiv:1908.11133}], neural networks with rectified linear unit (ReLU) activation function have been considered. The question here is, if the same rate of convergence for fully connected deep neural networks regression estimators with smooth activation function -- the sigmoid -- can be achieved. Indeed, the main result of the present paper, proves that under a set of sufficient conditions, the \(L_2\) -errors of least squares neural network regression estimators based on a set of fully connected DNNs with a fixed number of layers, achieve a similar rate of convergence as in the mentioned article.
0 references
curse of dimensionality
0 references
deep learning
0 references
neural networks
0 references
nonparametric regression
0 references
rate of convergence
0 references
0 references
0 references
0 references
0 references