Nonparametric regression using deep neural networks with ReLU activation function (Q2215715)

From MaRDI portal





scientific article
Language Label Description Also known as
default for all languages
No label defined
    English
    Nonparametric regression using deep neural networks with ReLU activation function
    scientific article

      Statements

      Nonparametric regression using deep neural networks with ReLU activation function (English)
      0 references
      14 December 2020
      0 references
      The author considers the already classical nonparametric regression model $Y_i =f_0(\mathbf{X} _i) +\epsilon _i $, where $\mathbf{X}_i\in[0,1]^d$, $Y_i\in \mathbb R$, $i=1,\dots,n$, $\mathbf{X} _i$ are the observed vectors, $Y_i$ are the responses and $\epsilon _i$ are the noise variables. The statistical problem is to recover $f_0$ from the samples ($\mathbf{X}_i$, $Y_i$). In the present paper, the estimation of the regression function is done by means of neural networks technology. In the second section, multilayer feedforward artificial neural networks are introduced and concepts as deep neural networks, deep learning, sparsity of the neural network, rectified linear unit (ReLU) are explained. The third section is devoted to the presentation of main results. The regression function is assumed to be a composition of several functions. The author proves that the estimators based on sparsely connected deep neural networks with ReLU activation function and for an appropriate network architecture can achieve minimax rates of convergence. Special cases of compositional constrains on the regression function are discussed in the fourth section. Suboptimality of wavelet series estimators is discussed in the fifth section and a summary on some achievements on statistical theory and neural networks can be found in the sixth section. The seventh section contains the proofs of results. The author reports that additional proofs are to be found in a supplementary material: ``Nonparametric regression using deep neural networks with ReLU activation function'', \url{doi:10.1214/19-AOS1875SUPP}.
      0 references
      nonparametric regression
      0 references
      multilayer neural networks
      0 references
      rectified linear activation function
      0 references
      rectified linear unit (ReLU)
      0 references
      minimax estimation risk
      0 references
      additive models
      0 references
      wavelets
      0 references
      0 references
      0 references
      0 references
      0 references
      0 references
      0 references
      0 references

      Identifiers

      0 references
      0 references
      0 references
      0 references
      0 references
      0 references
      0 references
      0 references
      0 references