Nonparametric regression using deep neural networks with ReLU activation function (Q2215715)

From MaRDI portal
scientific article
Language Label Description Also known as
English
Nonparametric regression using deep neural networks with ReLU activation function
scientific article

    Statements

    Nonparametric regression using deep neural networks with ReLU activation function (English)
    0 references
    14 December 2020
    0 references
    The author considers the already classical nonparametric regression model $Y_i =f_0(\mathbf{X} _i) +\epsilon _i $, where $\mathbf{X}_i\in[0,1]^d$, $Y_i\in \mathbb R$, $i=1,\dots,n$, $\mathbf{X} _i$ are the observed vectors, $Y_i$ are the responses and $\epsilon _i$ are the noise variables. The statistical problem is to recover $f_0$ from the samples ($\mathbf{X}_i$, $Y_i$). In the present paper, the estimation of the regression function is done by means of neural networks technology. In the second section, multilayer feedforward artificial neural networks are introduced and concepts as deep neural networks, deep learning, sparsity of the neural network, rectified linear unit (ReLU) are explained. The third section is devoted to the presentation of main results. The regression function is assumed to be a composition of several functions. The author proves that the estimators based on sparsely connected deep neural networks with ReLU activation function and for an appropriate network architecture can achieve minimax rates of convergence. Special cases of compositional constrains on the regression function are discussed in the fourth section. Suboptimality of wavelet series estimators is discussed in the fifth section and a summary on some achievements on statistical theory and neural networks can be found in the sixth section. The seventh section contains the proofs of results. The author reports that additional proofs are to be found in a supplementary material: ``Nonparametric regression using deep neural networks with ReLU activation function'', \url{doi:10.1214/19-AOS1875SUPP}.
    0 references
    nonparametric regression
    0 references
    multilayer neural networks
    0 references
    rectified linear activation function
    0 references
    rectified linear unit (ReLU)
    0 references
    minimax estimation risk
    0 references
    additive models
    0 references
    wavelets
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references