Optimal approximation rate of ReLU networks in terms of width and depth
From MaRDI portal
Publication:2065073
Abstract: This paper concentrates on the approximation power of deep feed-forward neural networks in terms of width and depth. It is proved by construction that ReLU networks with width and depth can approximate a H"older continuous function on with an approximation rate , where and are H"older order and constant, respectively. Such a rate is optimal up to a constant in terms of width and depth separately, while existing results are only nearly optimal without the logarithmic factor in the approximation rate. More generally, for an arbitrary continuous function on , the approximation rate becomes , where is the modulus of continuity. We also extend our analysis to any continuous function on a bounded set. Particularly, if ReLU networks with depth and width are used to approximate one-dimensional Lipschitz continuous functions on with a Lipschitz constant , the approximation rate in terms of the total number of parameters, , becomes , which has not been discovered in the literature for fixed-depth ReLU networks.
Recommendations
- Deep network approximation characterized by number of neurons
- Optimal approximation of piecewise smooth functions using deep ReLU neural networks
- Error bounds for approximations with deep ReLU networks
- On sharpness of an error bound for deep ReLU network approximation
- Deep network with approximation error being reciprocal of width to power of square root of depth
Cites work
- A note on the expressive power of deep rectified linear unit networks in high-dimensional spaces
- A priori estimates of the population risk for two-layer neural networks
- Approximation by superpositions of a sigmoidal function
- Approximation rates for neural networks with general activation functions
- Deep ReLU Networks Overcome the Curse of Dimensionality for Generalized Bandlimited Functions
- Deep network approximation characterized by number of neurons
- Deep network with approximation error being reciprocal of width to power of square root of depth
- Efficient distribution-free learning of probabilistic concepts
- Error bounds for approximations with deep ReLU networks
- Exponential convergence of the deep neural network approximation for analytic functions
- Make _1 regularization effective in training sparse CNN
- Multilayer feedforward networks are universal approximators
- Neural network approximation: three hidden layers are enough
- Nonlinear approximation via compositions
- Optimal approximation of piecewise smooth functions using deep ReLU neural networks
- Stochastic modified equations and dynamics of stochastic gradient algorithms. I: Mathematical foundations
- Universal approximation bounds for superpositions of a sigmoidal function
Cited in
(25)- Optimal approximation of piecewise smooth functions using deep ReLU neural networks
- Gauss Newton method for solving variational problems of PDEs with neural network discretizaitons
- Rates of approximation by ReLU shallow neural networks
- Deep Network Approximation for Smooth Functions
- Deep network with approximation error being reciprocal of width to power of square root of depth
- Towards Lower Bounds on the Depth of ReLU Neural Networks
- Deep nonparametric regression on approximate manifolds: nonasymptotic error bounds with polynomial prefactors
- Error bounds for ReLU networks with depth and width parameters
- Deep learning via dynamical systems: an approximation perspective
- How do noise tails impact on deep ReLU networks?
- On sharpness of an error bound for deep ReLU network approximation
- Active learning based sampling for high-dimensional nonlinear partial differential equations
- Computing ground states of Bose-Einstein condensation by normalized deep neural network
- Deep network approximation characterized by number of neurons
- DENSITY RESULTS BY DEEP NEURAL NETWORK OPERATORS WITH INTEGER WEIGHTS
- Weighted variation spaces and approximation by shallow ReLU networks
- ReLU neural networks of polynomial size for exact maximum flow computation
- Universal regular conditional distributions via probabilistic transformers
- Solving PDEs on unknown manifolds with machine learning
- Side effects of learning from low-dimensional data embedded in a Euclidean space
- Mini-workshop: Mathematics of entropic AI in the natural sciences. Abstracts from the mini-workshop held April 7--12, 2024
- Greedy training algorithms for neural networks and applications to PDEs
- Approximation results on nonlinear operators by \(P_p\)-statistical convergence
- Low dimensional approximation and generalization of multivariate functions on smooth manifolds using deep ReLU neural networks
- Deep Neural Networks with ReLU-Sine-Exponential Activations Break Curse of Dimensionality in Approximation on Hölder Class
This page was built for publication: Optimal approximation rate of ReLU networks in terms of width and depth
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2065073)