scientific article; zbMATH DE number 7255167
From MaRDI portal
Publication:4969246
Benjamin Gess, Arnulf Jentzen, Benjamin Fehrman
Publication date: 5 October 2020
Full work available at URL: https://arxiv.org/abs/1904.01517
Title: zbMATH Open Web Interface contents unavailable due to conflicting licenses.
Related Items (14)
A proof of convergence for gradient descent in the training of artificial neural networks for constant target functions ⋮ Landscape analysis for shallow neural networks: complete classification of critical points for affine target functions ⋮ A proof of convergence for stochastic gradient descent in the training of artificial neural networks with ReLU activation for constant target functions ⋮ Full error analysis for the training of deep neural networks ⋮ Stochastic gradient descent with noise of machine learning type. I: Discrete time analysis ⋮ On minimal representations of shallow ReLU networks ⋮ Overall error analysis for the training of deep neural networks via stochastic gradient descent with random initialisation ⋮ Theoretical analysis of Adam using hyperparameters close to one without Lipschitz smoothness ⋮ Convergence Analysis of a Quasi-Monte CarloBased Deep Learning Algorithm for Solving Partial Differential Equations ⋮ Central limit theorems for stochastic gradient descent with averaging for stable manifolds ⋮ Neural network regression for Bermudan option pricing ⋮ Solving the Kolmogorov PDE by means of deep learning ⋮ Convergence analysis for gradient flows in the training of artificial neural networks with ReLU activation ⋮ SABRINA: a stochastic subspace majorization-minimization algorithm
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Logarithmic regret algorithms for online convex optimization
- Error bounds and convergence analysis of feasible descent methods: A general approach
- Linear convergence of first order methods for non-strongly convex optimization
- Adaptivity of averaged stochastic gradient descent to local strong convexity for logistic regression
- Reducing the Dimensionality of Data with Neural Networks
- Large-Scale Machine Learning with Stochastic Gradient Descent
- Shorter Notes: Regularity of the Distance Function
- Robust Stochastic Approximation Approach to Stochastic Programming
- Degenerate Nonlinear Programming with a Quadratic Growth Condition
- Optimization Methods for Large-Scale Machine Learning
- Stochastic First- and Zeroth-Order Methods for Nonconvex Stochastic Programming
- An Asynchronous Parallel Stochastic Coordinate Descent Algorithm
- A Stochastic Approximation Method
- Mini-batch stochastic approximation methods for nonconvex stochastic composite optimization
This page was built for publication: