Fit without fear: remarkable mathematical phenomena of deep learning through the prism of interpolation
From MaRDI portal
Publication:5887828
Recommendations
Cites work
- scientific article; zbMATH DE number 823069 (Why is no real title available?)
- scientific article; zbMATH DE number 7626719 (Why is no real title available?)
- scientific article; zbMATH DE number 7626737 (Why is no real title available?)
- scientific article; zbMATH DE number 6781369 (Why is no real title available?)
- scientific article; zbMATH DE number 5060482 (Why is no real title available?)
- scientific article; zbMATH DE number 3222478 (Why is no real title available?)
- scientific article; zbMATH DE number 3313108 (Why is no real title available?)
- scientific article; zbMATH DE number 3371284 (Why is no real title available?)
- 10.1162/153244303321897690
- A Correspondence Between Bayesian Estimation on Stochastic Processes and Smoothing by Splines
- A decision-theoretic generalization of on-line learning and an application to boosting
- A distribution-free theory of nonparametric regression
- A jamming transition from under- to over-parametrization affects generalization in deep learning
- A randomized Kaczmarz algorithm with exponential convergence
- Advanced Lectures on Machine Learning
- Benign overfitting in linear regression
- Boosting the margin: a new explanation for the effectiveness of voting methods
- Comment on: Boosting algorithms: regularization, prediction and model fitting
- Convex optimization: algorithms and complexity
- Deep double descent: where bigger models and more data hurt*
- Deep learning
- Gauss and the invention of least squares
- Just interpolate: kernel ``ridgeless regression can generalize
- Learning Theory
- Nearest neighbor pattern classification
- Occam's razor
- On early stopping in gradient descent learning
- Overparameterized neural networks implement associative memory
- Reconciling modern machine-learning practice and the classical bias-variance trade-off
- Scattered Data Approximation
- Stochastic gradient descent, weighted sampling, and the randomized Kaczmarz algorithm
- Surprises in high-dimensional ridgeless least squares interpolation
- The Generalization Error of Random Features Regression: Precise Asymptotics and the Double Descent Curve
- The Hilbert kernel regression estimate.
- The elements of statistical learning. Data mining, inference, and prediction
- Two models of double descent for weak features
- When do neural networks outperform kernel methods?*
- Wide neural networks of any depth evolve as linear models under gradient descent *
Cited in
(19)- The Modern Mathematics of Deep Learning
- Mini-workshop: Interpolation and over-parameterization in statistics and machine learning. Abstracts from the mini-workshop held September 17--22, 2023
- A moment-matching approach to testable learning and a new characterization of Rademacher complexity
- A data-dependent approach for high-dimensional (robust) Wasserstein alignment
- An adaptively weighted stochastic gradient MCMC algorithm for Monte Carlo simulation and global optimization
- Convergence analysis for over-parameterized deep learning
- Reconciling modern machine-learning practice and the classical bias-variance trade-off
- New equivalences between interpolation and SVMs: kernels and structured features
- Overparameterized maximum likelihood tests for detection of sparse vectors
- Benign overfitting and adaptive nonparametric regression
- Tractability from overparametrization: the example of the negative perceptron
- Recent Theoretical Advances in Non-Convex Optimization
- Overparameterization and generalization error: weighted trigonometric interpolation
- Double data piling: a high-dimensional solution for asymptotically perfect multi-category classification
- Stopping rules for gradient methods for non-convex problems with additive noise in gradient
- Differentiability in unrolled training of neural physics simulators on transient dynamics
- A mathematical perspective of machine learning
- The energy landscape of the Kuramoto model in random geometric graphs in a circle
- Training adaptive reconstruction networks for blind inverse problems
This page was built for publication: Fit without fear: remarkable mathematical phenomena of deep learning through the prism of interpolation
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5887828)