Deep double descent: where bigger models and more data hurt*
From MaRDI portal
Publication:5020041
DOI10.1088/1742-5468/ac3a74OpenAlexW4206410067MaRDI QIDQ5020041
Gal Kaplun, Yamini Bansal, Preetum Nakkiran, Tristan Yang, Ilya Sutskever, Boaz Barak
Publication date: 3 January 2022
Published in: Journal of Statistical Mechanics: Theory and Experiment (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1912.02292
Related Items (13)
Sensitivity-Informed Provable Pruning of Neural Networks ⋮ Fit without fear: remarkable mathematical phenomena of deep learning through the prism of interpolation ⋮ On the influence of optimizers in deep learning-based side-channel analysis ⋮ Binary Classification of Gaussian Mixtures: Abundance of Support Vectors, Benign Overfitting, and Regularization ⋮ Overparameterization and Generalization Error: Weighted Trigonometric Interpolation ⋮ Reliable extrapolation of deep neural operators informed by physics or sparse observations ⋮ Consistent Sparse Deep Learning: Theory and Computation ⋮ Semi-Supervised Node Classification via Semi-Global Graph Transformer Based on Homogeneity Augmentation ⋮ Stability of the scattering transform for deformations with minimal regularity ⋮ Fragility, robustness and antifragility in deep learning ⋮ On the robustness of sparse counterfactual explanations to adverse perturbations ⋮ On the properties of bias-variance decomposition for kNN regression ⋮ A Unifying Tutorial on Approximate Message Passing
Uses Software
Cites Work
This page was built for publication: Deep double descent: where bigger models and more data hurt*