Learning ReLU Networks on Linearly Separable Data: Algorithm, Optimality, and Generalization
From MaRDI portal
Publication:5238780
DOI10.1109/TSP.2019.2904921zbMath1458.68185arXiv1808.04685OpenAlexW2885208219MaRDI QIDQ5238780
Gang Wang, Jie Chen, Georgios B. Giannakis
Publication date: 28 October 2019
Published in: IEEE Transactions on Signal Processing (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1808.04685
Applications of mathematical programming (90C90) Learning and adaptive systems in artificial intelligence (68T05)
Related Items (10)
Revisiting Landscape Analysis in Deep Neural Networks: Eliminating Decreasing Paths to Infinity ⋮ High-Dimensional Learning Under Approximate Sparsity with Applications to Nonsmooth Estimation and Regularized Neural Networks ⋮ On the Benefit of Width for Neural Networks: Disappearance of Basins ⋮ Lagrangian dual theory and stability analysis for fuzzy optimization problems ⋮ Efficient deep data assimilation with sparse observations and time-varying sensors ⋮ Optimization for deep learning: an overview ⋮ Solving Allen-Cahn and Cahn-Hilliard Equations using the Adaptive Physics Informed Neural Networks ⋮ Bias of homotopic gradient descent for the hinge loss ⋮ Suboptimal Local Minima Exist for Wide Neural Networks with Smooth Activations ⋮ Generalised latent assimilation in heterogeneous reduced spaces with machine learning surrogate models
This page was built for publication: Learning ReLU Networks on Linearly Separable Data: Algorithm, Optimality, and Generalization