The following pages link to Entropy-SGD (Q1354032):
Displaying 20 items.
- Stochastic backward Euler: an implicit gradient descent algorithm for \(k\)-means clustering (Q1632217) (← links)
- Forward stability of ResNet and its variants (Q1988347) (← links)
- Selection dynamics for deep neural networks (Q2003969) (← links)
- Chaos and complexity from quantum neural network. A study with diffusion metric in machine learning (Q2032539) (← links)
- Bias of homotopic gradient descent for the hinge loss (Q2045131) (← links)
- On Bayesian posterior mean estimators in imaging sciences and Hamilton-Jacobi partial differential equations (Q2051535) (← links)
- Interpretable machine learning: fundamental principles and 10 grand challenges (Q2074414) (← links)
- A spin glass model for the loss surfaces of generative adversarial networks (Q2076047) (← links)
- Optimization for deep learning: an overview (Q2218095) (← links)
- Deep relaxation: partial differential equations for optimizing deep neural networks (Q2319762) (← links)
- Run-and-inspect method for nonconvex optimization and global optimality bounds for R-local minimizers (Q2425163) (← links)
- Building a telescope to look into high-dimensional image spaces (Q3121213) (← links)
- (Q4558200) (← links)
- Global Minima of Overparameterized Neural Networks (Q4999400) (← links)
- Entropic gradient descent algorithms and wide flat minima* (Q5020063) (← links)
- Wasserstein-Based Projections with Applications to Inverse Problems (Q5074785) (← links)
- Ensemble Kalman inversion: a derivative-free technique for machine learning tasks (Q5197869) (← links)
- Comparing dynamics: deep neural networks versus glassy systems (Q5854115) (← links)
- The committee machine: computational to statistical gaps in learning a two-layers neural network (Q5854128) (← links)
- Dynamics of stochastic gradient descent for two-layer neural networks in the teacher–student setup* (Q5857458) (← links)