Large-Scale Machine Learning with Stochastic Gradient Descent
From MaRDI portal
Publication:3298463
DOI10.1007/978-3-7908-2604-3_16zbMath1436.68293OpenAlexW114517082MaRDI QIDQ3298463
Publication date: 14 July 2020
Published in: Proceedings of COMPSTAT'2010 (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1007/978-3-7908-2604-3_16
Related Items (only showing first 100 items - show all)
Coupled Generation ⋮ Stochastic distributed learning with gradient quantization and double-variance reduction ⋮ Some Limit Properties of Markov Chains Induced by Recursive Stochastic Algorithms ⋮ An inexact first-order method for constrained nonlinear optimization ⋮ Randomized numerical linear algebra: Foundations and algorithms ⋮ Mean-field inference methods for neural networks ⋮ On block accelerations of quantile randomized Kaczmarz for corrupted systems of linear equations ⋮ An Augmented Lagrangian Deep Learning Method for Variational Problems with Essential Boundary Conditions ⋮ Unnamed Item ⋮ Quantile-Based Iterative Methods for Corrupted Systems of Linear Equations ⋮ The inverse variance–flatness relation in stochastic gradient descent is critical for finding flat minima ⋮ Wasserstein-Based Projections with Applications to Inverse Problems ⋮ Random Batch Particle Methods for the Homogeneous Landau Equation ⋮ A Consensus-Based Global Optimization Method with Adaptive Momentum Estimation ⋮ Unnamed Item ⋮ Unnamed Item ⋮ Unnamed Item ⋮ An Acceleration Strategy for Randomize-Then-Optimize Sampling Via Deep Neural Networks ⋮ Population based optimization via differential evolution and adaptive fractional gradient descent ⋮ An off-the-grid approach to multi-compartment magnetic resonance fingerprinting ⋮ Fractional-based stochastic gradient algorithms for time-delayed ARX models ⋮ Multistart global optimization with tunnelling and an evolutionary strategy supervised by a martingale ⋮ SCORE: approximating curvature information under self-concordant regularization ⋮ Privacy-preserving federated learning on lattice quantization ⋮ A distributed optimisation framework combining natural gradient with Hessian-free for discriminative sequence training ⋮ Fast Calibration for Computer Models with Massive Physical Observations ⋮ A mini-batch stochastic conjugate gradient algorithm with variance reduction ⋮ PFLM: privacy-preserving federated learning with membership proof ⋮ A literature survey of matrix methods for data science ⋮ Monte Carlo simulation of SDEs using GANs ⋮ Approximation capabilities of measure-preserving neural networks ⋮ A Bregman stochastic method for nonconvex nonsmooth problem beyond global Lipschitz gradient continuity ⋮ Traffic sensor location using Wardrop equilibrium ⋮ Solving nonconvex energy minimization problems in martensitic phase transitions with a mesh-free deep learning approach ⋮ Cluster‐based gradient method for stochastic optimal control problems with elliptic partial differential equation constraint ⋮ Bi-level algorithm for optimizing hyperparameters in penalized nonnegative matrix factorization ⋮ Scaling up stochastic gradient descent for non-convex optimisation ⋮ Reconciling privacy and utility: an unscented Kalman filter-based framework for differentially private machine learning ⋮ On Stochastic and Deterministic Quasi-Newton Methods for Nonstrongly Convex Optimization: Asymptotic Convergence and Rate Analysis ⋮ Ergodicity and long-time behavior of the Random Batch Method for interacting particle systems ⋮ Probabilistic forecast reconciliation: properties, evaluation and score optimisation ⋮ Regularized Newton Method with Global \({\boldsymbol{\mathcal{O}(1/{k}^2)}}\) Convergence ⋮ Spherical graph drawing by multi-dimensional scaling ⋮ Byzantine-robust variance-reduced federated learning over distributed non-i.i.d. data ⋮ Optimal Treatment Regimes: A Review and Empirical Comparison ⋮ A new large-scale learning algorithm for generalized additive models ⋮ A modified stochastic quasi-Newton algorithm for summing functions problem in machine learning ⋮ A Variational Neural Network Approach for Glacier Modelling with Nonlinear Rheology ⋮ Importance sampling in signal processing applications ⋮ Linear convergence of distributed estimation with constraints and communication delays ⋮ Stochastic projective splitting ⋮ Nonlinear Gradient Mappings and Stochastic Optimization: A General Framework with Applications to Heavy-Tail Noise ⋮ Handling high-dimensional data with missing values by modern machine learning techniques ⋮ Parametric level-set inverse problems with stochastic background estimation ⋮ Bi-fidelity modeling of uncertain and partially unknown systems using DeepONets ⋮ Accelerated doubly stochastic gradient descent for tensor CP decomposition ⋮ Using Wasserstein generative adversarial networks for the design of Monte Carlo simulations ⋮ Unnamed Item ⋮ On the discrepancy principle for stochastic gradient descent ⋮ Unnamed Item ⋮ Collaborative filtering for massive multinomial data ⋮ On a Scalable Entropic Breaching of the Overfitting Barrier for Small Data Problems in Machine Learning ⋮ A Continuous-Time Analysis of Distributed Stochastic Gradient ⋮ Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review ⋮ Recent Theoretical Advances in Non-Convex Optimization ⋮ Near optimal step size and momentum in gradient descent for quadratic functions ⋮ Surpassing Gradient Descent Provably: A Cyclic Incremental Method with Linear Convergence Rate ⋮ Optimization Methods for Large-Scale Machine Learning ⋮ Distributed Newton Methods for Deep Neural Networks ⋮ Unnamed Item ⋮ Unnamed Item ⋮ Quasi-Newton methods: superlinear convergence without line searches for self-concordant functions ⋮ A globally convergent gradient-like method based on the Armijo line search ⋮ Unnamed Item ⋮ Bayesian differential programming for robust systems identification under uncertainty ⋮ A Distributed Optimal Control Problem with Averaged Stochastic Gradient Descent ⋮ An Adaptive Surrogate Modeling Based on Deep Neural Networks for Large-Scale Bayesian Inverse Problems ⋮ An Optimal Algorithm for Decentralized Finite-Sum Optimization ⋮ Solving Allen-Cahn and Cahn-Hilliard Equations using the Adaptive Physics Informed Neural Networks ⋮ IQN: An Incremental Quasi-Newton Method with Local Superlinear Convergence Rate ⋮ Stochastic gradient descent, weighted sampling, and the randomized Kaczmarz algorithm ⋮ Scalable estimation strategies based on stochastic approximations: classical results and new insights ⋮ CoverBLIP: accelerated and scalable iterative matched-filtering for magnetic resonance fingerprint reconstruction* ⋮ One Step Back, Two Steps Forward: Interference and Learning in Recurrent Neural Networks ⋮ Making the Last Iterate of SGD Information Theoretically Optimal ⋮ Stochastic Primal-Dual Coordinate Method for Regularized Empirical Risk Minimization ⋮ Probabilistic Line Searches for Stochastic Optimization ⋮ Unnamed Item ⋮ Unnamed Item ⋮ Unnamed Item ⋮ On the fast convergence of random perturbations of the gradient flow ⋮ Approximate survey propagation for statistical inference ⋮ A dual approach for optimal algorithms in distributed optimization over networks ⋮ Unnamed Item ⋮ Joint Online Parameter Estimation and Optimal Sensor Placement for the Partially Observed Stochastic Advection-Diffusion Equation ⋮ Spurious minimizers in non uniform Fourier sampling optimization ⋮ Reconstructing the Thermal Phonon Transmission Coefficient at Solid Interfaces in the Phonon Transport Equation ⋮ A Stochastic Proximal Alternating Minimization for Nonsmooth and Nonconvex Optimization ⋮ On Gradient-Based Learning in Continuous Games ⋮ Stochastic subgradient descent method for large-scale robust chance-constrained support vector machines
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Optimal aggregation of classifiers in statistical learning.
- Support-vector networks
- Acceleration of Stochastic Approximation by Averaging
- The importance of convexity in learning with squared loss
- On‐line learning for very large data sets
- On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities
- Some applications of concentration inequalities to statistics
This page was built for publication: Large-Scale Machine Learning with Stochastic Gradient Descent