Large-Scale Machine Learning with Stochastic Gradient Descent

From MaRDI portal
Publication:3298463

DOI10.1007/978-3-7908-2604-3_16zbMath1436.68293OpenAlexW114517082MaRDI QIDQ3298463

Léon Bottou

Publication date: 14 July 2020

Published in: Proceedings of COMPSTAT'2010 (Search for Journal in Brave)

Full work available at URL: https://doi.org/10.1007/978-3-7908-2604-3_16




Related Items (only showing first 100 items - show all)

Coupled GenerationStochastic distributed learning with gradient quantization and double-variance reductionSome Limit Properties of Markov Chains Induced by Recursive Stochastic AlgorithmsAn inexact first-order method for constrained nonlinear optimizationRandomized numerical linear algebra: Foundations and algorithmsMean-field inference methods for neural networksOn block accelerations of quantile randomized Kaczmarz for corrupted systems of linear equationsAn Augmented Lagrangian Deep Learning Method for Variational Problems with Essential Boundary ConditionsUnnamed ItemQuantile-Based Iterative Methods for Corrupted Systems of Linear EquationsThe inverse variance–flatness relation in stochastic gradient descent is critical for finding flat minimaWasserstein-Based Projections with Applications to Inverse ProblemsRandom Batch Particle Methods for the Homogeneous Landau EquationA Consensus-Based Global Optimization Method with Adaptive Momentum EstimationUnnamed ItemUnnamed ItemUnnamed ItemAn Acceleration Strategy for Randomize-Then-Optimize Sampling Via Deep Neural NetworksPopulation based optimization via differential evolution and adaptive fractional gradient descentAn off-the-grid approach to multi-compartment magnetic resonance fingerprintingFractional-based stochastic gradient algorithms for time-delayed ARX modelsMultistart global optimization with tunnelling and an evolutionary strategy supervised by a martingaleSCORE: approximating curvature information under self-concordant regularizationPrivacy-preserving federated learning on lattice quantizationA distributed optimisation framework combining natural gradient with Hessian-free for discriminative sequence trainingFast Calibration for Computer Models with Massive Physical ObservationsA mini-batch stochastic conjugate gradient algorithm with variance reductionPFLM: privacy-preserving federated learning with membership proofA literature survey of matrix methods for data scienceMonte Carlo simulation of SDEs using GANsApproximation capabilities of measure-preserving neural networksA Bregman stochastic method for nonconvex nonsmooth problem beyond global Lipschitz gradient continuityTraffic sensor location using Wardrop equilibriumSolving nonconvex energy minimization problems in martensitic phase transitions with a mesh-free deep learning approachCluster‐based gradient method for stochastic optimal control problems with elliptic partial differential equation constraintBi-level algorithm for optimizing hyperparameters in penalized nonnegative matrix factorizationScaling up stochastic gradient descent for non-convex optimisationReconciling privacy and utility: an unscented Kalman filter-based framework for differentially private machine learningOn Stochastic and Deterministic Quasi-Newton Methods for Nonstrongly Convex Optimization: Asymptotic Convergence and Rate AnalysisErgodicity and long-time behavior of the Random Batch Method for interacting particle systemsProbabilistic forecast reconciliation: properties, evaluation and score optimisationRegularized Newton Method with Global \({\boldsymbol{\mathcal{O}(1/{k}^2)}}\) ConvergenceSpherical graph drawing by multi-dimensional scalingByzantine-robust variance-reduced federated learning over distributed non-i.i.d. dataOptimal Treatment Regimes: A Review and Empirical ComparisonA new large-scale learning algorithm for generalized additive modelsA modified stochastic quasi-Newton algorithm for summing functions problem in machine learningA Variational Neural Network Approach for Glacier Modelling with Nonlinear RheologyImportance sampling in signal processing applicationsLinear convergence of distributed estimation with constraints and communication delaysStochastic projective splittingNonlinear Gradient Mappings and Stochastic Optimization: A General Framework with Applications to Heavy-Tail NoiseHandling high-dimensional data with missing values by modern machine learning techniquesParametric level-set inverse problems with stochastic background estimationBi-fidelity modeling of uncertain and partially unknown systems using DeepONetsAccelerated doubly stochastic gradient descent for tensor CP decompositionUsing Wasserstein generative adversarial networks for the design of Monte Carlo simulationsUnnamed ItemOn the discrepancy principle for stochastic gradient descentUnnamed ItemCollaborative filtering for massive multinomial dataOn a Scalable Entropic Breaching of the Overfitting Barrier for Small Data Problems in Machine LearningA Continuous-Time Analysis of Distributed Stochastic GradientDeep Convolutional Neural Networks for Image Classification: A Comprehensive ReviewRecent Theoretical Advances in Non-Convex OptimizationNear optimal step size and momentum in gradient descent for quadratic functionsSurpassing Gradient Descent Provably: A Cyclic Incremental Method with Linear Convergence RateOptimization Methods for Large-Scale Machine LearningDistributed Newton Methods for Deep Neural NetworksUnnamed ItemUnnamed ItemQuasi-Newton methods: superlinear convergence without line searches for self-concordant functionsA globally convergent gradient-like method based on the Armijo line searchUnnamed ItemBayesian differential programming for robust systems identification under uncertaintyA Distributed Optimal Control Problem with Averaged Stochastic Gradient DescentAn Adaptive Surrogate Modeling Based on Deep Neural Networks for Large-Scale Bayesian Inverse ProblemsAn Optimal Algorithm for Decentralized Finite-Sum OptimizationSolving Allen-Cahn and Cahn-Hilliard Equations using the Adaptive Physics Informed Neural NetworksIQN: An Incremental Quasi-Newton Method with Local Superlinear Convergence RateStochastic gradient descent, weighted sampling, and the randomized Kaczmarz algorithmScalable estimation strategies based on stochastic approximations: classical results and new insightsCoverBLIP: accelerated and scalable iterative matched-filtering for magnetic resonance fingerprint reconstruction*One Step Back, Two Steps Forward: Interference and Learning in Recurrent Neural NetworksMaking the Last Iterate of SGD Information Theoretically OptimalStochastic Primal-Dual Coordinate Method for Regularized Empirical Risk MinimizationProbabilistic Line Searches for Stochastic OptimizationUnnamed ItemUnnamed ItemUnnamed ItemOn the fast convergence of random perturbations of the gradient flowApproximate survey propagation for statistical inferenceA dual approach for optimal algorithms in distributed optimization over networksUnnamed ItemJoint Online Parameter Estimation and Optimal Sensor Placement for the Partially Observed Stochastic Advection-Diffusion EquationSpurious minimizers in non uniform Fourier sampling optimizationReconstructing the Thermal Phonon Transmission Coefficient at Solid Interfaces in the Phonon Transport EquationA Stochastic Proximal Alternating Minimization for Nonsmooth and Nonconvex OptimizationOn Gradient-Based Learning in Continuous GamesStochastic subgradient descent method for large-scale robust chance-constrained support vector machines


Uses Software


Cites Work


This page was built for publication: Large-Scale Machine Learning with Stochastic Gradient Descent