A variational perspective on accelerated methods in optimization
From MaRDI portal
Publication:4646231
DOI10.1073/PNAS.1614734113zbMATH Open1404.90098arXiv1603.04245OpenAlexW2963430672WikidataQ37451331 ScholiaQ37451331MaRDI QIDQ4646231FDOQ4646231
Authors: Andre Wibisono, Ashia C. Wilson, Michael Jordan
Publication date: 11 January 2019
Published in: Proceedings of the National Academy of Sciences (Search for Journal in Brave)
Abstract: Accelerated gradient methods play a central role in optimization, achieving optimal rates in many settings. While many generalizations and extensions of Nesterov's original acceleration method have been proposed, it is not yet clear what is the natural scope of the acceleration concept. In this paper, we study accelerated methods from a continuous-time perspective. We show that there is a Lagrangian functional that we call the emph{Bregman Lagrangian} which generates a large class of accelerated methods in continuous time, including (but not limited to) accelerated gradient descent, its non-Euclidean extension, and accelerated higher-order gradient methods. We show that the continuous-time limit of all of these methods correspond to traveling the same curve in spacetime at different speeds. From this perspective, Nesterov's technique and many of its generalizations can be viewed as a systematic way to go from the continuous-time curves generated by the Bregman Lagrangian to a family of discrete-time accelerated algorithms.
Full work available at URL: https://arxiv.org/abs/1603.04245
Recommendations
- Multiscale analysis of accelerated gradient methods
- A variational formulation of accelerated optimization on Riemannian manifolds
- Understanding the acceleration phenomenon via high-resolution differential equations
- A differential equation for modeling Nesterov's accelerated gradient method: theory and insights
- Generalized momentum-based methods: a Hamiltonian perspective
Cited In (only showing first 100 items - show all)
- A differential equation for modeling Nesterov's accelerated gradient method: theory and insights
- Preconditioned accelerated gradient descent methods for locally Lipschitz smooth objectives with applications to the solution of nonlinear PDEs
- Discrete processes and their continuous limits
- Fast primal-dual algorithm via dynamical system for a linearly constrained convex optimization problem
- Online optimization of switched LTI systems using continuous-time and hybrid accelerated gradient flows
- An optimal high-order tensor method for convex optimization
- Triggered gradient tracking for asynchronous distributed optimization
- Explicit stabilised gradient descent for faster strongly convex optimisation
- Title not available (Why is that?)
- Inducing strong convergence of trajectories in dynamical systems associated to monotone inclusions with composite structure
- PDE acceleration: a convergence rate analysis and applications to obstacle problems
- Conformal symplectic and relativistic optimization
- Stochastic Methods for Composite and Weakly Convex Optimization Problems
- A general system of differential equations to model first-order adaptive algorithms
- Proximal gradient flow and Douglas-Rachford splitting dynamics: global exponential stability via integral quadratic constraints
- First-order methods for convex optimization
- On dissipative symplectic integration with applications to gradient-based optimization
- A primal-dual flow for affine constrained convex optimization
- Novel projection neurodynamic approaches for constrained convex optimization
- Implicit Regularization and Momentum Algorithms in Nonlinearly Parameterized Adaptive Control and Prediction
- A Systematic Approach to Lyapunov Analyses of Continuous-Time Models in Convex Optimization
- Selection dynamics for deep neural networks
- Accelerated variational PDEs for efficient solution of regularized inversion problems
- A Continuous-Time Analysis of Distributed Stochastic Gradient
- On the convergence of gradient-like flows with noisy gradient input
- Title not available (Why is that?)
- Continuous relaxations for the traveling salesman problem
- Accelerated optimization on Riemannian manifolds via discrete constrained variational integrators
- Accelerated first-order primal-dual proximal methods for linearly constrained composite convex programming
- Accelerations for a variety of global optimization methods
- Stochastic mirror descent dynamics and their convergence in monotone variational inequalities
- An \(O(s^r)\)-resolution ODE framework for understanding discrete-time algorithms and applications to the linear convergence of minimax problems
- Neurodynamic optimization approaches with finite/fixed-time convergence for absolute value equations
- The approximate duality gap technique: a unified theory of first-order methods
- Title not available (Why is that?)
- Is there an analog of Nesterov acceleration for gradient-based MCMC?
- Accelerated differential inclusion for convex optimization
- Mass-spring-damper networks for distributed optimization in non-Euclidean spaces
- Fractional differential equation approach for convex optimization with convergence rate analysis
- Analysis of optimization algorithms via integral quadratic constraints: nonstrongly convex problems
- Title not available (Why is that?)
- User-friendly guarantees for the Langevin Monte Carlo with inaccurate gradient
- Accelerated information gradient flow
- A control-theoretic perspective on optimal high-order optimization
- Iterative ensemble Kalman methods: a unified perspective with some new variants
- Stochastic modified equations and dynamics of stochastic gradient algorithms. I: Mathematical foundations
- Accelerated schemes for a class of variational inequalities
- Accelerating optimization by tracing valley
- From differential equation solvers to accelerated first-order methods for convex optimization
- Projected dynamical systems on irregular, non-Euclidean domains for nonlinear optimization
- The connections between Lyapunov functions for some optimization algorithms and differential equations
- Revisiting the ODE method for recursive algorithms: fast convergence using quasi stochastic approximation
- Understanding the acceleration phenomenon via high-resolution differential equations
- Generalized momentum-based methods: a Hamiltonian perspective
- Robust hybrid zero-order optimization algorithms with acceleration via averaging in time
- How does momentum benefit deep neural networks architecture design? A few case studies
- Exponential convergence of distributed primal-dual convex optimization algorithm without strong convexity
- Unified acceleration of high-order algorithms under general Hölder continuity
- Accelerating Proximal Markov Chain Monte Carlo by Using an Explicit Stabilized Method
- Accelerations for global optimization covering methods using second derivatives
- Inertial primal-dual dynamics with damping and scaling for linearly constrained convex optimization problems
- Contractivity of Runge-Kutta methods for convex gradient systems
- Adaptive Hamiltonian variational integrators and applications to symplectic accelerated optimization
- A variational formulation of accelerated optimization on Riemannian manifolds
- Bregman Itoh-Abe methods for sparse optimisation
- Deterministic neural networks optimization from a continuous and energy point of view
- High-order symplectic Lie group methods on \(SO(n)\) using the polar decomposition
- Finding extremals of Lagrangian actions
- Convergence rates of inertial primal-dual dynamical methods for separable convex optimization problems
- Finding geodesics joining given points
- Online accelerated data-driven learning for optimal feedback control of discrete-time partially uncertain systems
- Dynamical Systems Theory and Algorithms for NP-hard Problems
- Conformal mirror descent with logarithmic divergences
- Linear convergence of forward-backward accelerated algorithms without knowledge of the modulus of strong convexity
- Potential Function-Based Framework for Minimizing Gradients in Convex and Min-Max Optimization
- A new dynamical system with self-adaptive dynamical stepsize for pseudomonotone mixed variational inequalities
- Practical perspectives on symplectic accelerated optimization
- Differential equations in data analysis
- A Recursively Recurrent Neural Network (R2N2) Architecture for Learning Iterative Algorithms
- Multiscale analysis of accelerated gradient methods
- Fast gradient method for low-rank matrix estimation
- Title not available (Why is that?)
- The regularized submodular maximization via the Lyapunov method
- Accelerated Optimization in the PDE Framework: Formulations for the Manifold of Diffeomorphisms
- Bregman dynamics, contact transformations and convex optimization
- Accelerated Optimization in the PDE Framework Formulations for the Active Contour Case
- Search direction correction with normalized gradient makes first-order methods faster
- From Halpern's fixed-point iterations to Nesterov's accelerated interpretations for root-finding problems
- An ordinary differential equation for modeling Halpern fixed-point Algorithm
- Exploring critical points of energy landscapes: from low-dimensional examples to phase field crystal PDEs
- A second order primal-dual dynamical system for a convex-concave bilinear saddle point problem
- Lagrangian and Hamiltonian dynamics for probabilities on the statistical bundle
- Factor-\(\sqrt{2}\) acceleration of accelerated gradient methods
- Hessian barrier algorithms for linearly constrained optimization problems
- Time-adaptive Lagrangian variational integrators for accelerated optimization
- Accelerated extra-gradient descent: a novel accelerated first-order method
- Near-optimal tensor methods for minimizing the gradient norm of convex functions and accelerated primal–dual tensor methods
- Decentralized concurrent learning with coordinated momentum and restart
- Essential convergence rate of ordinary differential equations appearing in optimization
- Fast symplectic integrator for Nesterov-type acceleration method
This page was built for publication: A variational perspective on accelerated methods in optimization
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4646231)